back to list

"Lp norm" naming - god damn it

🔗Mike Battaglia <battaglia01@gmail.com>

7/31/2012 4:23:31 AM

Here's what happens should one insist on calling it "the Lp norm" in
all cases, regardless of what JI group you're in.

Say we're in the 2.9.21.11/5 subgroup, and say we want to use a metric
for interval complexity that corresponds to Tenney height. I want to
make it so that ||a/b|| = log(a*b) for all ratios in this subgroup.
IOW, we want to put "the L1 norm" on these intervals.

But that's actually wrong, that's not what we want. The correct norm
is NOT the actual mathematical thing called the "L1 norm" on that
space. It's absolutely not the thing where you just weight the axes
with log(2) and log(9) and log(7*3) and log(11*5) and then start using
a taxicab metric. So we don't want to use an L1 norm at all. We want
to use some weird norm which has a unit sphere which is:

a) the intersection of the 11-limit weighted L1 unit sphere and the
2.3.7.11/5 subspace, which is going to be some weird polytope, and
then
b) sent through a dilation and shear that changes the axes to
2.9.21.11/5, which is now an even weirder polytope.

and whose name is "the norm induced down from the L1 norm".

Let's call these strange norms something else than the Lebesgue norm
on a vector space please. I'll suggest "Tenney norm" or "Tp norm"
again.

-Mike

🔗Mike Battaglia <battaglia01@gmail.com>

7/31/2012 4:31:48 AM

On Tue, Jul 31, 2012 at 7:23 AM, Mike Battaglia <battaglia01@gmail.com> wrote:
>
> Let's call these strange norms something else than the Lebesgue norm
> on a vector space please. I'll suggest "Tenney norm" or "Tp norm"
> again.

Also, this isn't just for the sake of pedantry. It's because it makes
it difficult to communicate and even more difficult to figure out what
other people are saying.

- Earlier tonight, I had a lot of fun trying to express the concept
that Graham's temperament finder is currently using The Actual Thing
Called The Lp Norm on subgroup spaces instead of using The Thing We're
Just Calling The Lp Norm.
- There was some confusion on Facebook about whether "the Lp norm" on
the 2.9.21 subgroup is actually the Lp norm on the 2.9.21 subgroup.
- Earlier this week in Gene's "optimal subgroup" thread I wasn't sure
if Gene were talking about weighted or unweighted Lp norms.

And this is on week one and entirely for conversations between the
only 4-5 people who actually know what these concepts mean.

You know what, screw it, I'm just going to make a wiki page about the
whole thing and we'll see where that leaves everything.

-Mike

🔗Keenan Pepper <keenanpepper@gmail.com>

7/31/2012 1:08:01 PM

--- In tuning-math@yahoogroups.com, Mike Battaglia <battaglia01@...> wrote:
>
> On Tue, Jul 31, 2012 at 7:23 AM, Mike Battaglia <battaglia01@...> wrote:
> >
> > Let's call these strange norms something else than the Lebesgue norm
> > on a vector space please. I'll suggest "Tenney norm" or "Tp norm"
> > again.

I like "Tenney norm" for the L1-based one. And the L2-based one can be the "Tenney-Euclidean norm", which is particularly easy to deal with because it actually is an L2 norm on the subspace. You just need the appropriate weighting matrix.

> Also, this isn't just for the sake of pedantry. It's because it makes
> it difficult to communicate and even more difficult to figure out what
> other people are saying.
>
> - Earlier tonight, I had a lot of fun trying to express the concept
> that Graham's temperament finder is currently using The Actual Thing
> Called The Lp Norm on subgroup spaces instead of using The Thing We're
> Just Calling The Lp Norm.

Right, and for me this is a big feature request for the temperament finder. Currently it uses the correct weighting only when the basis elements are all prime powers (in which case the two norms are identical). Unfortunately I think it'll be non-trivial for Graham to implement because it means you can't just treat the basis elements as floating point numbers.

Keenan

🔗Mike Battaglia <battaglia01@gmail.com>

7/31/2012 3:44:43 PM

On Tue, Jul 31, 2012 at 4:08 PM, Keenan Pepper <keenanpepper@gmail.com> wrote:
>
> I like "Tenney norm" for the L1-based one. And the L2-based one can be the "Tenney-Euclidean norm", which is particularly easy to deal with because it actually is an L2 norm on the subspace. You just need the appropriate weighting matrix.

Not always. Take a look at the 2.9.21 subgroup of the 7-limit.
Assuming we're using the vector space R^n, then the appropriate
subspace is actually just the same as the 2.3.7 subgroup. The induced
norm on the 2.3.7 subgroup will lead to an ordinary 2-sphere as the
unit sphere, meaning we'll still have an L2 norm - but we want 2.9.21,
not 2.3.7. To get to the former from the latter, we'll need to shrink
the 3-axis and apply a shear mapping, which will ruin it, and we won't
have an L2 norm anymore. For instance, |0 0 1> and |0 -1 1>,
corresponding to 21/1 and 7/3 respectively, must now have the same
norm. The unit sphere is now a rotated ellipsoid and not an ordinary
2-sphere. And even if you want to get away with this by weighting the
coordinates instead of changing the behavior of the norm, it's still
not an L2 norm, because then |0 zero something> has to have the same
norm as |0 nonzero something> for 21/1 to have the same norm as 7/3.

Put another way, if what you said were exactly how it worked, Graham's
temperament finder would already be exhibiting the correct behavior.

As for names, what I want is a name for all of these norms, such that
the norm is a function of the limit and that you use the same name for
the weird induced norm on a subspace as you do for the full-limit. If,
for instance, we say that Tenney-Euclidean height is the T2 norm,
where the "L" in "Lp" standing for Lebesgue has been replaced with "T"
for Tenney, then it should also apply to this case. The T2 norm can
reduce to the L2 norm in cases where the subgroup only involves primes
and prime powers, but doesn't necessarily for subgroups, and if a
function of the subgroup you're using.

That's the thing I want a name for. I like Tp norm, where it's
understood that the Tp norm involves the mapping back to full limit +
weighting matrix coordinate change + Lp norm all in one step. Then you
can say that the Tp norm reduces to the Lp norm in cases where only
the last two steps are needed.

> > Also, this isn't just for the sake of pedantry. It's because it makes
> > it difficult to communicate and even more difficult to figure out what
> > other people are saying.
> >
> > - Earlier tonight, I had a lot of fun trying to express the concept
> > that Graham's temperament finder is currently using The Actual Thing
> > Called The Lp Norm on subgroup spaces instead of using The Thing We're
> > Just Calling The Lp Norm.
>
> Right, and for me this is a big feature request for the temperament finder. Currently it uses the correct weighting only when the basis elements are all prime powers (in which case the two norms are identical). Unfortunately I think it'll be non-trivial for Graham to implement because it means you can't just treat the basis elements as floating point numbers.

I offered to do it when I have the time but I don't know when that's
going to be these days, especially because I'm not much of a python
expert. It doesn't seem like it'd be too difficult, but I don't know
how it's coded now.

-Mike

🔗Keenan Pepper <keenanpepper@gmail.com>

8/1/2012 7:09:50 AM

--- In tuning-math@yahoogroups.com, Mike Battaglia <battaglia01@...> wrote:
>
> On Tue, Jul 31, 2012 at 4:08 PM, Keenan Pepper <keenanpepper@...> wrote:
> >
> > I like "Tenney norm" for the L1-based one. And the L2-based one can be the "Tenney-Euclidean norm", which is particularly easy to deal with because it actually is an L2 norm on the subspace. You just need the appropriate weighting matrix.
>
> Not always. Take a look at the 2.9.21 subgroup of the 7-limit.
> Assuming we're using the vector space R^n, then the appropriate
> subspace is actually just the same as the 2.3.7 subgroup. The induced
> norm on the 2.3.7 subgroup will lead to an ordinary 2-sphere as the
> unit sphere, meaning we'll still have an L2 norm - but we want 2.9.21,
> not 2.3.7. To get to the former from the latter, we'll need to shrink
> the 3-axis and apply a shear mapping, which will ruin it, and we won't
> have an L2 norm anymore. For instance, |0 0 1> and |0 -1 1>,
> corresponding to 21/1 and 7/3 respectively, must now have the same
> norm. The unit sphere is now a rotated ellipsoid and not an ordinary
> 2-sphere. And even if you want to get away with this by weighting the
> coordinates instead of changing the behavior of the norm, it's still
> not an L2 norm, because then |0 zero something> has to have the same
> norm as |0 nonzero something> for 21/1 to have the same norm as 7/3.

Yeah, it's a ellipsoid with weird axes. All of this is why I said "weighting *matrix*", not just "weights". As you just pointed out, it's important that it can be an arbitrary matrix.

I could work out this example with the weighting matrix if necessary, but I think you get the idea now.

My point is, after you apply the weighting *matrix*, you end up with new coordinates, and the Tenney-Euclidean norm actually is the L2 norm of those new coordinates. Whereas for the Tenney norm (L1-based), there's no possible linear transformation you can do so that it ends up as an L1 norm. (An ellipse is a linear transformation of a circle, but a hexagon is not a linear transformation of a square.)

> Put another way, if what you said were exactly how it worked, Graham's
> temperament finder would already be exhibiting the correct behavior.

No, because the weighting matrix it currently uses is always diagonal, and always has the diagonal weights as the log of the basis element. If it used the correct, non-diagonal weighting matrices it would be correct, but then to get that you can't just treat the basis elements as floating point numbers. Their prime factorization matters.

> That's the thing I want a name for. I like Tp norm, where it's...

That's fine with me.

Keenan

🔗Mike Battaglia <battaglia01@gmail.com>

8/2/2012 5:01:55 AM

On Wed, Aug 1, 2012 at 10:09 AM, Keenan Pepper <keenanpepper@gmail.com>
wrote:
>
> Yeah, it's a ellipsoid with weird axes. All of this is why I said
> "weighting *matrix*", not just "weights". As you just pointed out, it's
> important that it can be an arbitrary matrix.
>
> I could work out this example with the weighting matrix if necessary, but
> I think you get the idea now.
>
> My point is, after you apply the weighting *matrix*, you end up with new
> coordinates, and the Tenney-Euclidean norm actually is the L2 norm of those
> new coordinates. Whereas for the Tenney norm (L1-based), there's no possible
> linear transformation you can do so that it ends up as an L1 norm. (An
> ellipse is a linear transformation of a circle, but a hexagon is not a
> linear transformation of a square.)

I don't understand what matrices you consider to be "weighting
matrices" then. The only things I've seen people call weighting
matrices around here are diagonal matrices. If on the other hand, by
"weighting matrix" you mean any arbitrary matrix which need not be
diagonal and need not even be square, then there definitely is an
injective but non-surjective linear transformation that you can do to
turn any subgroup Tenney norm back into a full-limit L1 norm. If V is
the V-map for the subgroup as expressed in some full limit, and W is
the diagonal weighting matrix for that limit, then W*V is the
weighting matrix you want. So for some smonzo M, ||(W*V)*M||_1 ends up
being equivalent to the weird hexagonal norm you're talking about.
This works for any Lp norm in general.

On the other hand, if your definition of a weighting matrix is matrix
that doesn't necessarily have to be diagonal, but must always be
square, then yeah, it's impossible.

-Mike

🔗Keenan Pepper <keenanpepper@gmail.com>

8/2/2012 8:26:49 AM

--- In tuning-math@yahoogroups.com, Mike Battaglia <battaglia01@...> wrote:
> I don't understand what matrices you consider to be "weighting
> matrices" then. The only things I've seen people call weighting
> matrices around here are diagonal matrices. If on the other hand, by
> "weighting matrix" you mean any arbitrary matrix which need not be
> diagonal and need not even be square, then there definitely is an
> injective but non-surjective linear transformation that you can do to
> turn any subgroup Tenney norm back into a full-limit L1 norm. If V is
> the V-map for the subgroup as expressed in some full limit, and W is
> the diagonal weighting matrix for that limit, then W*V is the
> weighting matrix you want. So for some smonzo M, ||(W*V)*M||_1 ends up
> being equivalent to the weird hexagonal norm you're talking about.
> This works for any Lp norm in general.

Oh, very good point.

> On the other hand, if your definition of a weighting matrix is matrix
> that doesn't necessarily have to be diagonal, but must always be
> square, then yeah, it's impossible.

That is indeed what I had in mind, because I was thinking of it as just weighting certain axes (which, however, can be different from the coordinate axes). For a non-square matrix the resulting vectors are in a higher-dimensional space, so they can hardly be called "weighted versions" of the same vectors. At least that's how I think of it.

So what I was saying is that for L2, and only L2, you can express the induced norm as an L2 norm applied to the result of a square matrix. The reason is because SVD exists. Any matrix can be decomposed as UDV where U and V are orthogonal (i.e. preserve L2 norm) and D is diagonal. So, even though UDV is non-square, you can drop U and lop off the parts of D and V corresponding to singular values of zero and obtain a square matrix with the required property. This doesn't work for any other Lp norm because there isn't a continuous family of matrices preserving that norm analogous to the orthogonal matrices.

None of which is particularly relevant for practical computation, because as you said you can just use the non-square matrix and be done with it.

Keenan

🔗Keenan Pepper <keenanpepper@gmail.com>

8/2/2012 8:39:47 AM

--- In tuning-math@yahoogroups.com, "Keenan Pepper" <keenanpepper@...> >
> So what I was saying is that for L2, and only L2, you can express the induced norm as an L2 norm applied to the result of a square matrix. The reason is because SVD exists. Any matrix can be decomposed as UDV where U and V are orthogonal (i.e. preserve L2 norm) and D is diagonal. So, even though UDV is non-square, you can drop U and lop off the parts of D and V corresponding to singular values of zero and obtain a square matrix with the required property. This doesn't work for any other Lp norm because there isn't a continuous family of matrices preserving that norm analogous to the orthogonal matrices.

BTW, I believe that for any p != 2, the matrices preserving the Lp norm form a finite group, the full group of symmetries of a hypercube. Only for p = 2 do they form a continuous, infinite group.

Which, again, is merely a curiosity and not relevant for our purposes. Right?

Keenan