Servicios
Descargas
Buscar
Idiomas
P. Completa
Effective bounds of the variance of statistics on multisets of necklaces
Arvydas Karbonskis; Eugenijus Manstavičius
Arvydas Karbonskis; Eugenijus Manstavičius
Effective bounds of the variance of statistics on multisets of necklaces
Vėrinių multiaibių statistikos dispersijos efektyvūs įverčiai
Lietuvos matematikos rinkinys, vol. 61 Ser. A, pp. 7-12, 2020
Vilniaus Universitetas
resúmenes
secciones
referencias
imágenes

Abstract: The variance of a linear statistics on multisets of necklaces is explored. The upper and lower bounds with optimal constants are obtained.

Keywords: Tur´an–Kubilius inequality, polynomials over finite field, additive function.

Summary: Nagrinėjama tiesinės statistikos, apibrėžtos atsitiktinių vėrinių multiaibėje, dispersija. Gauti tikslūs viršutinieji ir apatinieji įverčiai.

Keywords: Turanas–Kubiliaus nelygybė, daugianariai virš baigtinio lauko, priedų funkcija.

Carátula del artículo

Articles

Effective bounds of the variance of statistics on multisets of necklaces

Vėrinių multiaibių statistikos dispersijos efektyvūs įverčiai

Arvydas Karbonskis
Institute of Mathematics, Vilnius University, Lituania
Eugenijus Manstavičius
Institute of Mathematics, Vilnius University, Lituania
Lietuvos matematikos rinkinys, vol. 61 Ser. A, pp. 7-12, 2020
Vilniaus Universitetas

Recepción: 28 Octubre 2020

Publicación: 18 Febrero 2021

1 Introduction and results

Let (P, ǁ · ǁ) be an initial set of weighted objects and

π ( j ) : = | { p P : p = j } | <

for every j = 1, 2, . . . . Examine the set G with the extended weight function ǁ · ǁ of multisets comprised of p ∈ P. Namely, a ∈ G if a = {p1, . . . , pr} and ǁaǁ = ǁp1ǁ + · · · + ǁprǁ including the empty multiset ∅ of weight 0. Then

m ( n ) : = | ς n | : = | { α ς : a = n } | = ( k ) = n j = 1 n ( π ( j ) + k j - 1 k j ) ,

where l(k¯) = 1k1 + + nkn if k¯ = (k1, . . . , kn) N 0 n and n 0 : = { 0 } .

In the present paper, we deal with the multisets for which m(n) = qn, where q ≥ 2 is an arbitrary natural number. If q is a prime power, then may be interpreted as F q * [ t ] , the set of monic polynomials over a finite field Fq. Then P is the subset of irre- ducible polynomials. For an arbitrary such q, there exist combinatorial constructions, called multisets of necklaces satisfying m( n) = qn (see, [1, Example 2.12, p. 43]). For multisets, we have the following relations

π ( n ) = 1 n d | n q n / d μ ( d ) , q n = d | n d π ( d ) , (1)

where in the summations, d runs over natural divisors of n and µ(d) stands for the Möbius function. The equalities are equivalent to the formal power series relation

n = 0 q n x n = 1 1 - q x = j = 1 ( 1 - x 3 ) - π ( j ) .

Take an a ∈ Gn uniformly at random, that is, sample it with probabilityνn({a}) = q−n, n ∈ N and ν0({∅ }) = 1. If kj(a) ≥ 0 is the number of elements pi in a ∈ Gn of weight j, then k(a) = k1(a), . . . , kn(a) is the structure vector of a ∈ Gn satisfying l(k(a)) = n. Its distribution is

v n ( k ( a ) = s ) = 1 { ( s ) = n } q - n j = 1 ( π ( j ) + s j - 1 s j ) . (2)

where s = ( s 1 , . . . , s n ) 0 n and 1.{.} stands for the indicator function.

We are interested in the distribution with respect to νn of the linear statistics

h ( c ) : = h ( c , a ) = c 1 k 1 ( a ) + · · · + c n k n ( a ) , c = ( c 1 , . . . , c n ) n . (3)

The number of components in a is such a function, namely, it equals k1(a)+ +kn(a). We refer to [1] for more sophisticated examples.

The present paper is devoted to the variance of h ( c ) which is a sum of dependent random variables (r.vs) as the relation h ( j , a ) = ( k ( a ) ) = n for each a G n shows. Estimating it, we propose an approach to overcome technical obstacles stemming from dependence.

In the sequel, the expectations and variances with respect to νn will be denoted by En and Vn while, when the probability space (Ω, F , P ) is not specified, we will respectively use the notation E and V. The summation indexes i, j, l, k, m, m1 and m2 will be natural numbers.

Theorem 1 If c n and n 0 , then

V n h ( c ) = 1 j k n c j 2 π ( j ) k q - j k - i l + j k > n i l n , j k n c i c j π ( i ) π ( j ) q - i l - j k . (4)

The sketch of the proof is given at the beginning of Section 2.

t is known [1] that, for a fixed j, the r.v. kj(a) converges in distribution to the r.v. γj distributed according the negative binomial law NB(π(j), q−j). If {γ1, γ2, . .} . are mutually independent, define the statistics Yn = c1γ1 + + nγn. We shall see that the first sum on the right-hand side in (4) is close to VYn; therefore, estimating Vnh(c¯), we use the following quadratic forms:

B n ( c ) : = 1 j k n c j 2 π ( j ) k q - j k , R n ( c ) = m n m q - 2 m ( j | m c j π ( j ) ) 2 .

Theorem 2 If n ≥ 2, then

V n h ( c ) B n ( c ) + 1 2 R n ( c ) . (5)

The inequality becomes an equality for

c j = c j * : = 3 π ( j ) d | j d q d μ ( j d ) - ( 2 n + 1 ) j , 1 j n . (6)

Corollary 1 If n ≥ 2 and c 0 , , then

V n h ( c ) < 3 2 B n ( c ) < ( 3 2 - q - 1 q n q - n ) V Y n . (7)

The inequalities are trivial for functions proportional to h ( j , a ) = n if a G n because of V n h ( j ) = 0 then. A shift of c eliminates this inconvenience. Observe that either of B n ( c - t j ) and R n ( c - t j ) attain their minimums in t at

t = t c : = 2 ( n + 1 ) n m n m q - m j | m c j π ( j ) .

Theorem 3 If n ≥ 3, then

B n ( c t c j ) - 1 3 R n ( c t c j ) V n h ( c ) B n ( c t c j ) - 1 2 R n ( c t c j ) .

Both inequalities are sharp.

Corollary 2If n ≥ 3 and c α j for every α

3 2 B n ( c t c j ) V n h ( c ) < 3 2 B n ( c t c j ) .

The proofs of the last two theorems presented in Section 2 are built upon the ideas and auxiliary results obtained in [4], [2] and [5].

2 Proofs

We firstly recall known facts about random multisets which can be found in [3] and [1, Section 2.3]. Let y ( x ) = ( y 1 ( x ) , y 2 ( x ) , . . . ) be the infinite dimensional vector of independent r.vs having the negative binomial distributions NB(π(j), xj), namely,

P ( γ j ( x ) = m ) = ( π ( j ) + m - 1 m ) ( 1 - x j ) π ( j ) x j m , m = 0 , 1 , . . .

where 0 < x ≤ q−1. Then γj(q−1 ) = γjwhich has been introduced in Introduction. For convenience, we extend k(a) to k(a) := (k1(a), . . . , kn(a), 0, . . . ) and use infinite dimensional vectors. S e t θ ( x ) = 1 y 1 ( x ) + · · · + n , y n ( x ) + ( n + 1 ) y n + 1 ( x ) + , . . . The latter r.v. is well defined if 0 < x < q−1, since the condition of the Boreli–Cantelli lemma is satisfied:

j = 1 p ( γ j ( x ) 0 ) = j = 1 ( 1 - ( 1 - x j ) π ( j ) ) < .

Lemma 1 I f s = ( s 1 , . . . , s j , s j + 1 , . . . ) 0 and 0 < x < q−1, then

v n ( k ( a ) = s ) = p ( γ ( x ) = s | θ ( x ) = n ) .

Proof. Actually, this is Lemma 2.2 in [3] stated there for Fq[t]. The details remain the same in the more general case.

Lemma 2 For a functional Ψ : 0 such that E | Ψ ( γ ( x ) ) | < , , we have

E Ψ ( γ ( x ) ) = ( 1 - q x ) ( Ψ ( 0 ) + n = 1 E n Ψ ( k ( a ) ) q n x n ) , 0 < x < q - 1 .

Proof. Apply Lemma 1 in the double averaging as follows:

E Ψ ( γ ( x ) ) = n = 0 E ( Ψ ( γ ( x ) ) | θ ( x ) = n ) P ( θ ( x ) = n ) = n = 0 E n Ψ ( k ( a ) ) P ( θ ( x ) = n ) .

Proof of Theorem 1. It is straightforward. Applying the last lemma for the relevant Ψ , one can easily find the needed mixed moments of kj(a), 1 ≤ j ≤ n, and further, the variance of the linear combination h(a).

To prove Theorems 2 and 3, we will apply the following lemmas concerning par- ticular matrices and quadratic forms.

Lemma 3LetU = ((uij)), i, j ≤ n, be the symmetric matrix with the entries

u i j = 1 { i + j > n } ( i j ) - 1 / 2 .

The spectrum of U is the set{1, 1/2, 1/3, . . . , ( 1)n−1/n} . The eigenvectors corresponding to the first three eigenvalues areproportional to c r = c r 1 , . . . , c r n ) , wherer = 1, 2, 3and, for j ≤ n,

e 1 j = j , e 2 j = ( 3 j - 2 n - 1 ) j , e 3 j = ( 1 0 j 2 - 6 ( 2 n + 1 ) j + 3 n 2 + 3 n + 2 ) j .

Proof. This is the byproduct of works [4] and [2].

Afterwards, let e r , 1 m n , be the orthogonal basis of n comprised of the eigenvectors of U and x means the transposed vector x .

Lemma 4 If b m and 1 ≤ m ≤ n and n ≥ 2, then

- 1 2 1 m n m b m 2 1 m 1 , m 2 n m 1 + m 2 > n b m 1 b m 2 1 m n m b m 2 . (9)

If n ≥ 3and

m n m b m = 0 , (10)

then

1 m 1 , m 2 n m 1 + m 2 > n b m 1 b m 2 1 3 1 m n m b m 2 . (11)

Moreover, each bound in (9) and (11) are achieved, respectively, for b m = e r m / m , , wherer = 2, 1, 3and e r m have been defined in Lemma 3.

Proof. Inequalities (9) are seen from Lemma 3 after the substitution b m = x m / m , ,m<n since the extreme eigenvalues are 1 and 1/2.

After the same substitution, we further examine the quadratic form with the matrix U . Condition (10) reckons the subspace of vectors x = ( x 1 , . . . , x n ) satisfying x 1 + · · · + x j j + · · · + x n n = x · e 1 = 0 . This subspace is spanned over the first eigenvector. In other words, under (10), only the form values obtained in the subspace L ⊂ Rn spanned over the vectors e 2 , . . . , e n count. Hence

m a x x L ǁ x ǁ - 2 x x m a x 2 r n ( - 1 ) r - 1 / r = 1 / 3 .

Returning to bm, from this we obtain inequality (11).

Proof of Theorem 2. After grouping the summands, expression (4) can be rewritten as follows:

V n h ( c ) = b n ( c ) - 1 m 1 , m 2 n m 1 + m 2 > n ( q - m 1 1 m n c i π ( i ) ) ( q - m 2 j | m 2 c i π ( j ) ) .

Now evidently estimate (5) follows from Lemma 4 with

b m = q - m j | m c j π ( j ) , m n .

Moreover, it becomes an equality if we take c j = c j * satisfying

q - m j | m c j * π ( j ) = 3 m - 2 n - 1 ,

which by the Möbius inversion formula and (1) may be rewritten as (6).

To prove the first assertion of Corollary 1, it suffices to estimate the inner sum in R n ( c ) , , namely,

( j | m c j π ( j ) ) 2 j | m c j 3 π ( j ) j . j | m j π ( j ) = j | m c j 3 π ( j ) j . q m .

Further, using the expression of VYn, we just estimate the remainder:

V Y n - B n ( c ) = j m c j 3 π ( j ) k > n / j k q - j k n q - n j m c j 2 π ( j ) j q j ( q j - 1 ) 2 . q j - 1 q j n q - n - 1 ( q - 1 ) V Y n .

Plugging both estimates into (5), we obtain the first inequality in Corollary 1 with ≤ instead of <. In fact, we obtained the strict inequality since Cauchy’s inequality applied in the last step is strict if c is not proportional to j , and in this exceptional case, V h ( c ) = 0 . .

Proof of Theorem 3. Observe that V n h ( c ) = V n h ( c ) - t n ) - V n h ( c - t j ) for every t . Hence the right-hand inequality follows from (5) applied for the shifted statistics.

To get the lower bound of variance, we combine (4) and (11). We start with

V n h ( c - t c j ) - B n ( c - t c j ) - m 1 , m 2 n m 1 , m 2 > n b m 1 b m 2 ,

where

b m = q - m j | m ( c j - t c j ) π ( j )

and m ≤ n. By the definition of tc the latter sequence satisfies condition (10). Hence by (11),

m 1 , m 2 n m 1 + m 2 > n b m 1 b m 2 , 1 3 m n m b m 2 = 1 3 R n ( c - t c j ) .

This and (4) imply the lower bound. Moreover, the latter is sharp since Lemma 4 assures this by a choice of a particular sequence b m , m n . .

Material suplementario
References
[1] R. Arratia, A.D. Barbour, S. Tavaré. Logarithmic Combinatorial Structures: A Prob- abilistic Approach. EMS Monographs in Mathematics. EMS Publishing House, Zürich, 2003.
[2] Ž. Baronėnas, E. Manstavičius, P. Šapokaitė. A sharp inequality for the variance with respect to the ewens sampling formula. 2019, arXiv:2003.05975v1.
[3] J.C. Hansen. Factorization in fq [.] and brownian motion. Combin. Probab. Comput., .:285–299, 1993.
[4] J. Klimavičius, E. Manstavičius. The Tur´an–Kubilius inequality on permutations. Annales Univ. Sci. Budapest., Sect. Comp., 48:45–51, 2018.
[5] E. Manstavičius. Sharp bounds for the variance of linear statistics on random permuta- tions. Random Struct. Alg., 2020. https://doi.org/10.1002/rsa.20951.
Notas
Buscar:
Contexto
Descargar
Todas
Imágenes
Visor de artículos científicos generados a partir de XML-JATS4R por Redalyc