Abstract: The variance of a linear statistics on multisets of necklaces is explored. The upper and lower bounds with optimal constants are obtained.
Keywords: Tur´an–Kubilius inequality, polynomials over finite field, additive function.
Summary: Nagrinėjama tiesinės statistikos, apibrėžtos atsitiktinių vėrinių multiaibėje, dispersija. Gauti tikslūs viršutinieji ir apatinieji įverčiai.
Keywords: Turanas–Kubiliaus nelygybė, daugianariai virš baigtinio lauko, priedų funkcija.
Articles
Effective bounds of the variance of statistics on multisets of necklaces
Vėrinių multiaibių statistikos dispersijos efektyvūs įverčiai

Recepción: 28 Octubre 2020
Publicación: 18 Febrero 2021
Let (P, ǁ · ǁ) be an initial set of weighted objects and
for every j = 1, 2, . . . . Examine the set G with the extended weight function ǁ · ǁ of multisets comprised of p ∈ P. Namely, a ∈ G if a = {p1, . . . , pr} and ǁaǁ = ǁp1ǁ + · · · + ǁprǁ including the empty multiset ∅ of weight 0. Then
where l(k¯) = 1k1 + + nkn if k¯ = (k1, . . . , kn) and n .
In the present paper, we deal with the multisets for which m(n) = qn, where q ≥ 2 is an arbitrary natural number. If q is a prime power, then may be interpreted as , the set of monic polynomials over a finite field Fq. Then P is the subset of irre- ducible polynomials. For an arbitrary such q, there exist combinatorial constructions, called multisets of necklaces satisfying m( n) = qn (see, [1, Example 2.12, p. 43]). For multisets, we have the following relations
where in the summations, d runs over natural divisors of n and µ(d) stands for the Möbius function. The equalities are equivalent to the formal power series relation
Take an a ∈ Gn uniformly at random, that is, sample it with probabilityνn({a}) = q−n, n ∈ N and ν0({∅ }) = 1. If kj(a) ≥ 0 is the number of elements pi in a ∈ Gn of weight j, then k(a) = k1(a), . . . , kn(a) is the structure vector of a ∈ Gn satisfying l(k(a)) = n. Its distribution is
where and 1.{.} stands for the indicator function.
We are interested in the distribution with respect to νn of the linear statistics
The number of components in a is such a function, namely, it equals k1(a)+ +kn(a). We refer to [1] for more sophisticated examples.
The present paper is devoted to the variance of which is a sum of dependent random variables (r.vs) as the relation for each a shows. Estimating it, we propose an approach to overcome technical obstacles stemming from dependence.
In the sequel, the expectations and variances with respect to νn will be denoted by En and Vn while, when the probability space (Ω, F , P ) is not specified, we will respectively use the notation E and V. The summation indexes i, j, l, k, m, m1 and m2 will be natural numbers.
Theorem 1 If and , then
The sketch of the proof is given at the beginning of Section 2.
t is known [1] that, for a fixed j, the r.v. kj(a) converges in distribution to the r.v. γj distributed according the negative binomial law NB(π(j), q−j). If {γ1, γ2, . .} . are mutually independent, define the statistics Yn = c1γ1 + + nγn. We shall see that the first sum on the right-hand side in (4) is close to VYn; therefore, estimating Vnh(c¯), we use the following quadratic forms:
Theorem 2 If n ≥ 2, then
The inequality becomes an equality for
Corollary 1 If n ≥ 2 and, then
The inequalities are trivial for functions proportional to if because of then. A shift of eliminates this inconvenience. Observe that either of and attain their minimums in at
Theorem 3 If n ≥ 3, then
Both inequalities are sharp.
Corollary 2If n ≥ 3 and for every
The proofs of the last two theorems presented in Section 2 are built upon the ideas and auxiliary results obtained in [4], [2] and [5].
We firstly recall known facts about random multisets which can be found in [3] and [1, Section 2.3]. Let be the infinite dimensional vector of independent r.vs having the negative binomial distributions NB(π(j), xj), namely,
where 0 < x ≤ q−1. Then γj(q−1 ) = γjwhich has been introduced in Introduction. For convenience, we extend k(a) to k(a) := (k1(a), . . . , kn(a), 0, . . . ) and use infinite dimensional vectors.The latter r.v. is well defined if 0 < x < q−1, since the condition of the Boreli–Cantelli lemma is satisfied:
Lemma 1 and 0 < x < q−1, then
Proof. Actually, this is Lemma 2.2 in [3] stated there for Fq[t]. The details remain the same in the more general case.
Lemma 2 For a functional such that , we have
Proof. Apply Lemma 1 in the double averaging as follows:
Proof of Theorem 1. It is straightforward. Applying the last lemma for the relevant Ψ , one can easily find the needed mixed moments of kj(a), 1 ≤ j ≤ n, and further, the variance of the linear combination h(a).
To prove Theorems 2 and 3, we will apply the following lemmas concerning par- ticular matrices and quadratic forms.
Lemma 3LetU = ((uij)), i, j ≤ n, be the symmetric matrix with the entries
The spectrum of U is the set{1, 1/2, 1/3, . . . , ( 1)n−1/n} . The eigenvectors corresponding to the first three eigenvalues areproportional to wherer = 1, 2, 3and, for j ≤ n,
Proof. This is the byproduct of works [4] and [2].
Afterwards, let be the orthogonal basis of comprised of the eigenvectors of U and means the transposed vector .
Lemma 4 If and 1 ≤ m ≤ n and n ≥ 2, then
If n ≥ 3and
then
Moreover, each bound in (9) and (11) are achieved, respectively, for , wherer = 2, 1, 3and have been defined in Lemma 3.
Proof. Inequalities (9) are seen from Lemma 3 after the substitution ,m<n since the extreme eigenvalues are 1 and 1/2.
After the same substitution, we further examine the quadratic form with the matrix U . Condition (10) reckons the subspace of vectors satisfying This subspace is spanned over the first eigenvector. In other words, under (10), only the form values obtained in the subspace L ⊂ Rn spanned over the vectors count. Hence
Returning to bm, from this we obtain inequality (11).
Proof of Theorem 2. After grouping the summands, expression (4) can be rewritten as follows:
Now evidently estimate (5) follows from Lemma 4 with
Moreover, it becomes an equality if we take satisfying
which by the Möbius inversion formula and (1) may be rewritten as (6).
To prove the first assertion of Corollary 1, it suffices to estimate the inner sum in , namely,
Further, using the expression of VYn, we just estimate the remainder:
Plugging both estimates into (5), we obtain the first inequality in Corollary 1 with ≤ instead of <. In fact, we obtained the strict inequality since Cauchy’s inequality applied in the last step is strict if is not proportional to, and in this exceptional case, .
Proof of Theorem 3. Observe that for every . Hence the right-hand inequality follows from (5) applied for the shifted statistics.
To get the lower bound of variance, we combine (4) and (11). We start with
where
and m ≤ n. By the definition of tc the latter sequence satisfies condition (10). Hence by (11),
This and (4) imply the lower bound. Moreover, the latter is sharp since Lemma 4 assures this by a choice of a particular sequence .