×
Well done. You've clicked the tower. This would actually achieve something if you had logged in first. Use the key for that. The name takes you home. This is where all the applicables sit. And you can't apply any changes to my site unless you are logged in.

Our policy is best summarized as "we don't care about _you_, we care about _them_", no emails, so no forgetting your password. You have no rights. It's like you don't even exist. If you publish material, I reserve the right to remove it, or use it myself.

Don't impersonate. Don't name someone involuntarily. You can lose everything if you cross the line, and no, I won't cancel your automatic payments first, so you'll have to do it the hard way. See how serious this sounds? That's how serious you're meant to take these.

×
Register


Required. 150 characters or fewer. Letters, digits and @/./+/-/_ only.
  • Your password can’t be too similar to your other personal information.
  • Your password must contain at least 8 characters.
  • Your password can’t be a commonly used password.
  • Your password can’t be entirely numeric.

Enter the same password as before, for verification.
Login

Grow A Dic
Define A Word
Make Space
Set Task
Mark Post
Apply Votestyle
Create Votes
(From: saved spaces)
Exclude Votes
Apply Dic
Exclude Dic

Click here to flash read.

Estimating the underlying distribution from \textit{iid} samples is a
classical and important problem in statistics. When the alphabet size is large
compared to number of samples, a portion of the distribution is highly likely
to be unobserved or sparsely observed. The missing mass, defined as the sum of
probabilities $\text{Pr}(x)$ over the missing letters $x$, and the Good-Turing
estimator for missing mass have been important tools in large-alphabet
distribution estimation. In this article, given a positive function $g$ from
$[0,1]$ to the reals, the missing $g$-mass, defined as the sum of
$g(\text{Pr}(x))$ over the missing letters $x$, is introduced and studied. The
missing $g$-mass can be used to investigate the structure of the missing part
of the distribution. Specific applications for special cases such as
order-$\alpha$ missing mass ($g(p)=p^{\alpha}$) and the missing Shannon entropy
($g(p)=-p\log p$) include estimating distance from uniformity of the missing
distribution and its partial estimation. Minimax estimation is studied for
order-$\alpha$ missing mass for integer values of $\alpha$ and exact minimax
convergence rates are obtained. Concentration is studied for a class of
functions $g$ and specific results are derived for order-$\alpha$ missing mass
and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal
worst-case variance factors are derived. Two new notions of concentration,
named strongly sub-Gamma and filtered sub-Gaussian concentration, are
introduced and shown to result in right tail bounds that are better than those
obtained from sub-Gaussian concentration.

Click here to read this post out
ID: 161586; Unique Viewers: 0
Unique Voters: 0
Total Votes: 0
Votes:
Latest Change: May 30, 2023, 7:30 a.m. Changes:
Dictionaries:
Words:
Spaces:
Views: 12
CC:
No creative common's license
Comments: