-
3
-
-
0041877169
-
Estimation of entropy and mutual information
-
Paninski, L. Estimation of entropy and mutual information. Neural Comput. 2003, 15, 1191-1253.
-
(2003)
Neural Comput.
, vol.15
, pp. 1191-1253
-
-
Paninski, L.1
-
5
-
-
0027297096
-
Covariation of mutations in the V3 loop of human immunodeficiency virus type 1 envelope protein: An information theoretic analysis
-
Korber, B.; Farber, R.M.; Wolpert, D.H.; Lapedes, A.S. Covariation of mutations in the V3 loop of human immunodeficiency virus type 1 envelope protein: An information theoretic analysis. Proc. Natl. Acad. Sci. USA 1993, 90, 7176-7180.
-
(1993)
Proc. Natl. Acad. Sci. USA
, vol.90
, pp. 7176-7180
-
-
Korber, B.1
Farber, R.M.2
Wolpert, D.H.3
Lapedes, A.S.4
-
6
-
-
0000328645
-
Estimating functions of probability distributions from a finite set of samples
-
Wolpert, D.; Wolf, D. Estimating functions of probability distributions from a finite set of samples. Phys. Rev. E 1995, 52, 6841-6854.
-
(1995)
Phys. Rev. E
, vol.52
, pp. 6841-6854
-
-
Wolpert, D.1
Wolf, D.2
-
7
-
-
84887165416
-
Estimating functions of probability distributions from a finite set of samples, Part II
-
arXiv:comp-gas/9403002.
-
Wolf, D.R.; Wolf, D.R.; Wolpert, D.H. Estimating functions of probability distributions from a finite set of samples, Part II: Bayes Estimators for Mutual Information, Chi-Squared, Covariance, and other Statistics. 1994, arXiv:comp-gas/9403002.
-
(1994)
Bayes Estimators for Mutual Information, Chi-Squared, Covariance, and other Statistics
-
-
Wolf, D.R.1
Wolf, D.R.2
Wolpert, D.H.3
-
8
-
-
85035242587
-
Erratum: Estimating functions of probability distributions from a finite set of samples
-
Wolpert, D.; Wolf, D. Erratum: Estimating functions of probability distributions from a finite set of samples. Phys. Rev. E 1996, 54, 6973.
-
(1996)
Phys. Rev. E
, vol.54
, pp. 6973
-
-
Wolpert, D.1
Wolf, D.2
-
10
-
-
84887132218
-
-
MIT Lincoln Laboratory Technical Report 1169; Massachusetts Institute of Technology: Lexington, MA, USA, (accessed on 28 October 2013).
-
Hurley, M.; Kao, E. Numerical Estimation of Information Numerical Estimation of Information Theoretic Measures for Large datasets; MIT Lincoln Laboratory Technical Report 1169; Massachusetts Institute of Technology: Lexington, MA, USA, 2013; Available online: http://www.dtic.mil/dtic/tr/fulltext/u2/a580524.pdf (accessed on 28 October 2013).
-
(2013)
Numerical Estimation of Information Numerical Estimation of Information Theoretic Measures for Large datasets
-
-
Hurley, M.1
Kao, E.2
-
11
-
-
84877790753
-
Bayesian estimation of discrete entropy with mixtures of stick-breaking priors
-
Proceedings of the 26th Annual Conference on Neural Information Processing Systems 2012, 3-6 December, Lake Tahoe, NV, USA
-
Archer, E.; Park, I.; Pillow, J. Bayesian estimation of discrete entropy with mixtures of stick-breaking priors. In Advances in Neural Information Processing Systems 25, Proceedings of the 26th Annual Conference on Neural Information Processing Systems 2012, 3-6 December 2012, Lake Tahoe, NV, USA; pp. 2024-2032.
-
(2012)
Advances in Neural Information Processing Systems 25
, pp. 2024-2032
-
-
Archer, E.1
Park, I.2
Pillow, J.3
-
12
-
-
0042777690
-
Entropy and inference, revisited
-
Dietterich, T., Ed.; MIT Press: Cambridge, MA, USA
-
Nemenman, I.; Shafee, F.; Bialek, W. Entropy and inference, revisited. In Advances in Neural Information Processing System; Dietterich, T., Ed.; MIT Press: Cambridge, MA, USA, 2003.
-
(2003)
Advances in Neural Information Processing System
-
-
Nemenman, I.1
Shafee, F.2
Bialek, W.3
-
13
-
-
85036164237
-
Entropy and information in neural spike trains: Progress on the sampling problem
-
Nemenman, I.; Bialek, W.; van Steveninck, R.D.R. Entropy and information in neural spike trains: Progress on the sampling problem. Phys. Rev. E 2004, 69, e056111.
-
(2004)
Phys. Rev. E
, vol.69
-
-
Nemenman, I.1
Bialek, W.2
van Steveninck, R.D.R.3
-
14
-
-
41849129516
-
Neural coding of natural stimuli: information at sub-millisecond resolution
-
Nemenman, I.; Lewen, G.D.; Bialek, W.; van Steveninck, R.R.D.R. Neural coding of natural stimuli: information at sub-millisecond resolution. PLoS Comput. Biol. 2008, 4, e1000025.
-
(2008)
PLoS Comput. Biol.
, vol.4
-
-
Nemenman, I.1
Lewen, G.D.2
Bialek, W.3
van Steveninck, R.R.D.R.4
-
15
-
-
84878363315
-
Bayesian and quasi-Bayesian estimators for mutual information from discrete data
-
Archer, E.; Park, I.M.; Pillow, J.W. Bayesian and quasi-Bayesian estimators for mutual information from discrete data. Entropy 2013, 15, 1738-1755.
-
(2013)
Entropy
, vol.15
, pp. 1738-1755
-
-
Archer, E.1
Park, I.M.2
Pillow, J.W.3
-
16
-
-
34548391942
-
Coverage-adjusted entropy estimation
-
Vu, V.; Yu, B.; Kass, R. Coverage-adjusted entropy estimation. Stat. Med. 2007, 26, 4039-4060.
-
(2007)
Stat. Med.
, vol.26
, pp. 4039-4060
-
-
Vu, V.1
Yu, B.2
Kass, R.3
-
19
-
-
84855371905
-
Coincidences and estimation of entropies of random variables with large cardinalities
-
Nemenman, I. Coincidences and estimation of entropies of random variables with large cardinalities. Entropy 2011, 13, 2013-2023.
-
(2011)
Entropy
, vol.13
, pp. 2013-2023
-
-
Nemenman, I.1
-
20
-
-
0345591789
-
Reconciling Bayesian and non-Bayesian Analysis
-
Kluwer Academic Publishers: Dordrecht, Netherlands
-
Wolpert, D.H. Reconciling Bayesian and non-Bayesian Analysis. In Maximum Entropy and Bayesian Methods; Kluwer Academic Publishers: Dordrecht, Netherlands, 1996; pp. 79-86.
-
(1996)
Maximum Entropy and Bayesian Methods
, pp. 79-86
-
-
Wolpert, D.H.1
-
21
-
-
85128067285
-
The Relationship between PAC, the Statistical Physics Framework, the Bayesian Framework, and the VC Framework
-
Addison-Wesley: Indianapolis, IN, USA
-
Wolpert, D.H. The Relationship between PAC, the Statistical Physics Framework, the Bayesian Framework, and the VC Framework. In The Mathematics of Generalization; Addison-Wesley: Indianapolis, IN, USA, 1995; pp. 117-215.
-
(1995)
The Mathematics of Generalization
, pp. 117-215
-
-
Wolpert, D.H.1
-
22
-
-
84887203320
-
-
Note that no particular property is required of the relation among multiple unseen random variables that may (or may not) exist.
-
Note that no particular property is required of the relation among multiple unseen random variables that may (or may not) exist.
-
-
-
-
23
-
-
21144477583
-
Estimating the number of species: A review
-
Bunge, J.; Fitzpatrick, M. Estimating the number of species: A review. J. Am. Stat. Assoc. 1993, 88, 364-373.
-
(1993)
J. Am. Stat. Assoc.
, vol.88
, pp. 364-373
-
-
Bunge, J.1
Fitzpatrick, M.2
-
24
-
-
84887156499
-
-
The proof of this proposition uses moment-generating functions with Fourier decompositions of the prior, π. To ensure we do not divide by zero, we have to introduce the constant 1+ε into that proof, and to ensure the convergence of our resultant Taylor decompositions, we have to assume infinite differentiability. This is the reason for the peculiar technical condition.
-
The proof of this proposition uses moment-generating functions with Fourier decompositions of the prior, π. To ensure we do not divide by zero, we have to introduce the constant 1+ε into that proof, and to ensure the convergence of our resultant Taylor decompositions, we have to assume infinite differentiability. This is the reason for the peculiar technical condition.
-
-
-
-
26
-
-
84887148475
-
-
This is an example of a more general phenomenon, that in some statistical scenarios there is a low probability that the likelihood of a randomly sampled dataset is concentrated near the truth (various "paradoxes" of statistics can be traced to this phenomenon). See also the discussion of Bayes factors below.
-
This is an example of a more general phenomenon, that in some statistical scenarios there is a low probability that the likelihood of a randomly sampled dataset is concentrated near the truth (various "paradoxes" of statistics can be traced to this phenomenon). See also the discussion of Bayes factors below.
-
-
-
-
27
-
-
84887191504
-
-
Intuitively, for that P(c), the likelihood says that a dataset, f125, 125,. g, would imply a relatively high value of c and, therefore, a high probability that ρ is close to uniform over all bins. Given this, it also implies a low value of |Z|, since if there were any more bins than the eight that have counts, we almost definitely would have seen them seen them for almost-uniform ρ. Conversely, for the dataset, {691,., 6}, the implication is that c must be low where some rare bins trail out, and as a result, there might be a few more bins who had zero counts.
-
Intuitively, for that P(c), the likelihood says that a dataset, f125, 125,. g, would imply a relatively high value of c and, therefore, a high probability that ρ is close to uniform over all bins. Given this, it also implies a low value of |Z|, since if there were any more bins than the eight that have counts, we almost definitely would have seen them seen them for almost-uniform ρ. Conversely, for the dataset, {691,., 6}, the implication is that c must be low where some rare bins trail out, and as a result, there might be a few more bins who had zero counts.
-
-
-
-
28
-
-
84887130917
-
-
Of course, since the formulas in WW implicitly assume c = m, care must be taken to insert appropriate pseudo-counts into those formulas if we want to use a value, c, that differs from m.
-
Of course, since the formulas in WW implicitly assume c = m, care must be taken to insert appropriate pseudo-counts into those formulas if we want to use a value, c, that differs from m.
-
-
-
-
29
-
-
84887144255
-
Estimating functions of probability distributions from a finite set of samples, Part 1
-
arXiv:comp-gas/9403001.
-
Wolpert, D.; Wolf, D. Estimating functions of probability distributions from a finite set of samples, Part 1: Bayes Estimators and the Shannon Entropy. 1994, arXiv:comp-gas/9403001.
-
(1994)
Bayes Estimators and the Shannon Entropy
-
-
Wolpert, D.1
Wolf, D.2
|