# Hartley (unit)

The **hartley** (symbol **Hart**), also called a **ban**, or a **dit** (short for **d**ecimal dig**it**),^{[1]}^{[2]}^{[3]} is a logarithmic unit that measures information or entropy, based on base 10 logarithms and powers of 10. One hartley is the information content of an event if the probability of that event occurring is 1⁄10.^{[4]} It is therefore equal to the information contained in one decimal digit (or dit), assuming *a priori* equiprobability of each possible value. It is named after Ralph Hartley.

If base 2 logarithms and powers of 2 are used instead, then the unit of information is the bit, or shannon, which is the information content of an event if the probability of that event occurring is 1⁄2. Natural logarithms and powers of e define the nat.

One ban corresponds to ln(10) nat = log_{2}(10) bit or Sh, or approximately 2.303 nat, or 3.322 bit.^{[a]} A **deciban** is one tenth of a ban (or about 0.332 bit); the name is formed from *ban* by the SI prefix *deci-*.

Though there is no associated SI unit, information entropy is part of the International System of Quantities, defined by International Standard IEC 80000-13 of the International Electrotechnical Commission.

The term *hartley* is named after Ralph Hartley, who suggested in 1928 to measure information using a logarithmic base equal to the number of distinguishable states in its representation, which would be the base 10 for a decimal digit.^{[5]}^{[6]}

The *ban* and the *deciban* were invented by Alan Turing with Irving John "Jack" Good in 1940, to measure the amount of information that could be deduced by the codebreakers at Bletchley Park using the Banburismus procedure, towards determining each day's unknown setting of the German naval Enigma cipher machine. The name was inspired by the enormous sheets of card, printed in the town of Banbury about 30 miles away, that were used in the process.^{[7]}

Good argued that the sequential summation of *decibans* to build up a measure of the weight of evidence in favour of a hypothesis, is essentially Bayesian inference.^{[7]} Donald A. Gillies, however, argued the *ban* is, in effect, the same as Karl Popper's measure of the severity of a test.^{[8]}

The deciban is a particularly useful unit for log-odds, notably as a measure of information in Bayes factors, odds ratios (ratio of odds, so log is difference of log-odds), or weights of evidence. 10 decibans corresponds to odds of 10:1; 20 decibans to 100:1 odds, etc. According to Good, a change in a weight of evidence of 1 deciban (i.e., a change in the odds from evens to about 5:4) is about as finely as humans can reasonably be expected to quantify their degree of belief in a hypothesis.^{[9]}

Odds corresponding to integer decibans can often be well-approximated by simple integer ratios; these are collated below. Value to two decimal places, simple approximation (to within about 5%), with more accurate approximation (to within 1%) if simple one is inaccurate: