Jump to content

Hartley (unit): Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
There might be some people who would like the hartley to supersede the ban, but that has not occurred in practice.
 
(29 intermediate revisions by 19 users not shown)
Line 1: Line 1:
{{Short description|Unit of information}}
{{lowercase title}}
{{redirect|dit (information)|other informational topics|DIT (disambiguation)}}

{{Fundamental info units}}
{{Fundamental info units}}


The '''hartley''' (symbol '''Hart'''), also called a '''ban''', or a '''dit''' (short for '''d'''ecimal dig'''it'''), is a [[logarithmic unit]] which measures [[information]] or [[information entropy|entropy]], based on base 10 [[logarithm]]s and powers of 10, rather than the powers of 2 and [[binary logarithm|base 2 logarithms]] which define the [[bit]], or [[shannon (unit)|shannon]]. One ban or hartley is the information content of an event if the [[probability]] of that event occurring is 1/10.<ref>{{cite web|title=IEC 80000-13:2008|url=http://www.iso.org/iso/catalogue_detail?csnumber=31898|publisher=[[International Organization for Standardization]]|accessdate=21 July 2013}}</ref> It is therefore equal to the information contained in one [[decimal digit]] (or dit), assuming ''[[A priori probability|a priori]]'' equiprobability of each possible value.
The '''hartley''' (symbol '''Hart'''), also called a '''ban''', or a '''dit''' (short for "decimal digit"),<ref name="Klar_1970"/><ref name="Klar_1989"/><ref name="Lukoff_1979"/> is a [[logarithmic unit]] that measures [[information]] or [[information entropy|entropy]], based on base 10 [[logarithm]]s and powers of 10. One hartley is the information content of an event if the [[probability]] of that event occurring is {{frac|1|10}}.<ref name="IEC"/> It is therefore equal to the information contained in one [[decimal digit]] (or dit), assuming ''[[A priori probability|a priori]]'' equiprobability of each possible value. It is named after [[Ralph Hartley]].


If [[binary logarithm|base 2 logarithms]] and powers of 2 are used instead, then the unit of information is the [[shannon (unit)|shannon]] or [[bit]], which is the information content of an event if the [[probability]] of that event occurring is {{frac|1|2}}. [[Natural logarithm]]s and powers of [[E (mathematical constant)|e]] define the [[nat (unit)|nat]].
As a [[bit]] corresponds to a [[binary digit]], so a ban is a [[decimal digit]]. A '''deciban''' is one tenth of a ban; the name is formed from ''ban'' by the [[SI prefix]] ''[[deci-]]''.


One hartley corresponds to log<sub>2</sub>(10) [[bit]] = ln(10) [[Nat (unit)|nat]], or approximately 3.322 [[shannon (unit)|Sh]],{{efn|This value, approximately 10/3, but slightly less, can be understood simply because <math>10^3 = 1,000 \lesssim 1,024 = 2^{10}</math>: 3 decimal digits are slightly ''less'' information than 10 binary digits, so 1 decimal digit is slightly ''less'' than 10/3 binary digits.}} or 2.303 [[nat (unit)|nat]]. A deciban is about 0.332 Sh.
One ban corresponds to ln(10) [[Nat (unit)|nat]] = log<sub>2</sub>(10) [[shannon (unit)|Sh]], or approximately 2.303 [[nat (unit)|nat]], or 3.322 bit (3.322 Sh).{{efn|This value, approximately {{frac|10|3}}, but slightly less, can be understood simply because <math>10^3 = 1,000 \lesssim 1,024 = 2^{10}</math>: 3 decimal digits are slightly ''less'' information than 10 binary digits, so 1 decimal digit is slightly ''less'' than {{frac|10|3}} binary digits.}} A '''deciban''' is one tenth of a ban (or about 0.332 Sh); the name is formed from ''ban'' by the [[SI prefix]] ''[[deci-]]''.


Though not an [[SI unit]], the hartley is part of the [[International System of Quantities]], defined by International Standard [[IEC 80000-13]] of the [[International Electrotechnical Commission]]. It is named after [[Ralph Hartley]].
Though there is no associated [[SI unit]], [[Entropy (information theory)|information entropy]] is part of the [[International System of Quantities]], defined by International Standard [[IEC 80000-13]] of the [[International Electrotechnical Commission]].


== History ==
== History ==
The term ''hartley'' is named after [[Ralph Hartley]], who suggested in 1928 to measure information using a logarithmic base equal to the number of distinguishable states in its representation, which would be the base 10 for a decimal digit.<ref name="Hartley_1928"/><ref name="Reza_1994"/>
The ''ban'' and the ''deciban'' were invented by [[Alan Turing]] with [[I. J. Good]] in 1940, to measure the amount of information that could be deduced by the codebreakers at [[Bletchley Park]] using the [[Banburismus]] procedure, towards determining each day's unknown setting of the German naval [[Enigma machine|Enigma]] cipher machine. The name was inspired by the enormous sheets of card, printed in the town of [[Banbury]] about 30 miles away, that were used in the process.<ref name="good">{{cite journal|last=Good|first=I.J.|authorlink=I. J. Good|title=Studies in the History of Probability and Statistics. XXXVII A. M. Turing's statistical work in World War II|journal=[[Biometrika]]|year=1979|volume=66|issue=2|pages=393&ndash;396|doi=10.1093/biomet/66.2.393|mr=0548210}}</ref>

[[I. J. Good|Jack Good]] argued that the sequential summation of ''decibans'' to build up a measure of the weight of evidence in favour of a hypothesis, is essentially [[Bayesian inference]].<ref name = good/> [[Donald A. Gillies]], however, argued the ''ban'' is, in effect, the same as [[Karl Popper|Karl Popper's]] measure of the severity of a test.<ref>{{cite journal | last = Gillies | first = Donald A. | author-link = Donald A. Gillies | year = 1990 | title = The Turing-Good Weight of Evidence Function and Popper's Measure of the Severity of a Test | periodical = British Journal for the Philosophy of Science | volume = 41 |issue=1 | pages = 143&ndash;146 | mr=055678 | jstor=688010 | doi=10.1093/bjps/41.1.143 }} </ref>


The ''ban'' and the ''deciban'' were invented by [[Alan Turing]] with [[Irving John Good|Irving John "Jack" Good]] in 1940, to measure the amount of information that could be deduced by the codebreakers at [[Bletchley Park]] using the [[Banburismus]] procedure, towards determining each day's unknown setting of the German naval [[Enigma machine|Enigma]] cipher machine. The name was inspired by the enormous sheets of card, printed in the town of [[Banbury]] about 30 miles away, that were used in the process.<ref name="Good_1979"/>
The term ''hartley'' is after [[Ralph Hartley]], who suggested this unit in 1928.<ref>{{cite journal |last=Hartley |first=R.V.L. |authorlink=Ralph Hartley |title=Transmission of Information |date=July 1928 |volume=VII |issue=3 |journal=[[Bell System Technical Journal]] |pages=535&ndash;563 |url=http://dotrose.com/etext/90_Miscellaneous/transmission_of_information_1928b.pdf|accessdate=2008-03-27}}</ref><ref>Reza, Fazlollah M. ''An Introduction to Information Theory.'' New York: Dover, 1994. {{ISBN|0-486-68210-2}}.</ref>


Good argued that the sequential summation of ''decibans'' to build up a measure of the weight of evidence in favour of a hypothesis, is essentially [[Bayesian inference]].<ref name="Good_1979"/> [[Donald A. Gillies]], however, argued the ''ban'' is, in effect, the same as [[Karl Popper|Karl Popper's]] measure of the severity of a test.<ref name="Gillies_1990"/>
The ban pre-dates [[Claude Shannon|Shannon]]'s use of ''bit'' as a unit of information by at least eight years, and remains in use in the early 21st century.<ref>{{cite web |url=http://www.theregister.co.uk/2012/10/05/turing_gchq_boss_speech/ |title=GCHQ boss: Crypto-genius Turing brought tech to British spooks |deadurl=no |accessdate=2013-07-08}}</ref> In the [[IEC_80000-13|International System of Quantities]] it is replaced by the hartley.


==Usage as a unit of odds==
==Usage as a unit of odds==
The deciban is a particularly useful unit for [[log-odds]], notably as a measure of information in [[Bayes factor]]s, [[odds ratio]]s (ratio of odds, so log is difference of log-odds), or [[weight of evidence|weights of evidence]]. 10 decibans corresponds to odds of 10:1; 20 decibans to 100:1 odds, etc. According to [[I. J. Good]], a change in a weight of evidence of 1 deciban (i.e., a change in the odds from evens to about 5:4) is about as finely as humans can reasonably be expected to quantify their degree of belief in a hypothesis.<ref>{{cite journal |title=Weight of Evidence: A Brief Survey |last=Good |first=I.J. |year=1985 |journal=Bayesian Statistics |volume=2 |pages=253 |url=http://www.waterboards.ca.gov/water_issues/programs/tmdl/docs/303d_policydocs/207.pdf |accessdate=2012-12-13}}</ref>
The deciban is a particularly useful unit for [[log-odds]], notably as a measure of information in [[Bayes factor]]s, [[odds ratio]]s (ratio of odds, so log is difference of log-odds), or weights of evidence. 10 decibans corresponds to odds of 10:1; 20 decibans to 100:1 odds, etc. According to Good, a change in a weight of evidence of 1 deciban (i.e., a change in the odds from evens to about 5:4) is about as finely as humans can reasonably be expected to quantify their degree of belief in a hypothesis.<ref name="Good_1985"/>


Odds corresponding to integer decibans can often be well-approximated by simple integer ratios; these are collated below. Value to two decimal places, simple approximation (to within about 5%), with more accurate approximation (to within 1%) if simple one is inaccurate:
Odds corresponding to integer decibans can often be well-approximated by simple integer ratios; these are collated below. Value to two decimal places, simple approximation (to within about 5%), with more accurate approximation (to within 1%) if simple one is inaccurate:
Line 51: Line 51:
==See also==
==See also==
* [[bit]]
* [[bit]]
* [[decibel]]


==Notes==
==Notes==
Line 56: Line 57:


==References==
==References==
{{reflist}}
{{reflist|refs=
<ref name="Lukoff_1979">{{cite book |author-last=Lukoff |author-first=Herman |author-link=Herman Lukoff |title=From Dits to Bits: A personal history of the electronic computer |date=1979 |publisher=Robotics Press |location=Portland, Oregon, USA |isbn=0-89661-002-0 |lccn=79-90567}}</ref>
<ref name="IEC">{{cite web |title=IEC 80000-13:2008 |publisher=[[International Organization for Standardization]] (ISO) |url=http://www.iso.org/iso/catalogue_detail?csnumber=31898 |access-date=2013-07-21}}</ref>
<ref name="Hartley_1928">{{cite journal |author-last=Hartley |author-first=Ralph Vinton Lyon |author-link=Ralph Vinton Lyon Hartley |title=Transmission of Information |date=July 1928 |volume=VII |issue=3 |journal=[[Bell System Technical Journal]] |pages=535–563 |url=http://dotrose.com/etext/90_Miscellaneous/transmission_of_information_1928b.pdf |access-date=2008-03-27}}</ref>
<ref name="Reza_1994">{{cite book |author-last=Reza |author-first=Fazlollah M. |author-link=Fazlollah M. Reza |title=An Introduction to Information Theory |location=New York |publisher=[[Dover Publications]] |date=1994 |isbn=0-486-68210-2}}</ref>
<ref name="Good_1979">{{cite journal |author-last=Good |author-first=Irving John |author-link=Irving John Good |title=Studies in the History of Probability and Statistics. XXXVII A. M. Turing's statistical work in World War II |journal=[[Biometrika]] |date=1979 |volume=66 |issue=2 |pages=393–396 |doi=10.1093/biomet/66.2.393 |mr=0548210}}</ref>
<ref name="Gillies_1990">{{cite journal |author-last=Gillies |author-first=Donald A. |author-link=Donald A. Gillies |date=1990 |title=The Turing-Good Weight of Evidence Function and Popper's Measure of the Severity of a Test |periodical=[[British Journal for the Philosophy of Science]] |volume=41 |issue=1 |pages=143–146 |mr=055678 |jstor=688010 |doi=10.1093/bjps/41.1.143}}</ref>
<ref name="Good_1985">{{cite journal |title=Weight of Evidence: A Brief Survey |author-last=Good |author-first=Irving John |author-link=Irving John Good |date=1985 |journal=Bayesian Statistics |volume=2 |pages=253 |url=http://www.waterboards.ca.gov/water_issues/programs/tmdl/docs/303d_policydocs/207.pdf |access-date=2012-12-13}}</ref>
<ref name="Klar_1970">{{cite book |title=Digitale Rechenautomaten – Eine Einführung |language=de |trans-title=Digital Computers – An Introduction |chapter=1.8.1 Begriffe aus der Informationstheorie |trans-chapter=1.8.1 Terms used in information theory |author-first=Rainer |author-last=Klar |publisher=[[Walter de Gruyter & Co.]] / {{ill|G. J. Göschen'sche Verlagsbuchhandlung|de|G. J. Göschen’sche Verlagsbuchhandlung}} |publication-place=Berlin, Germany |series=Sammlung Göschen |volume=1241/1241a |date=1970-02-01 |isbn=3-11-083160-0 |id={{ISBN|978-3-11-083160-3}}. Archiv-Nr. 7990709. |page=35 |edition=1 |url=https://books.google.com/books?id=QnqVDwAAQBAJ&pg=PA35 |access-date=2020-04-13 |url-status=live |archive-url=https://web.archive.org/web/20200418205642/https://books.google.de/books?redir_esc=y&hl=de&id=QnqVDwAAQBAJ&q=dit#v=snippet&q=dit&f=false |archive-date=2020-04-18}} (205 pages) (NB. A 2019 reprint of the first edition is available under {{ISBN|3-11002793-3|978-3-11002793-8}}. A reworked and expanded [[#Klar-1989|4th edition]] exists as well.)</ref>
<ref name="Klar_1989">{{anchor|Klar-1989}}{{cite book |title=Digitale Rechenautomaten – Eine Einführung in die Struktur von Computerhardware |language=de |trans-title=Digital Computers – An Introduction into the structure of computer hardware |chapter=1.9.1 Begriffe aus der Informationstheorie |trans-chapter=1.9.1 Terms used in information theory |author-first=Rainer |author-last=Klar |publisher=[[Walter de Gruyter & Co.]] |publication-place=Berlin, Germany |series=Sammlung Göschen |volume=2050 |date=1989 |orig-year=1988-10-01 |isbn=3-11011700-2 |id={{ISBN|978-3-11011700-4}} |page=57 |edition=4th reworked}} (320 pages)</ref>
}}


[[Category:Units of information]]
[[Category:Units of information]]
[[Category:Logarithmic scales of measurement]]
[[Category:Units of level]]

Latest revision as of 02:29, 16 October 2023

The hartley (symbol Hart), also called a ban, or a dit (short for "decimal digit"),[1][2][3] is a logarithmic unit that measures information or entropy, based on base 10 logarithms and powers of 10. One hartley is the information content of an event if the probability of that event occurring is 110.[4] It is therefore equal to the information contained in one decimal digit (or dit), assuming a priori equiprobability of each possible value. It is named after Ralph Hartley.

If base 2 logarithms and powers of 2 are used instead, then the unit of information is the shannon or bit, which is the information content of an event if the probability of that event occurring is 12. Natural logarithms and powers of e define the nat.

One ban corresponds to ln(10) nat = log2(10) Sh, or approximately 2.303 nat, or 3.322 bit (3.322 Sh).[a] A deciban is one tenth of a ban (or about 0.332 Sh); the name is formed from ban by the SI prefix deci-.

Though there is no associated SI unit, information entropy is part of the International System of Quantities, defined by International Standard IEC 80000-13 of the International Electrotechnical Commission.

History

[edit]

The term hartley is named after Ralph Hartley, who suggested in 1928 to measure information using a logarithmic base equal to the number of distinguishable states in its representation, which would be the base 10 for a decimal digit.[5][6]

The ban and the deciban were invented by Alan Turing with Irving John "Jack" Good in 1940, to measure the amount of information that could be deduced by the codebreakers at Bletchley Park using the Banburismus procedure, towards determining each day's unknown setting of the German naval Enigma cipher machine. The name was inspired by the enormous sheets of card, printed in the town of Banbury about 30 miles away, that were used in the process.[7]

Good argued that the sequential summation of decibans to build up a measure of the weight of evidence in favour of a hypothesis, is essentially Bayesian inference.[7] Donald A. Gillies, however, argued the ban is, in effect, the same as Karl Popper's measure of the severity of a test.[8]

Usage as a unit of odds

[edit]

The deciban is a particularly useful unit for log-odds, notably as a measure of information in Bayes factors, odds ratios (ratio of odds, so log is difference of log-odds), or weights of evidence. 10 decibans corresponds to odds of 10:1; 20 decibans to 100:1 odds, etc. According to Good, a change in a weight of evidence of 1 deciban (i.e., a change in the odds from evens to about 5:4) is about as finely as humans can reasonably be expected to quantify their degree of belief in a hypothesis.[9]

Odds corresponding to integer decibans can often be well-approximated by simple integer ratios; these are collated below. Value to two decimal places, simple approximation (to within about 5%), with more accurate approximation (to within 1%) if simple one is inaccurate:

decibans exact
value
approx.
value
approx.
ratio
accurate
ratio
probability
0 100/10 1 1:1 50%
1 101/10 1.26 5:4 56%
2 102/10 1.58 3:2 8:5 61%
3 103/10 2.00 2:1 67%
4 104/10 2.51 5:2 71.5%
5 105/10 3.16 3:1 19:6, 16:5 76%
6 106/10 3.98 4:1 80%
7 107/10 5.01 5:1 83%
8 108/10 6.31 6:1 19:3, 25:4 86%
9 109/10 7.94 8:1 89%
10 1010/10 10 10:1 91%

See also

[edit]

Notes

[edit]
  1. ^ This value, approximately 103, but slightly less, can be understood simply because : 3 decimal digits are slightly less information than 10 binary digits, so 1 decimal digit is slightly less than 103 binary digits.

References

[edit]
  1. ^ Klar, Rainer (1970-02-01). "1.8.1 Begriffe aus der Informationstheorie" [1.8.1 Terms used in information theory]. Digitale Rechenautomaten – Eine Einführung [Digital Computers – An Introduction]. Sammlung Göschen (in German). Vol. 1241/1241a (1 ed.). Berlin, Germany: Walter de Gruyter & Co. / G. J. Göschen'sche Verlagsbuchhandlung [de]. p. 35. ISBN 3-11-083160-0. ISBN 978-3-11-083160-3. Archiv-Nr. 7990709. Archived from the original on 2020-04-18. Retrieved 2020-04-13. (205 pages) (NB. A 2019 reprint of the first edition is available under ISBN 3-11002793-3, 978-3-11002793-8. A reworked and expanded 4th edition exists as well.)
  2. ^ Klar, Rainer (1989) [1988-10-01]. "1.9.1 Begriffe aus der Informationstheorie" [1.9.1 Terms used in information theory]. Digitale Rechenautomaten – Eine Einführung in die Struktur von Computerhardware [Digital Computers – An Introduction into the structure of computer hardware]. Sammlung Göschen (in German). Vol. 2050 (4th reworked ed.). Berlin, Germany: Walter de Gruyter & Co. p. 57. ISBN 3-11011700-2. ISBN 978-3-11011700-4. (320 pages)
  3. ^ Lukoff, Herman (1979). From Dits to Bits: A personal history of the electronic computer. Portland, Oregon, USA: Robotics Press. ISBN 0-89661-002-0. LCCN 79-90567.
  4. ^ "IEC 80000-13:2008". International Organization for Standardization (ISO). Retrieved 2013-07-21.
  5. ^ Hartley, Ralph Vinton Lyon (July 1928). "Transmission of Information" (PDF). Bell System Technical Journal. VII (3): 535–563. Retrieved 2008-03-27.
  6. ^ Reza, Fazlollah M. (1994). An Introduction to Information Theory. New York: Dover Publications. ISBN 0-486-68210-2.
  7. ^ a b Good, Irving John (1979). "Studies in the History of Probability and Statistics. XXXVII A. M. Turing's statistical work in World War II". Biometrika. 66 (2): 393–396. doi:10.1093/biomet/66.2.393. MR 0548210.
  8. ^ Gillies, Donald A. (1990). "The Turing-Good Weight of Evidence Function and Popper's Measure of the Severity of a Test". British Journal for the Philosophy of Science. 41 (1): 143–146. doi:10.1093/bjps/41.1.143. JSTOR 688010. MR 0055678.
  9. ^ Good, Irving John (1985). "Weight of Evidence: A Brief Survey" (PDF). Bayesian Statistics. 2: 253. Retrieved 2012-12-13.