bit

n.

[from the mainstream meaning and “Binary digIT”]

  1. [techspeak] The unit of information; the amount of information obtained from knowing the answer to a yes-or-no question for which the two outcomes are equally probable.

  2. [techspeak] A computational quantity that can take on one of two values, such as true and false or 0 and 1.

  3. A mental flag: a reminder that something should be done eventually. “I have a bit set for you.” (I haven’t seen you for a while, and I’m supposed to tell or ask you something.)

  4. More generally, a (possibly incorrect) mental state of belief. “I have a bit set that says that you were the last guy to hack on EMACS.” (Meaning “I think you were the last guy to hack on EMACS, and what I am about to say is predicated on this, so please stop me if this isn’t true.”) “I just need one bit from you” is a polite way of indicating that you intend only a short interruption for a question that can presumably be answered yes or no.

A bit is said to be set if its value is true or 1, and reset or clear if its value is false or 0. One speaks of setting and clearing bits. To toggle or invert a bit is to change it, either from 0 to 1 or from 1 to 0. See also flag, trit, mode bit.

The term bit first appeared in print in the computer-science sense in a 1948 paper by information theorist Claude Shannon, and was there credited to the early computer scientist John Tukey (who also seems to have coined the term software). Tukey records that bit evolved over a lunch table as a handier alternative to bigit or binit, at a conference in the winter of 1943-44.