Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

In what situations is octal base used?

People also ask

What is the octal system used for?

What are the Uses of Octal Numbers? The Octal Number system is widely used in computer application sectors and digital numbering systems. The computing systems use 16-bit, 32-bit or 64-bit word which is further divided into 8-bits words. The octal number is also used in the aviation sector in the form of a code.

Is octal ever used?

Octal became widely used in computing when systems such as the UNIVAC 1050, PDP-8, ICL 1900 and IBM mainframes employed 6-bit, 12-bit, 24-bit or 36-bit words. Octal was an ideal abbreviation of binary for these machines because their word size is divisible by three (each octal digit represents three binary digits).

Where are octal and hexadecimal numbers used?

The octal and hexadecimal number systems are introduced to interpret large binary numbers in computer languages. Ans. Decimal and hexadecimal are number systems, which have a base of 10 and 16, respectively.


Octal is used when the number of bits in one word is a multiple of 3. Examples are ancient systems with 18bit word sizes, systems with 9bit bytes, or unix file permissions with 9bits or 12bits.

If the number of bits in your word is a multiple of 4, however, please do use hex, by all means.


Octal is used as a shorthand for representing file permissions on UNIX systems. For example, file mode rwxr-xr-x would be 0755.


Octal is used when the syntax is a relic from the ages when it perhaps made sense on some platform (system words haven't always been a multiple of 8 bits). Nowadays hex is the thing to use.


Didn't think of this but Digital displays!

Several other uses from: http://en.wikipedia.org/wiki/Octal


One of the main reasons octal used to be more frequently used was that it is easier to convert between octal and binary in your head than hex to binary: you only have to remember the binary representation of the 8 octal digits (0-7).

Back in the days when debugging meant reading register contents from a row of LEDs, or entering data with an array of toggle switches, this was a big concern. The panels on many of these early computers grouped the LEDs and switches in groups of threes to facilitate this.

However, hex began to win out as word sizes that are multiples of 8-bit bytes began to win out, and the need to read and enter data in binary became unecessary (with console text UI and later GUI debuggers).