A BCD digit only uses the binary patterns that represent decimal numbers, ie 0000 - 1001; this requires 4 bits (1 nybble) so there can be 2 BCD digits to a byte. Therefore in 3 bytes there can be 3 × 2 = 6 BCD digits. The largest BCD digit is 1001 = 9.
Assuming non-signed, the maximum 3 byte BCD number is 999,999.
It depends what exact system is used - it seems there are several methods to encode numbers as BCD.For example, if "packed" BCD is used (two digits per byte), and one nibble (half-byte) is reserved for the sign, that allows a total of 5 decimal digits.
999
The Largest 4Bytes Hex number is FFFF FFFF which is 65535 in decimal.
If using the compressed format, where a byte holds two decimal digits (because only 4 bits are needed to make nine), so two bytes would be four decimal digits, the largest which is 9999.
255
I would say a monoicosebyte which is an astonishing 1,000,000,000,000,000,000,000,000,000,000,000,000,000 bytes. This unit of storage will most likely never come up due to how large this unit of measurement is for storage. I bet that there isn't even a single device on earth that can hold this many bytes.
A Kilobyte is equal to 1000 bytes
The Largest 4Bytes Hex number is FFFF FFFF which is 65535 in decimal.
If using the compressed format, where a byte holds two decimal digits (because only 4 bits are needed to make nine), so two bytes would be four decimal digits, the largest which is 9999.
255
11b which is 1*2 + 1*1 = 3 would be for two bits. But a byte is 8 bits, so 2 bytes is 16 bits. The largest binary number is [2^16 - 1], which is 65535 (base ten)
1024 bytes is binary counting while 1000 bites is decimal counting.
Tera Bytes are the largest unit of measurement.
The way "gigabyte" is usually used, it means 10243 bytes. In other words, 1,073,741,824 bytes.
A Mac address (Media Access Control address) is 48 bits long, which is equivalent to 6 bytes.
A zettabyte is a massive amount of bytes and referencing from wikipedia (yes it is correct) it is 1,000,000,000,000,000,000,000 bytes in decimal or 1021
GB or gigabytes
Yes. The standard definition is now 10^6 bytes. Historically, it could have represented 1,048,576 bytes (2^20 bytes), a value now defined as a mebibyte (million-binary byte).
Four bytes represent 32 bits. 32 bits represent 4,294,967,296 possibilities.