I don't understand what is:
-32 bit cheat
Here a simple Explanation:
How it looks in Memory:
u/s 08 Bit Cheat = { [00] [00] [00] [00] [00] [00] [00] [00] } = {1 Byte}
u/s 16 Bit Cheat = { [00 00] [00 00] [00 00] [00 00] } = {1 Byte}
u/s 32 Bit Cheat = { [00 00 00 00] [00 00 00 00] } = {1 Byte}
u/s 64 Bit Cheat = { [00 00 00 00 00 00 00 00] } = {1 Byte}
u/s 08 Bit Cheat = { [01] [02] [03] [04] [05] [06] [07] [08] } = {1 Byte}
u/s 16 Bit Cheat = { [02 01] [04 03] [06 05] [08 07] } = {1 Byte}
u/s 32 Bit Cheat = { [04 03 02 01] [08 07 06 05] } = {1 Byte}
u/s 64 Bit Cheat = { [80 70 60 50 40 30 20 10] } = {1 Byte}
Which Numbers each Data Type can hold:
unsigned = Positive Numbers
u08 Hex = 0 to FF
u08 Dec = 0 to 255
u16 Hex = 0 to FFFF
u16 Dec = 0 to 65’535
u32 Hex = 0 to FFFFFFFF
u32 Dec = 0 to 4’294’967’295
u64 Hex = 0 to FFFFFFFFFFFFFFFF
u64 Dec = 0 to 18'446'744'073'709'551'615
signed = Positive & Negative Numbers
s08 Hex = FF80 to 7F
s08 Dec = -128 to 127
s16 Hex = 8000 to 7FFF
s16 Dec = -32'768 to 32'767
s32 Hex = 80000000 to 7FFFFFFF
s32 Dec = -2'147'483'648 to 2'147'483'647
s64 Hex = 8000000000000000 to 7FFFFFFFFFFFFFFF
s64 Dec = -9'223'372'036'854'775'808 to 9'223'372'036'854'775'807
Float = 1.2E-38 to 3.4E+38
(-120 followed by another 36 zeros then a decimal Point to
340 followed by another 36 zeros then a decimal Point)
Double = 2.3E-308 to 1.7E+308
(-230 followed by another 306 zeros then a decimal Point to
170 followed by another 306 zeros then a decimal Point)
Float Precision up to 6 Digits after decimal Point
Double Precision up to 15 Digits after decimal Point
u/s 08 Bit Cheat = { [00] [00] [00] [00] [00] [00] [00] [00] } = {1 Byte}
u/s 16 Bit Cheat = { [00 00] [00 00] [00 00] [00 00] } = {1 Byte}
u/s 32 Bit Cheat = { [00 00 00 00] [00 00 00 00] } = {1 Byte}
u/s 64 Bit Cheat = { [00 00 00 00 00 00 00 00] } = {1 Byte}
u/s 08 Bit Cheat = { [01] [02] [03] [04] [05] [06] [07] [08] } = {1 Byte}
u/s 16 Bit Cheat = { [02 01] [04 03] [06 05] [08 07] } = {1 Byte}
u/s 32 Bit Cheat = { [04 03 02 01] [08 07 06 05] } = {1 Byte}
u/s 64 Bit Cheat = { [80 70 60 50 40 30 20 10] } = {1 Byte}
Which Numbers each Data Type can hold:
unsigned = Positive Numbers
u08 Hex = 0 to FF
u08 Dec = 0 to 255
u16 Hex = 0 to FFFF
u16 Dec = 0 to 65’535
u32 Hex = 0 to FFFFFFFF
u32 Dec = 0 to 4’294’967’295
u64 Hex = 0 to FFFFFFFFFFFFFFFF
u64 Dec = 0 to 18'446'744'073'709'551'615
signed = Positive & Negative Numbers
s08 Hex = FF80 to 7F
s08 Dec = -128 to 127
s16 Hex = 8000 to 7FFF
s16 Dec = -32'768 to 32'767
s32 Hex = 80000000 to 7FFFFFFF
s32 Dec = -2'147'483'648 to 2'147'483'647
s64 Hex = 8000000000000000 to 7FFFFFFFFFFFFFFF
s64 Dec = -9'223'372'036'854'775'808 to 9'223'372'036'854'775'807
Float = 1.2E-38 to 3.4E+38
(-120 followed by another 36 zeros then a decimal Point to
340 followed by another 36 zeros then a decimal Point)
Double = 2.3E-308 to 1.7E+308
(-230 followed by another 306 zeros then a decimal Point to
170 followed by another 306 zeros then a decimal Point)
Float Precision up to 6 Digits after decimal Point
Double Precision up to 15 Digits after decimal Point
@TomSwitch please correct me if i'm wrong.
Last edited by crazy_p,