I imagine this is a question that has been asked a thousand times but I can't find a straight answer.
I'm trying to port a very short, C function (initially written for Windows) to OSX, using XCode.
I've created an XCode, Terminal application project and I've copied and pasted the code into the project. XCode is giving me a bunch of errors telling me that the type uint32
and uint64
are unknown types.
What are the equivalent value types in OSX?
In C#, UInt32 struct is used to represent 32-bit unsigned integers(also termed as uint data type) starting from range 0 to 4,294,967,295.
The UInt64 value type represents unsigned integers with values ranging from 0 to 18,446,744,073,709,551,615. Important. The UInt64 type is not CLS-compliant. The CLS-compliant alternative type is Decimal. Int64 can be used instead to replace a UInt64 value that ranges from zero to MaxValue.
The range for the UInt64 type is from 0 through 2^64-1 . The size of UInt64 is 64 bits across all 64-bit and 32-bit platforms.
You can use uint32_t
and uint64_t
by importing stdint.h
.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With