I have a byte array which I believe correctly stores a UTF-16 encoded Surrogate Pair for the unicode character ????
Running that byte array through .Net System.Text.Encoding.Unicode.GetString()
returns non-expected results.
Actual results: ��
Expected results: ????
Code example:
byte[] inputByteArray = new byte[4];
inputByteArray[0] = 0x91;
inputByteArray[1] = 0xDF;
inputByteArray[2] = 0x00;
inputByteArray[3] = 0xD8;
// System.Text.Encoding.Unicode accepts little endian UTF-16
// Least significant byte first within the byte array [0] MSByete in [3]
string str = System.Text.Encoding.Unicode.GetString(inputByteArray);
// This returns �� rather than the excpected symbol: ????
Console.WriteLine(str);
Detail on how I got to that particular byte array from the character : ????
This character is within the Supplementary Multilingual Plane. This character in Unicode is 0x10391. Encoded into a UTF-16 surrogate pair, this should be :
Minus the Unicode value with 0x10000 : val = 0x00391 = (0x10391 - 0x10000)
High surrogate: 0xD800 = ( 0xD800 + (0x00391 >> 10 ))
top 10 bits
Low surrogate: 0xDF91 = (0xDC00 + (0x00391 & 0b_0011_1111_1111))
bottom 10 bits