As far as I understand a coded character set maps/assigns numbers (called code points) to an (abstract) character (eg. the german character ü to the code point 00FC in unicode).
This code point can be encoded (e.g. represented in byte patterns) in different ways:
UTF-8 (1 Byte), UTF-16 (2 Bytes) and UTF-32 (4 Bytes)
So the process is:
(abstract) characters ---> maps to code points (coded) character set --> encoding of code points to 1...n bytes
Why this intermediate stage of code points? Why are (abstract) characters not directly mapped to 1...n bytes?
(abstract) characters --> maps to 1...n bytes
This intermediate stage (assign numbers to characters) is also done in other (encoded) character sets. So there must be good reasons for it.
I want to understand why no direct mapping to bytes is done and if there are character sets which doesn't have this intermediate stage and directly maps to bytes.
Thanks in advance...