It is possible to store a pair of 32-bit single precision floating point numbers in the same space that would be taken by a 64-bit double precision number. For example, the XMM registers of the SSE2 instruction set, can store four single precision numbers or two double precision numbers.
By the IEEE 754 standard, the difference between single and double precision is not only the precision per se, but also the available range: 8 and 11 exponent bits respectively.
Intuitively, it seems to me that if you were designing an FPU to process either 2N single precision numbers or N double precision numbers in parallel, the circuit design should be simpler if you deviate from the IEEE standard and make both use the same number of exponent bits. For example, the bfloat16 half precision format, trades away some mantissa bits to keep the same number of exponent bits as single precision; part of the justification given for this, is that it's easier to convert between bfloat16 and single precision.
Do any actual vector instruction sets use the same number of exponent bits for single and double precision? If so, do they stick closer to the 8 bits typical for single precision, or 11 bits typical for double precision?
F
(single precision) andD
(double precision) formats; both used an 8-bit exponent field. However, the small exponent range caused numerical issues for double-precision computation in some contexts, so aG
format (basically IEEE-754 double precision) was added later. – njuffa