It is well documented that variable length string datasets cannot be compressed. That is clear and understandable.
But if one chunks and compresses such a dataset (especially a large one), something is getting compressed. The resultant file size is considerably smaller than a file where the dataset is not chunked and compressed.
So, what is actually getting compressed?