Their working is exceptionally thorough, all the concerns about additional maths and transformations is well considered. I really have to wonder if introducing any semantic-changing transformations for what I’m guessing are minuscule savings of bytes is worth it though. Safe things like stripping white space and shortening identifiers would be the major contributors to minification surely. How many CSS colours need to be defined for oklab rounding to even affect how many tcp packets are sent for instance?
While it’s hardly insightful that SQL is useful, I would have liked to read more about what the actual workload involving duckdb on a local machine looked like. I’m fully on board that local or single vm workloads can do an awful lot, but I’ve never been particularly satisfied with the pipelines I’ve seen (including my own). Usually they’re piles of scripts and intermediate data files sitting around and are hard to make idempotent and understand if you aren’t the author.
Also fwiw there’s no such thing as an M4 Ultra chip. That detail was either a mistake or hallucinated.