Hacker News new | past | comments | ask | show | jobs | submit login

I was comparing it with Parquet, which is much more complex, but has features that help you access the data in less than O(n), like row groups and pages.



you mentioned NLJSON and CSV, which would require to read all columns from the disk.


Yes, but you would usually have to read at least two columns anyway. What are the datasets that are too large to be ingested completely, but too small for a proper columnar format?

If ZSV is meant to occupy the gap between CSV/NLJSON (smaller datasets) and Parquet/DuckDB (larger datasets), this niche is actually really small, if not nonexistent.


yes it's unclear to me what is the advantage over parquet with compression. And there are enough file formats flying around already.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: