I built an in-browser data explorer for datasets too large for spreadsheets but too small to justify spinning up SQL. It runs entirely client-side using DuckDB-WASM, no server uploads accounts or anything.
The core idea is a visual DAG where each transformation (filter, join, aggregate, pivot) creates a view node. Nothing materializes until you need it, DuckDB executes the full chain on demand so you can build deep pipelines without copying data at each step.
Input files can be CSV/Parquet/Excel (Excel might not work great). There's a SQL editor with schema-aware autocomplete, pivot tables with drill-down to underlying rows, and sessions can be exported as files or shareable URLs (the entire pipeline gets encoded in the hash).
Sharing can be granular and you can choose not to embed the files or if files are too big they will not be embedded and the user when opening the link will have to upload the files to restore the session.
The part I find most useful: you can replay pipelines on new data. Share your January analysis, and a colleague runs the same transformations on February's data with schema validation.
Privacy-first since files never leave your browser, it's a static website actually. I will open source soon, and make it probably MIT licensed.
Also it's a WIP and so it may be buggy (there's not even images on the homepage) https://repere.ai