Thanks for taking a look!
Agreed regarding the privacy/security hesitations. Running the models locally with ollama is an option, but of course there's the hardware requirements and limitations of open source models to contend with. ultimately it's a balance between privacy and ease of use, and I'm not sure that there's a good one-size-fits-all for that balance.