logoalt Hacker News

andai01/21/20252 repliesview on HN

Could you elaborate?


Replies

XMasterrrr01/21/2025

I think I know what he means. I use AI Chat. I load Qwen2.5-1.5B-Instruct with llama.cpp server, fully offloaded to the CPU, and then I config AI Chat to connect to the llama.cpp endpoint.

Checkout the demo they have below

https://github.com/sigoden/aichat#shell-assistant

iamnotagenius01/22/2025

I just run llama-cli with the model. Every time I want some "awk" or "find" trickery, I just ask model. Good for throwaway python scripts too.

show 1 reply