logoalt Hacker News

throwa356262yesterday at 8:35 PM4 repliesview on HN

"LLM backends: Anthropic, OpenAI, OpenRouter."

And here I was hoping that this was local inference :)


Replies

micwyesterday at 8:51 PM

Sure. Why purchase a H200 if you can go with an ESP32 ^^

show 1 reply
__tnmyesterday at 9:29 PM

haha well I got something ridiculous coming soon for zclaw that will kinda work on board.. will require the S3 variant tho, needs a little more memory. Training it later today.

peteriszayesterday at 9:06 PM

right, 888 kB would be impossible for local inference

however, it is really not that impressive for just a client

show 1 reply