Anyone tried using this as a sub-agent for a more capable model like Claude/Codex?
I was thinking about the other way: Could you use this in front of Claude to summarize inputs and so reduce your token counts?
If you’re looking into small models for tiny local tasks, you should try Qwen coder 0,5B. It’s more of an experiment, but it can output decent functions given the right context instructions.
project started with
trying to run openclaw with it in ultra token saving mode, did totally not work.
great for shell scripts though (my major use case now)
The combined (input/output) context window length is 4K. Claude would blow through that even when trying to read and summarize a small file.