I does not look like it support streaming of responses from llm into channel. Big issue for local inferrence.