Building an AI Inference Gateway, Part 2: Unifying Streaming and Non-Streaming Processing
# Building an AI Inference Gateway, Part 2: Unifying Streaming and Non-Streaming Processing
Recently, while developing [llm-converter](https://github.com/putao520/llm-converter)—a Rust library for LLM protocol conversion—I encountered a problem: **h...
putao282.com5 min read