arxiv preprint - Speculative Streaming: Fast LLM Inference without Auxiliary Models | AI Breakdown | Podwise