Consistency LLM: converting LLMs to parallel decoders accelerates inference 3.5x May 8, 2024 by Comments