Novita AI, a number one world synthetic intelligence (AI) cloud platform, is proud to announce a strategic partnership with SGLang, a quick serving engine for giant language fashions and imaginative and prescient language fashions. By this collaboration, Novita AI will present high-performance GPU cloud sources for SGLang’s ongoing analysis, benchmarking, and optimization efforts.
Additionally Learn: Why multimodal AI is taking up communication
SGLang is a number one inference engine that co-designs a structured technology language with a extremely optimized runtime, enabling highly effective efficiency good points reminiscent of environment friendly RadixAttention cache reuse and zero-overhead batch scheduling for giant language and vision-language fashions. By aligning language-level management with backend optimizations, it empowers builders to construct advanced technology workflows, multi-modal purposes, and parallel inference pipelines with reliability and scale. SGLang is supported by main establishments together with NVIDIA, AMD, xAI, Oracle Cloud, Google Cloud, LinkedIn, Cursor, alongside analysis teams at Stanford, College of California, Berkeley, and College of California, Los Angeles—proof of robust group engagement and broad trade adoption.
“SGLang’s integration of language-level primitives with runtime optimizations demonstrates the worth of aligning software program and {hardware} to unlock new efficiency ranges,” mentioned Junyu Huang, Co-Founder & COO at Novita AI. “By contributing our infrastructure and experience, we’ve already supported the event of SGLang’s first end-to-end multi-turn reinforcement studying (RL) framework and the Prism multi-large language mannequin serving system, and stay dedicated to fueling its ongoing improvements for builders all over the place.”
Additionally Learn: AiThority Interview with Nicole Janssen, Co-Founder and Co-CEO of AltaML
“We’re thrilled to associate with the SGLang group,” added Junyu Huang. “Having supported their RL framework and multi-LLM serving system, we’re excited to see these achievements speed up their work and convey highly effective inference efficiency to purposes throughout industries.”
Novita AI can also be collaborating on SGLang’s large-scale professional parallelism venture, an open-source implementation designed to method the throughput benchmarks detailed within the official DeepSeek weblog, partnering to convey this milestone to fruition.
This collaboration displays Novita AI’s ongoing dedication to advancing an open ecosystem of inference engines and supporting various analysis initiatives by shared infrastructure and joint improvement efforts.
By collaborations with pioneering open-source tasks like SGLang, Novita AI continues to advance its mission of democratizing AI, making cutting-edge inference capabilities available to builders worldwide.
[To share your insights with us, please write to psen@itechseries.com]