Together AI is a research-driven AI cloud infrastructure provider offering a purpose-built GPU platform for training and running advanced AI models. Serving leading SaaS companies and pioneering startups, Together AI champions open source AI and decentralized computing, advocating for transparency to drive innovation and societal benefits.
Together AI is building the Inference Platform to power generative AI models. The role involves optimizing request routing, scaling systems, designing trade-offs between latency and throughput, and engineering features for efficient inference. The team works with LLMs, multimodal models, and image processing in a distributed environment, collaborating with research and the open-source community.
Together AI is a research-driven AI cloud infrastructure provider offering a purpose-built GPU platform for training and running advanced AI models. Serving leading SaaS companies and pioneering startups, Together AI champions open source AI and decentralized computing, advocating for transparency to drive innovation and societal benefits.