Together AI is a research-driven AI cloud infrastructure provider offering a purpose-built GPU platform for training and running advanced AI models. Serving leading SaaS companies and pioneering startups, Together AI champions open source AI and decentralized computing, advocating for transparency to drive innovation and societal benefits.
Together.ai is building infrastructure to enable efficient and scalable inference for large language models (LLMs). They are seeking an Inference Frameworks and Optimization Engineer to design, develop, and optimize distributed inference engines for high-performance serving. The role focuses on low-latency, high-throughput inference, GPU/accelerator optimizations, and software-hardware co-design.
Together AI is a research-driven AI cloud infrastructure provider offering a purpose-built GPU platform for training and running advanced AI models. Serving leading SaaS companies and pioneering startups, Together AI champions open source AI and decentralized computing, advocating for transparency to drive innovation and societal benefits.