Sr. Deployment Engineer, AI Inference - Toronto
1 day ago

Job summary
Cerebras Systems builds the world's largest AI chip which is 56 times larger than GPUs.
Responsibilities
- Deploy AI inference replicas and cluster software across multiple datacenters
- Operate across heterogeneous datacenter environments undergoing rapid 10x growth
- Maximize capacity allocation and optimize replica placement using constraint-solver algorithms
Job description
, consectetur adipiscing elit. Nullam tempor vestibulum ex, eget consequat quam pellentesque vel. Etiam congue sed elit nec elementum. Morbi diam metus, rutrum id eleifend ac, porta in lectus. Sed scelerisque a augue et ornare.
Donec lacinia nisi nec odio ultricies imperdiet.
Morbi a dolor dignissim, tristique enim et, semper lacus. Morbi laoreet sollicitudin justo eget eleifend. Donec felis augue, accumsan in dapibus a, mattis sed ligula.
Vestibulum at aliquet erat. Curabitur rhoncus urna vitae quam suscipit
, at pulvinar turpis lacinia. Mauris magna sem, dignissim finibus fermentum ac, placerat at ex. Pellentesque aliquet, lorem pulvinar mollis ornare, orci turpis fermentum urna, non ullamcorper ligula enim a ante. Duis dolor est, consectetur ut sapien lacinia, tempor condimentum purus.
Access all high-level positions and get the job of your dreams.
Similar jobs
AI Inference Engineer
4 weeks ago
We are looking for highly skilled engineers with a focus on C/C++, low level systems, performance and power optimization to join our team full-time. · Innovate on the inference optimization pipeline through algorithmic and system optimization · Own end to end system characterizat ...
AI Inference Engineer
4 weeks ago
We are building the Inference infrastructure for Edge, where we are unlocking data center ai inference capability closer to where the data lives. · We are looking for highly skilled engineers with a focus on C/C++, low level systems, performance and power optimization to join our ...
Audio Inference Engineer, Model Efficiency
1 month ago
We re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation semantic search RAG and agents. · We obsess over what we build Each one of us is responsible for contributing to increasing ...
+We're training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. · + · +Developing, deploying, and operating the AI platform delivering Cohere's large l ...
We're looking for a Site Reliability Engineer to join our Model Serving team. The team develops, deploys, and operates AI platforms delivering large language models through easy-to-use API endpoints. · You will build self-service systems that automate managing, deploying, and ope ...
We're training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. · Build self-service systems that automate managing, deploying and operating services. · ...
Full-Stack Software Engineer, Inference
4 weeks ago
We're training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, · semantic search, · RAG, · and agents.Cohere customers self-serve our API without any intervention. · ...
Full-Stack Software Engineer, Inference
4 weeks ago
Cohere is training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, · and agents. · We believe that our work is instrumental to the widespread adoption of AI. · We ...
Sr. Inference ML Runtime Engineer
1 hour ago
+Job summary · Cerebras Systems builds the world's largest AI chip, · +Drive and provide technical guidance to a team of software engineers working on complex machine learning integration projects. · Design and implement ML features (e.g., structured outputs, biased sampling, pre ...
Senior Software Engineer, AI Inference Systems
3 weeks ago
We are seeking highly skilled and motivated software engineers to join us and build AI inference systems that serve large-scale models with extreme efficiency. · Contribute features to vLLM that empower the newest models with the latest NVIDIA GPU hardware features; profile and o ...
+We're training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, · and agents.We believe that our work is instrumental to the widespread adoption of AI. · +In this ...
We're training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI. · ...
We're training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. · Build self-service systems that automate managing, deploying and operating services. · ...
Full-Stack Software Engineer, Inference
1 month ago
We're training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, · and agents.We believe that our work is instrumental to the widespread adoption of AI. · Improve th ...
Full-Stack Software Engineer, Inference
1 month ago
We are training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. · ...
Senior Software Engineer, AI Inference Systems
4 weeks ago
We are seeking highly skilled software engineers to build AI inference systems. You'll architect and implement high-performance stacks, optimize GPU kernels and compilers, drive industry benchmarks, and scale workloads across environments. · You'll collaborate with teams to push ...
We are seeking an exceptional Neural Rendering Research Inference Engineer, Advanced Graphics Program who has deep technical expertise in translating neural network models and algorithms to efficient inference solutions. · ...
We are seeking an exceptional Neural Rendering Research Inference Engineer, Advanced Graphics Program who has deep technical expertise in translating neural network models and algorithms to efficient inference solutions. · AMD is looking for a strategic research inference enginee ...
Software Engineer
4 weeks ago
+Job summary · We're hiring Software Engineers to support and evolve the API gateway to our inference cloud. The API Gateway is constantly evolving in response to the unending stream of new models, modalities, clients and inference load. · +Responsibilitiesundertake feature devel ...
Senior Product Manager – Inference
1 month ago
We are looking for a Senior Product Manager to join our growing Product Management team. · This is a challenging position that sits at the critical intersection of our predictive AI engine and internal product lines. · You will act as the custodian of our core inference and predi ...