Qualifications
Ownership of Model Serving: Design, develop, and sustain low-latency, highly reliable serving stacks for internal ML model serving while integrating with LLM serving partners. Automate Training Pipelines: Coordinate data preparation, training, evaluation, and registry workflows on Kubernetes, adhering to robust MLOps practices. Optimization at Scale: Analyze and enhance throughput, memory, and cost efficiency; implement caching, sharding, batching, and GPU/CPU autoscaling where beneficial. Build Platform Primitives: Develop reusable SDKs, templates, and CLI tools that enable research and product teams to deploy models independently and securely. Elevate Standards: Implement deep observability (tracing, metrics, alerts), conduct blameless post-mortems, and mentor engineers on best practices for production ML.
About the job
Cresta is dedicated to transforming customer conversations into a competitive edge by harnessing the full potential of the contact center. Our innovative platform merges cutting-edge AI with human intelligence, enabling contact centers to uncover valuable customer insights, optimize communication processes, and empower team members to work with greater efficiency. Founded by Sebastian Thrun, a pioneer from the esteemed Stanford AI lab and the visionary behind Google X, Waymo, and Udacity, Cresta's leadership team also includes CEO Ping Wu, co-founder of Google Contact Center AI and Vertex AI platform, and Tim Shi, an early member of OpenAI.
Join us on this exciting journey to reshape the workforce using the power of AI. The future of work is here, and it starts at Cresta.
About Cresta
Cresta is on a mission to revolutionize the customer interaction landscape by integrating the prowess of AI with the nuance of human understanding. Our solutions empower contact centers to maximize their efficiency and effectiveness by leveraging real-time insights and automation, ultimately transforming how businesses engage with their customers.