We use cookies. Find out more about it here. By continuing to browse this site you are agreeing to our use of cookies.
#alert
Back to search results

Software Engineer 2 - CoreAI

Microsoft
United States, Washington, Redmond
Sep 14, 2025
OverviewAt Microsoft, we're a community of passionate innovators driven by curiosity and purpose. We collaborate to imagine what's possible and accelerate our careers in a cloud-powered world where openness and innovation unlock limitless potential. Artificial Intelligence is central to Microsoft's strategy-and the Azure AI Platform is leading the charge. As part of our team, you'll contribute to cutting-edge projects that solve real-world challenges using transformative technologies. We are looking for a Software Engineer 2 - Core AI to join our agile team at the core of Microsoft's AI infrastructure. This team is building the Next Gen Scheduling & Optimization Platform-a foundational infrastructure layer that powers OpenAI models and other large-scale AI workloads across Azure. In this role, you will be responsible for managing inferencing capacity that fuels Microsoft's AI ambitions. Our fleet of premium AI accelerators runs state-of-the-art OpenAI models, forming the backbone of Microsoft's Copilots and the Azure OpenAI Service. You'll help dynamically allocate resources across models and customer offerings, monitor usage in near real-time, and rebalance capacity to drive massive efficiency gains. You'll work on high-impact distributed systems that support low-latency, high-volume, mission-critical customer scenarios, solving complex challenges in resource orchestration, telemetry, and performance optimization. You'll collaborate across Azure, OpenAI, CoreAI Services, and infrastructure teams to shape the future of scalable, cost-efficient AI.
ResponsibilitiesDesign and implement scalable services for GPU scheduling, allocation, and optimization across diverse AI workloads. Build reliable orchestrations to monitor GPU usage near real time and drive automated rebalancing decisions. Integrate with fleet health dashboards and GPU lifecycle management systems to ensure reliability and performance. Collaborate with partner teams across Azure ML, AOAI, and Core AI to align architecture, APIs, and operational readiness. Contribute to platform evolution supporting new hardware and real-time inference APIs.
Applied = 0

(web-759df7d4f5-j8zzc)