Senior Software Engineer in AI and Machine Learning
Oracle
About the role
About Us
Oracle integrates data, infrastructure, applications, and expertise to drive industry innovations and enhance lives. By embedding AI into our products and services, we empower customers to craft a promising future. Join us at the forefront of transforming AI and cloud solutions.
We are committed to fostering a diverse workforce and promoting opportunities for all while providing a comprehensive benefits package to our employees. We encourage community engagement through volunteer programs and are dedicated to ensuring access for candidates with disabilities throughout the recruitment process.
Oracle is an Equal Employment Opportunity Employer. Qualified applicants will receive fair consideration for employment without regard to various protected characteristics in accordance with the law.
Responsibilities
- Evaluate and enhance advanced technologies to optimize performance for training and inference tasks.
- Lead crucial discussions that shape AI infrastructure offerings.
- Design and develop scalable orchestration solutions for efficient AI/ML model serving and training, emphasizing model parallelism.
- Incorporate the latest research on generative AI and inference systems into our software stack for large language models.
- Drive initiatives for Generative AI systems, including Retrieval-Augmented Generation (RAG) and fine-tuning of large language models.
- Create robust services and tools supporting GPU-accelerated AI pipelines leveraging Kubernetes, Python/Go, and observability frameworks.
Qualifications
- Bachelor's, Master's, or Ph.D. in Computer Science, Engineering, Machine Learning, or related field (or equivalent experience).
- Strong foundation in Machine Learning and Deep Learning concepts, algorithms, and models.
- Proficient in orchestration and containerization technologies, especially Kubernetes and Docker.
- Deep understanding of container networking and storage architecture.
- Experience in managing and optimizing large-scale distributed training and inference workloads.
- Thorough knowledge of AI/ML workflows, including data processing, model training, and inference pipelines.
- Familiarity with parallel computing frameworks and paradigms.
- Proficient programming skills and extensive experience with major deep learning frameworks.
Disclaimer
Certain US customer or client-facing roles may be required to comply with applicable requirements, such as immunization and occupational health mandates.
Salary & Compensation
The salary range for this position in the US is from $96,800 to $251,600 per annum, with additional eligibility for bonuses, equity, and compensation deferrals. Oracle maintains a range of salaries for its roles to reflect various factors, including knowledge, skills, experience, market conditions, and internal equity.
Benefits
Oracle US offers an extensive benefits package that includes:
- Medical, dental, and vision insurance
- Disability coverage
- Life insurance
- Flexible spending accounts
- 401(k) plans with company match
- Paid time off
- Additional voluntary benefits
Requirements
- Strong foundation in Machine Learning and Deep Learning concepts, algorithms, and models.
- Proficient in orchestration and containerization technologies, especially Kubernetes and Docker.
- Deep understanding of container networking and storage architecture.
- Experience in managing and optimizing large-scale distributed training and inference workloads.
- Thorough knowledge of AI/ML workflows, including data processing, model training, and inference pipelines.
- Familiarity with parallel computing frameworks and paradigms.
- Proficient programming skills and extensive experience with major deep learning frameworks.
Responsibilities
- Evaluate and enhance advanced technologies to optimize performance for training and inference tasks.
- Lead crucial discussions that shape AI infrastructure offerings.
- Design and develop scalable orchestration solutions for efficient AI/ML model serving and training, emphasizing model parallelism.
- Incorporate the latest research on generative AI and inference systems into our software stack for large language models.
- Drive initiatives for Generative AI systems, including Retrieval-Augmented Generation (RAG) and fine-tuning of large language models.
- Create robust services and tools supporting GPU-accelerated AI pipelines leveraging Kubernetes, Python/Go, and observability frameworks.
Benefits
Skills
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free