M

DevOps Engineer - AWS

Millennium
Full-time
On-site
London, United Kingdom
DevOps Engineer - AWS

Job Description

Millennium is a top tier global hedge fund with a strong commitment to leveraging innovations in technology and data science to solve complex problems for the business. Β We are assembling a strong DevOps Engineering team to build next generation in-house platforms for several business lines.Β  Our technology teams provide a dynamic and fast-paced environment with excellent growth opportunities.

ResponsibilitiesΒ 

  • Work closely with quants, portfolio manager, risk managers and other engineers in New York, Chicago, Miami, Singapore, and Bangalore to develop data intensive and multi-asset analytics for our Fixed Income, Risk, and Commodities platforms
  • Gather requirements and user feedback in collaboration with fellow engineers and project leads
  • Design, build and refactor robust software applications with clean and concise code following Agile and continuous delivery practices
  • Automation of system maintenance tasks, end-of-day processing jobs, data integrity checks and bulk data loads/extracts
  • Staying abreast of industry trends, new platforms and tools, and coming up with a business case to adopt new technologies
  • Develop new tools and infrastructure using Python (Flask/Fast API) or Java (Spring Boot) and relational data backend (AWS – Aurora/Redshift/Athena/S3)
  • Support users and operational flows for quantitative risk, senior management and portfolio management teams using the tools developed

Qualifications/Skills Required

  • Advance degree in computer science or any other scientific fields
  • 5+ years of experience in CI/CD Tools like TeamCity, Jenkins, Octopus Deploy and ArgoCD.
  • AWS Cloud infrastructure design, implementation and support. Experience with multiple AWS Services.
  • Infrastructure as Code deploying cloud infrastructure using Terraform or Cloudformation.
  • Knowledge of Python (Flask/FastAPI/Django)
  • Demonstrated expertise in the process of containerization for applications and their subsequent orchestration within Kubernetes environments.
  • Experience working on at least one monitoring/observability stack (Datadog, ELK, Splunk, Loki, Grafana).
  • Strong knowledge of Unix or Linux
  • Strong communication skills to collaborate with various stakeholders
  • Able to work independently in a fast-paced environment
  • Detail oriented, organized, demonstrating thoroughness and strong ownership of work
  • Experience working in a production environment.
  • Some experience with relational and non-relational databases.

Qualifications/Nice to have

  • Experience with a messaging middleware platform like Solace, Kafka or RabbitMQ.
  • GCP experience nice to have
  • Experience with Snowflake and distributed processing technologies (e.g., Hadoop, Flink, Spark)