Skip to content

Smaato - Engineering Manager - Data

RemoteIndia, Mahārāshtra, PuneIT

Job description

What We Do

Smaato’s digital ad tech platform is a completely omnichannel, self-serve monetization solution and ad server. Our controls make monetization simple. Publishers can bring their first-party data and manage all inventory in one place. Marketers get access to the highest-quality inventory so they can reach audiences around the world and on any device. Smaato is headquartered in San Francisco, with additional offices in Hamburg, New York City, Beijing and Singapore. Learn more at www.smaato.com.


Verve Group/Smaato provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state or local laws


The Data Engineering team at Smaato works on a few very interesting technology problems related to big-data, distributed computing, low-latency, analytics, visualization, machine learning, and highly scalable platforms development. We build reliable, peta bytes scale distributed systems using technologies such as Spark, Hadoop, Apache Flink, Airflow, Kafka, and Druid. As part of Smaato, you will work on the applications where all threads come together: Streaming, Processing, Storage the Ad exchange. Our ultra-efficient exchange processes more than 125 billion ad requests daily, e.g. whopping 3.5 trillions in a month. Every line of code you write matters as it is likely to be executed several billion times a day. We are one of the biggest AWS users with a presence in four different regions.

Smaato’s data platform is a symphony of streams, orchestrations, micro services, schedulers, analytics, and visualization technologies. Platform is supported by polyglot persistence using Druid, DynamoDB, Vertica, MySQL, and bunch of orchestration & streaming frameworks like Airflow, Spark, Flink.


The job will involve constant feature development, performance tuning and platform stability assurance. The mission of our analytics team is “data driven decisions at your fingertips”. You own and provide the system that all business decisions will be based on. Precision and high-quality results are essential in this role.


Our engineers are passionate problem solvers. Be it Java, Python, Scala, Groovy, or typescripting, we are up for all games J.

What You’ll Do

  • Design and architect data platform components supporting millions of requests per second.
  • This role is 70% hands on, enhancing and supporting data pipeline components.
  • Lead small team, coordinate with product team, lead automations, review code.
  • Delivering high quality software, meeting timelines, and being agile.
  • Maintain the current platform and shape it’s evolution by adopting new technologies.
  • Closely collaborate with stakeholders to gather requirements and design new product features in a short feedback loop

Job requirements

What We'll Need


  • 12+ years of experience in Big-data platforms, or distributed systems, with a deep understanding of Apache Druid, AWS, Spark, and Airflow.
  • Exposure to highly scalable, low latency microservices development.
  • Strong exposure to application, enterprise, and microservice design patterns.
  • Strong understanding of computer science fundamentals, algorithms, and data structures.
  • Proficient in one or more programming languages – java, python, scala.
  • Exposure to AWS, Automation, and DevOps (Kubernetes, Jenkins CICD pipelines).
  • Experience leading a small team of highly productive engineers.
  • Proven experience in owning the products and driving them to the end, all the way from gathering requirements, development, testing, and deployment to ensuring high availability post-deployment.
  • Contribute to architectural and coding standards, evolving engineering best practices.
  • Nice to have - Open-source committer/contributor / PMC in any of these Apache big data open-source technologies
  • You enjoy operating your applications in production and strive to make on-calls obsolete, debugging issues in their live/production/test clusters and implementing solutions to restore service with minimal disruption to the business. Perform root cause analysis.

The team is looking for passionate engineers, ready to embrace technology to deliver next-generation data platforms.

or