Lead Data Engineer - Kafka

Asapp-2 in Bangalore

At ASAPP, we are on a mission to build transformative machine learning-powered products that push the boundaries of artificial intelligence and customer experience. We focus on solving complex, data-rich problems — the kind where there are huge systemic inefficiencies and where a real solution will have a significant economic impact. Our CX performance platform uses machine learning across both voice and digital engagement channels to augment and automate human work, radically increasing productivity and improving the efficiency and effectiveness of customer experience teams.
The Data Engineering team at ASAPP designs, builds and maintains our mission-critical core data infrastructure and analytics platform. Accurate, easy-to-access, and secure data is critical to our natural language processing (NLP) customer interaction platform which interacts with tens of millions of end-users in real-time.
We’re looking to hire a Lead Data Engineer with the knack for building out data infrastructure systems that can handle our ever-growing volumes of data and the demands we want to make of it. Automation is a key part of our workflow, so you’ll help design and build highly-available data processing pipelines that self-monitor and report anomalies. As part of our fast-growing data engineering team, this role will also play an integral role in shaping the future of data infrastructure as it applies to improving our existing metric-driven development and machine learning capabilities.
ASAPP is committed to creating a diverse environment and is proud to be an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, national origin, disability, age, or veteran status. If you have a disability and need assistance with our employment application process, please email us at careers@asapp.com to obtain assistance. #LI-Hybrid #LI-PR1
    • Manage large scale multi-nodes Kafka cluster environments residing on AWS
    • Handle all Kafka environment upgrades, including capacity planning, cluster setup, performance tuning and ongoing monitoring
    • Perform high-level, day-to-day operational maintenance, support for the Kafka Cluster
    • Recognize data flow patterns and generalizations to automate as much as possible to drive productivity gains
    • Expand our logging and monitoring processes to discover and resolve anomalies and issues before they become problems
    • Develop state-of-the-art automation and data solutions in Python, Spark and Flink
    • Maintain, Manage, Monitor our infrastructure related including Kafka, Kubernetes, Spark, Flink, Jenkins, general OLAP and RDBMS databases, S3 objects buckets, permissions
    • Know how to make the tradeoffs required to ship without compromising quality
    • 8+ years of experience in general software development and/or dev-ops, sre roles.
    • 5+ years experience in data engineering, data systems, pipeline and stream processing.
    • Expertise in at least one flavor of SQL, e.g. Redshift, Postgres, MySQL, Presto, Spark SQL, Hive
    • Proficiency in a high-level programming language. We use Python, Scala, Java, Kotlin, and Go
    • Experience with CI/CD (continuous integration and deployment)
    • Experience with workflow management systems such as Airflow, Oozie, Luigi, and Azkaban
    • Confidence operating in a devops-like capacity working with AWS, Kubernetes, Jenkins, Terraform, etc. thinking about automation, alerting, monitoring, and security and other declarative infrastructure
    • Bachelor’s Degree in a field of science, technology, engineering, or math, or equivalent hands-on experience
    • Experience in maintaining and managing kafka (not just using). Kafka Admin experience is a huge plus
    • Experience in maintaining and managing OLAP/HA database systems (not just using)
    • Familiarity handling Kubernetes clusters for various jobs, apps, and high throughput
    • Technical knowledge of data exchange and serialization formats such as Protobuf, Avro, or Thrift
    • Experience in either deploying and creating Spark Scala and/or Flink applications
    • Competitive compensation
    • Stock options
    • ICICI Lombard General Insurance LTD
    • Onsite lunch & dinner stipend
    • Connectivity (mobile phone & internet) stipend
    • Wellness perks
    • Mac equipment
    • Learning & development stipend
    • Parental leave, including 6 weeks paternity leave
Apply