• Data Engineer

    Location IN-KA-Bangalore
    Posted Date 5 months ago(01-03-2018 9:43)
    Job ID
    ADCI - Karnataka
  • Job Description

    Amazon Music is awash in data! To help make sense of it all, the Data Authority and Warehouse team enables repeatable, easy, in depth analysis of music customer behaviors. We reduce the cost in time and effort of analysis, data set building, model building, and user segmentation. Our goal is to empower all teams at Amazon to make data driven decisions and effectively measure their results by providing high quality, highly available data, and democratizing data access through self-service tools.

    If you love the challenges that come with big data then this role is for you. We collect billions of events a day, manage hundreds of terabytes of data on Redshift, Oracle, and S3, and develop data pipelines using SQL based ETL, Java/Scala Spark-EMR, and Java services. And we are just getting started!

    You are a talented, enthusiastic, and detail-oriented Data Engineer who wants to take on big challenges in an agile way. Duties include data modeling and information design, ETL on Redshift using SQL, ETL using python/scala/java MR jobs on Spark/EMR, fact and dimension development and maintenance, and Tableau setup, administration, and development. With the launch of Amazon Music Unlimited and the continued growth of Amazon Prime Music, providing high quality data to our internal customers has a direct impact on the quality of our customer experience.

    We focus our development on core platform features our internal customers need and deal in AWS technologies like Redshift, S3, EMR, EC2, DynamoDB, and Kinesis Firehose for the rest. In 2017 this team will help design and deliver the next generation big data platform, and roll out the Data Authority service for master data management. If these sound like the right kind of challenges for you, please apply today!

    Basic Qualifications

    • M.S. degree in computer science, mathematics, statistics or a similar quantitative field
    • Experience with Redshift or another columnar store DW
    • Experience with cloud solutions / AWS
    • Experience building reports and/or data visualization
    • Experience with Hadoop/MapReduce/AWS/EMR
    • Experience working with predictive analytics/decision models/data mining libraries as well as the tools for developing such
    • Experience building or administering reporting/analytics platforms
    • Experience building flexible data APIs that consumers use to power other parts of the business
    • Experience with scalable service architecture and design

    Preferred Qualifications

    1. MS in Computer Science · A proven ability to learn and adapt to new, complex development environments with deep dive analytical skills
      • Experience in algorithm design and problem solving · Experience leading small teams of engineers
      • Experience with Amazon technologies like DynamoDB is highly desirable
      • Have previous mobile web development experience
      • Experience developing full stack web applications using Perl, Java, Python or PHP
      • Experience in any of the top JavaScript MVC frameworks, like Backbone, Angular, Ember, Spine or Knockout
      • You stay updated on current web technologies
      • Passion for creating efficient, scalable, highly usable web interfaces
      • Technical Lead experience exploring other opensource frameworks and technologies and integrating with project
      • A love for music

    2. Internal Job Description
      • We manage an event collection service framework, the TULIP VIRT in DW, are a top consumer of Datanet, manage a Redshift VIRT, and process hundreds of million events a day.
      • We are focusing on HDFS / EMR / Hive, Kinesis in the near future.
      • We plan to integrate with and run analytics on Clickstream & Web Labs for all our device clients.
      • We will be doing algorithmic and machine learning platform support for search, personalization, audio quality, and a host of other customer facing enhancements driven by data.

    Sorry the Share function is not working properly at this moment. Please refresh the page and try again later.
    Share this job