Data Engineer – II (Level 5) – Job ID: 2272444 | ADCI HYD 16 SEZ – IN, TS, Hyderabad

Amazon

  • Full Time

DESCRIPTION

Job summary
Job Description:
Are you interested in being part of rapidly expanding Amazon initiative to redefine how we deliver packages to Amazon customers and take your skills and career to the next level? Join the Last Mile Data Engineering Team (LMDET) to raise the bar for our delivery channels and redefines Amazon’s global Last Mile delivery network.

Key job responsibilities
Last Mile Data Engineering Team is looking for a Data Engineer (DE) to play a significant role in building their large-scale, high-volume, high-performance data integration and delivery services. These data solutions would be primarily used in periodic reporting, and drive business decision making while dealing efficiently with the massive scale of data available through our Data Warehouse as well as our software systems. You will be responsible for designing, developing, troubleshooting, evaluating, deploying, and documenting data warehouse/data lake solutions, enabling stakeholders to manage the business and make effective decisions. You are expected to build efficient, flexible, extensible, and scalable data models, ETL designs and data integration services. You are required to support and manage growth of these data solutions. You must be a self-starter and be able to learn on the go. Excellent written and verbal communication skills are required as you will work very closely with diverse teams. As a Data Engineer, you will be working in one of the largest cloud-based data lakes. You should be skilled in the architecture of data warehouse solutions for the Enterprise using multiple platforms (EMR, RDBMS, Columnar, Cloud). You should have extensive experience in the design, creation, management, and business use of extremely large datasets. You should have excellent business and communication skills to be able to work with business owners to develop and define key business questions, and to build data sets that answer those questions. Above all you should be passionate about working with huge data sets and someone who loves to bring datasets together to answer business questions and drive change.

About the team
The Last Mile Data Engineering Team (LMDET) is part of the Last Mile Geospatial Org. The team comprises of Data engineers skilled in Data Lake implementation, Big Data technology stack, data warehouse solution design, developing complex ETL pipelines using AWS and traditional data engineering tools, owning Last Mile level data marts/aggregated/de-norm datasets across various spaces such as maps, geocoding, routing, places, etc., owning Red Certified AWS infrastructure for handling critical and restricted data access, near real time data ingestions related to transportation world using AWS, dealing with peta bytes of data for reporting and analytics.

BASIC QUALIFICATIONS

• Bachelor’s degree in software engineering or a relevant quantitative discipline
• 4+ years relevant professional experience in Data Engineering or in a similar role
• Strong knowledge of data warehousing concepts, including data warehouse technical architectures, infrastructure components, ETL/ ELT and reporting/analytic tools and environments, data structures, data modeling and performance tuning.
• 3+ years in with Advanced SQL (analytical functions), ETL, Data Warehousing
• 2+ years development experience in Python/Java/Scala for automation and building data pipelines
• 2+ years experience working with core AWS data and analytics services. Understanding on the applicability, limitations, and trade-offs between a wide set of AWS database and analytics technologies.
• 2+ years experience working with distributed computing and associated technologies such as Spark, EMR, etc.
• 2+ years experience with Redshift. Tangible experience working with Redshift Spectrum, AWS Glue, DynamoDB, and S3
• Experience building data products incrementally, integrating and managing large data sets from multiple sources
• A real passion for technology. We are looking for someone who is keen to demonstrate their existing skills while trying new approaches.

PREFERRED QUALIFICATIONS

•. Prefer 5-10 years experience
•. Experience with Massively Parallel Processing (MPP) databases – Redshift, Teradata etc
•. Experience with distributed systems and NoSQL databases
•. Experience with Big Data technologies e.g. Hadoop, Hive, Oozie, Presto, Hue, Spark, Scala and more!
•. Excellent oral and written communication skills including the ability to communicate effectively with both technical and non-technical stakeholders.
•. Proven ability to meet tight deadlines, multi-task, and prioritise workload – A work ethic based on a strong desire to exceed expectations.
•. Excellent analytical skills

Job Overview