Project Role : Data Engineer
Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems.
Must have skills : AWS Glue
Good to have skills : Amazon Web Services (AWS), Data Engineering, PySpark
Minimum 7.5 Year(s) Of Experience Is Required
Educational Qualification : 15 years full time education
Summary:
As a Data Engineer, a typical day involves designing, developing, and maintaining comprehensive data solutions that support the generation, collection, and processing of data. This role requires creating efficient data pipelines and ensuring the integrity and quality of data throughout its lifecycle. The position also involves implementing extract, transform, and load processes to facilitate seamless migration and deployment of data across various systems, enabling smooth data flow and accessibility for business needs. Collaboration with different teams to align data strategies and troubleshoot data-related challenges is an integral part of the daily routine.
Roles & Responsibilities:
- Expected to be an SME, collaborate and manage the team to perform.
- Responsible for team decisions.
- Engage with multiple teams and contribute on key decisions.
- Provide solutions to problems for their immediate team and across multiple teams.
- Lead the design and implementation of scalable data architectures to support business objectives.
- Mentor junior team members to enhance their technical skills and understanding of data engineering practices.
- Coordinate with stakeholders to gather requirements and translate them into technical specifications.
- Continuously evaluate and improve data processes to optimize performance and reliability.
Professional & Technical Skills:
- Must To Have Skills: Proficiency in AWS Glue, PySpark, Data Engineering, Amazon Web Services (AWS).
- Strong experience in building and managing data pipelines and workflows using cloud-based tools.
- In-depth knowledge of data integration techniques and best practices for ETL processes.
- Ability to work with large datasets and optimize data processing for performance and scalability.
- Familiarity with cloud infrastructure and services to support data storage, processing, and security.
- Experience in troubleshooting and resolving data quality and pipeline issues efficiently.
Additional Information:
- The candidate should have minimum 7.5 years of experience in AWS Glue.
- This position is based at our Pune office.
- A 15 years full time education is required.