Roles and Responsibilities
- Design, develop, and maintain large-scale data pipelines using Python and SQL.
- Collaborate with cross-functional teams to gather requirements and deliver high-quality solutions.
- Develop scalable and efficient ETL processes to extract, transform, and load data from various sources into a centralized repository.
- Ensure data quality by implementing robust validation checks on incoming data streams.
- Troubleshoot issues related to data processing workflows and resolve them promptly.
Desired Candidate Profile
- 3-6 years of experience in Data Engineering with expertise in Python programming language.
- Strong understanding of SQL database management systems (e.g., PostgreSQL) for querying large datasets.
- Experience working with cloud-based technologies such as AWS or GCP for deploying big data applications.
- Proficiency in designing complex data architectures using relational databases like MySQL or NoSQL databases like MongoDB.