Mandatory Skills -Data Engineer, SQL, Azure, Python, AWS data pipelines
Relevant years experience - 6+ Years
Job Description:
Role Overview
We are seeking a highly skilled Data Engineer with 6+ years of experience to design, develop, and optimize our data pipelines and systems. The ideal candidate will have strong proficiency in Python, extensive knowledge of SQL programming, and a good understanding of DevOps practices. Hands-on experience with AWS S3 or Azure Databricks is essential for this role.
Key Responsibilities:
- Design, build, and maintain efficient, scalable, and reliable data pipelines to support various business needs.
- Develop and optimize ETL processes to integrate data from various structured and unstructured data sources into centralized storage.
- Work with AWS S3 for data storage and management, ensuring scalability and security of data systems.
- Implement data processing workflows and transformations using Azure Databricks.
- Write complex SQL queries for data extraction, transformation, and analysis.
- Monitor and troubleshoot data pipelines and systems to ensure smooth and uninterrupted operation.
- Collaborate with DevOps teams to implement CI/CD pipelines for data solutions, ensuring high availability and reliability.
- Optimize system performance and scalability by implementing best practices in data engineering.
- Ensure data quality, integrity, and compliance with organizational standards and policies.
- Work closely with cross-functional teams, including data scientists, analysts, and software engineers, to deliver high-impact data solutions.
Required Skills and Qualifications:
- Experience: 6+ years in data engineering or a similar role.
- Programming: Proficient in Python, with a strong understanding of its data structures and libraries (e.g., pandas, NumPy, etc.).
- SQL: Advanced skills in SQL programming for complex data manipulation and analysis.
- DevOps: good experience with CI/CD pipelines, version control systems (e.g., Git), and infrastructure-as-code tools.
- Cloud Platforms: Hands-on experience with AWS S3 or Azure Databricks.
- Data Pipelines: Expertise in building, testing, and maintaining ETL processes and workflows.
- Data Formats: Familiarity with data serialization formats such as JSON, Parquet, and Avro.
- Problem-Solving: Strong analytical and troubleshooting skills for debugging data pipeline issues.
- Collaboration: Excellent communication and teamwork skills to work effectively in a cross-functional environment.
Exposure to visualization tools ( PowerBI/Tableau) would be an added advantage.
Work location & Mode - Pune and Bangalore (Hybrid-3 days)