Job Description :
Key Responsibilities :
- Design, develop, and maintain scalable data pipelines and ETL processes to support data integration from multiple sources.
- Perform data modeling for both structured and semi-structured datasets.
- Implement and manage data warehousing solutions to store and organize large volumes of data efficiently.
- Optimize ETL performance, monitor data workflows, and troubleshoot data-related issues.
- Collaborate with data scientists, analysts, and business stakeholders to understand data requirements and deliver high-quality solutions.
- Ensure data quality, consistency, and governance across all data assets.
- Stay current with emerging big data and cloud technologies to recommend modern solutions.
Required Qualifications :
- Bachelor's or Masters degree in Computer Science, Engineering, Information Systems, or a related field.
- Minimum 5+ years of experience in data engineering, with strong knowledge of data pipelines and ETL processes.
- Proficiency in SQL and experience with scripting languages (e.g., Python, Scala).
- Strong understanding of data modeling techniques (Star/Snowflake schemas).
- Experience with data warehousing solutions (e.g., Amazon Redshift, Google BigQuery, Snowflake).
- Hands-on experience with ETL tools and orchestration frameworks such as Apache Airflow, Talend, or Informatica.
- Exposure to big data technologies such as Hadoop, Spark, Hive, etc., is a plus.
- Familiarity with cloud platforms (AWS, GCP, or Azure).
- Excellent problem-solving skills and the ability to work in a collaborative team environment.
- Strong communication skills, both written and verbal.
Nice to Have :
- Experience working in hybrid or remote teams.
- Background in data analytics or working closely with BI teams.
- Familiarity with CI/CD pipelines and DevOps practices in a data environment.