Job Summary
As a Data Engineer, you’ll play a pivotal role in building and maintaining scalable data pipelines and infrastructure to support our organization’s growing data needs. You’ll work closely with data scientists, analysts, and other stakeholders to ensure data quality, consistency, and accessibility.
Job Description
- Design and implement robust data pipelines and batch/streaming data processing workflows to collect, transform, and load data from various sources.
- Develop highly scalable, fault-tolerant, and secure data storage and management solutions using technologies such as data warehouses, data lakes, and NoSQL databases.
- Optimize data processing performance and efficiency using distributed computing frameworks, in-memory processing, and other advanced techniques.
- Develop reusable data engineering components and libraries to support the self-service needs of data scientists and business analysts.
- Collaborate with the data science team to ensure the reliable and timely availability of high-quality data for model training and deployment.
- Automate data quality monitoring, data lineage tracking, and metadata management to ensure the integrity and governance of our data assets.
- Contribute to the evolution of our data architecture and technology stack, staying up to date with the latest industry trends and best practices.
- Mentor and guide junior data engineers, fostering a culture of continuous learning and improvement.
Job Requirements
- Bachelor’s or master’s degree in computer science, Data Engineering, or a related technical field.
- 5+ years of experience in data engineering, with a proven track record of designing and implementing large-scale data platforms.
- Proficient in SQL and data modeling.
- Experience with any orchestration/ETL tool.
- Strong programming skills, preferably in Python.
- Familiarity with data engineering best practices, including data modeling, data quality, and data governance.
- Excellent problem-solving skills, attention to detail, and the ability to work collaboratively in a team.
- Strong communication and interpersonal skills to effectively interact with cross-functional stakeholders.
Qualifications Requirements
- Teradata and/or Informatica experience is a plus.
- Proficiency in one/two programming languages like Python, Scala, or Java, and experience with SQL and NoSQL databases.
- Familiarity with modern data visualization and business intelligence tools (e.g., Tableau, Power BI, Looker).
- Experience in developing and deploying machine learning models as part of data-driven applications.
- Proficiency in Informatica PowerCenter for ETL and data integration.
- Experience in Big Data Management (BDM) and the implementation of data governance and data quality frameworks.