
Data Engineer
Required Skills
sql
python
aws
data pipeline development
etl
data architecture
cloud platforms
data integration
docker
kubernetes
apache airflow
spark
hadoop
data governance
data security
problem-solving
analytical skills
written communication
verbal communication
remote collaboration
Job Description
Job Title: Data Engineer
Job Type: Full-time
Location: Remote
Job Summary:
Join our customer's team as a Data Engineer and play a pivotal role in building and optimizing scalable data pipelines and architectures. You will collaborate closely with cross-functional teams to deliver high-impact data solutions that empower data-driven decision-making across the organization.
Key Responsibilities:
- Design, construct, install, and maintain robust data pipelines and architectures for large-scale data processing.
- Integrate data from various sources, ensuring accuracy, consistency, and reliability.
- Collaborate with data scientists, analysts, and business stakeholders to understand requirements and deliver effective data solutions.
- Develop, implement, and optimize ETL processes to support business intelligence and analytics needs.
- Monitor, troubleshoot, and enhance performance of data systems, ensuring data quality and availability.
- Document data workflows, processes, and technical decisions clearly for both technical and non-technical audiences.
- Champion best practices in data engineering, including coding standards, version control, and continuous integration.
Required Skills and Qualifications:
- Proven experience as a Data Engineer or in a similar role working with complex data systems.
- Expertise in building and optimizing data pipelines, architectures, and data sets.
- Strong SQL skills and proficiency with at least one programming language (e.g., Python, Java, Scala).
- Experience with cloud-based data solutions (AWS, GCP, or Azure).
- Demonstrated ability to communicate complex concepts clearly in both written and verbal form.
- Strong analytical and problem-solving skills with high attention to detail.
- Ability to work independently in a fully remote environment while collaborating effectively with a distributed team.
Preferred Qualifications:
- Experience with containerization and orchestration tools (e.g., Docker, Kubernetes, Airflow).
- Background in big data technologies such as Spark or Hadoop.
- Prior exposure to data governance and security best practices.