Overview
Role involves designing and building distributed data systems and large-scale data warehouses.
Ideal candidate has 8+ years of experience in data engineering with expertise in data platforms and pipelines.
remotefull-timeEnglishPythonSQLPySparkSnowflakeNoSQLApache AirflowAWSGCPLinuxGitHubJenkinsTerraformDockerKubernetes+ 5 more
Locations
Requirements
8+ years experience in data engineering Proficiency in Python, SQL, and PySpark Experience with AWS or GCP Strong presentation and communication skills Experience in building and deploying data pipelines Experience with Docker and Kubernetes Knowledge of disaster recovery planning
Responsibilities
Design and build distributed data systems Architect data platforms for large-scale data Build robust data pipelines Integrate external databases and APIs Conduct experiments for RAG workflows Monitor system performance and optimize Collaborate with onshore/offshore teams Stay updated with industry trends
Benefits
Career development opportunities High degree of individual responsibility