We are seeking an experienced and dedicated GCP Data Engineer to join our team. You will be responsible for designing, building, and optimizing robust, scalable, and highly available data pipelines and ETL/ELT solutions exclusively within the Google Cloud Platform (GCP).
Requirements
- Design, build, and maintain scalable data pipelines and ETL/ELT processes using core GCP data services such as Cloud Dataflow (Apache Beam), Cloud Dataproc, and BigQuery.
- Develop and optimize data infrastructure on GCP to ensure reliable, high-speed data ingestion (e.g., using Cloud Pub/Sub and Cloud Storage).
- Implement data quality checks, monitoring, and validation to ensure accuracy and integrity of data across all GCP systems.
- Collaborate closely with Data Scientists and Data Analysts to ensure data readiness for reporting, analytics, and Machine Learning initiatives (e.g., integrating with Vertex AI).
- Automate deployment, monitoring, and testing of data infrastructure and pipelines using Infrastructure as Code (IaC) tools like Terraform and CI/CD practices.
- Manage and optimize GCP data storage solutions, primarily BigQuery (data warehouse) and Cloud Storage (data lake), for performance and cost efficiency.
- Provide technical guidance and recommendations on data architecture and technology choices within the GCP ecosystem.
Benefits
- Generous Paid Time Off
- 401k Matching
- Retirement Plan
- Relocation Assistance