Design, construct, test, and maintain robust, reliable, and scalable data pipeline and services, and investigate and integrate up-and-coming big data technologies into existing requirements.
Requirements
- Designing and modeling data warehouse to support data analytics, reporting, and data applications;
- Implementing data warehouse and ETL processes using Hive, Spark, Hadoop, and/or SQL Server to optimize the data processing performance;
- Programming using SQL, Python, and/or Java;
- Estimating SLAs for the data pipeline, and triaging and debugging data infrastructure issues;
- Building data services or applications using HBase, Elastic Search, ClickHouse, and/or SQL Server.
Benefits
- Commuter benefits
- Flexible PTO
- Health insurance
- Life insurance
- Retirement Plan