Data Engineer

Data Engineer

Contract Type:

Contractor

Location:

Newport Beach - California

Industry:

Technology

Reference Number:

CR/508810

Salary:

$30 - $40 Hourly

Contact Name:

Tandym Group

Contact Email:

cole.ehrsam@tandymgroup.com

Contact Phone:


Date Published:

22-Dec-2025

A recognized property management company in California is actively seeking a new Data Engineer to join their growing team. In this role, the Data Engineer will support the design, automation, and operation of modern data infrastructure. 

 Responsibilities:
  • Design, build, and maintain scalable and resilient CI/CD pipelines for data applications and infrastructure, with a strong focus on Snowflake, dbt, and modern data tooling
  • Implement and manage Snowflake dbt projects, including developing dbt models, tests, and documentation, and integrating dbt workflows into CI/CD pipelines
  • Develop and manage Infrastructure as Code (IaC) using Terraform to provision and configure data infrastructure on Google Cloud Platform (GCP)
  • Automate the deployment, monitoring, and management of Snowflake data warehouse environments, ensuring optimal performance, security, reliability, and cost efficiency
  • Collaborate with data engineers and data scientists to understand requirements and deliver automated solutions for data ingestion, transformation, and delivery
  • Implement and maintain monitoring, logging, and alerting for data pipelines and infrastructure to ensure high availability and proactive issue resolution
  • Develop and maintain automation scripts and tooling using Python (primary) and Bash for operational tasks
  • Apply and maintain security best practices across data infrastructure, pipelines, and CI/CD processes
  • Troubleshoot and resolve issues related to data pipelines, infrastructure, and deployments
  • Participate in code reviews for Terraform, dbt models, and automation scripts
  • Create and maintain clear technical documentation for architectures, configurations, and operational processes
  • Perform other duties, as needed
Qualifications:
  • 5+ years of experience in Data Engineering, Analytics Engineering, or DevOps for data platforms
  • Strong hands-on experience with Snowflake in production environments
  • Proven expertise with dbt, including model development, testing, and documentation
  • Experience building and maintaining CI/CD pipelines for data and cloud infrastructure
  • Hands-on experience with Terraform and Infrastructure as Code practices
  • Experience working with Google Cloud Platform (GCP) for data storage, processing, and analytics
  • Strong programming skills in Python and working knowledge of Bash/shell scripting
  • Experience implementing monitoring, logging, and alerting for data systems
  • Solid understanding of data pipeline architectures and modern data stack concepts
  • Strong troubleshooting and problem-solving skills
  • Ability to collaborate effectively with cross-functional technical teams
Desired Skills:
  • Experience with Data Governance and Data Quality frameworks
  • Familiarity with Cost Optimization strategies in Snowflake and cloud environments
  • Experience supporting High-Availability or Mission-Critical Data platforms
  • Prior contract or consulting experience in fast-paced environments
APPLY NOW

Share this job

Interested in this job?
Save Job
Create As Alert

Similar Jobs

SCHEMA MARKUP ( This text will only show on the editor. )