A recognized property management company in California is actively seeking a new Data Engineer to join their growing team. In this role, the Data Engineer will support the design, automation, and operation of modern data infrastructure.

Responsibilities:
- Design, build, and maintain scalable and resilient CI/CD pipelines for data applications and infrastructure, with a strong focus on Snowflake, dbt, and modern data tooling
- Implement and manage Snowflake dbt projects, including developing dbt models, tests, and documentation, and integrating dbt workflows into CI/CD pipelines
- Develop and manage Infrastructure as Code (IaC) using Terraform to provision and configure data infrastructure on Google Cloud Platform (GCP)
- Automate the deployment, monitoring, and management of Snowflake data warehouse environments, ensuring optimal performance, security, reliability, and cost efficiency
- Collaborate with data engineers and data scientists to understand requirements and deliver automated solutions for data ingestion, transformation, and delivery
- Implement and maintain monitoring, logging, and alerting for data pipelines and infrastructure to ensure high availability and proactive issue resolution
- Develop and maintain automation scripts and tooling using Python (primary) and Bash for operational tasks
- Apply and maintain security best practices across data infrastructure, pipelines, and CI/CD processes
- Troubleshoot and resolve issues related to data pipelines, infrastructure, and deployments
- Participate in code reviews for Terraform, dbt models, and automation scripts
- Create and maintain clear technical documentation for architectures, configurations, and operational processes
- Perform other duties, as needed
Qualifications:
- 5+ years of experience in Data Engineering, Analytics Engineering, or DevOps for data platforms
- Strong hands-on experience with Snowflake in production environments
- Proven expertise with dbt, including model development, testing, and documentation
- Experience building and maintaining CI/CD pipelines for data and cloud infrastructure
- Hands-on experience with Terraform and Infrastructure as Code practices
- Experience working with Google Cloud Platform (GCP) for data storage, processing, and analytics
- Strong programming skills in Python and working knowledge of Bash/shell scripting
- Experience implementing monitoring, logging, and alerting for data systems
- Solid understanding of data pipeline architectures and modern data stack concepts
- Strong troubleshooting and problem-solving skills
- Ability to collaborate effectively with cross-functional technical teams
Desired Skills:
- Experience with Data Governance and Data Quality frameworks
- Familiarity with Cost Optimization strategies in Snowflake and cloud environments
- Experience supporting High-Availability or Mission-Critical Data platforms
- Prior contract or consulting experience in fast-paced environments



