Data Engineer

Tandym Tech Logo
  • Logistics
  • FlexTime
  • Applications have closed

A recognized property management company in California is actively seeking a new Data Engineer to join their growing team. In this role, the Data Engineer will support the design, automation, and operation of modern data infrastructure.

Responsibilities:

  • Design, build, and maintain scalable and resilient CI/CD pipelines for data applications and infrastructure, with a strong focus on Snowflake, dbt, and modern data tooling

  • Implement and manage Snowflake dbt projects, including developing dbt models, tests, and documentation, and integrating dbt workflows into CI/CD pipelines

  • Develop and manage Infrastructure as Code (IaC) using Terraform to provision and configure data infrastructure on Google Cloud Platform (Google Cloud Platform)

  • Automate the deployment, monitoring, and management of Snowflake data warehouse environments, ensuring optimal performance, security, reliability, and cost efficiency

  • Collaborate with data engineers and data scientists to understand requirements and deliver automated solutions for data ingestion, transformation, and delivery

  • Implement and maintain monitoring, logging, and alerting for data pipelines and infrastructure to ensure high availability and proactive issue resolution

  • Develop and maintain automation scripts and tooling using Python (primary) and Bash for operational tasks

  • Apply and maintain security best practices across data infrastructure, pipelines, and CI/CD processes

  • Troubleshoot and resolve issues related to data pipelines, infrastructure, and deployments

  • Participate in code reviews for Terraform, dbt models, and automation scripts

  • Create and maintain clear technical documentation for architectures, configurations, and operational processes

  • Perform other duties, as needed

Qualifications:

  • 5+ years of experience in Data Engineering, Analytics Engineering, or DevOps for data platforms

  • Strong hands-on experience with Snowflake in production environments

  • Proven expertise with dbt, including model development, testing, and documentation

  • Experience building and maintaining CI/CD pipelines for data and cloud infrastructure

  • Hands-on experience with Terraform and Infrastructure as Code practices

  • Experience working with Google Cloud Platform (Google Cloud Platform) for data storage, processing, and analytics

  • Strong programming skills in Python and working knowledge of Bash/shell scripting

  • Experience implementing monitoring, logging, and alerting for data systems

  • Solid understanding of data pipeline architectures and modern data stack concepts

  • Strong troubleshooting and problem-solving skills

  • Ability to collaborate effectively with cross-functional technical teams

Desired Skills:

  • Experience with Data Governance and Data Quality frameworks

  • Familiarity with Cost Optimization strategies in Snowflake and cloud environments

  • Experience supporting High-Availability or Mission-Critical Data platforms

  • Prior contract or consulting experience in fast-paced environments