Job Search and Career Advice Platform

Enable job alerts via email!

Big Data Engineer

Methodfi

Snowflake (AZ)

On-site

USD 100,000 - 130,000

Full time

30+ days ago

Generate a tailored resume in minutes

Land an interview and earn more. Learn more

Job summary

A leading tech company in the United States is looking for a Big Data Engineer to design and optimize data pipelines and manage data storage. The ideal candidate has proven experience with big data ecosystems, SQL and NoSQL databases, and familiarity with cloud-native architectures. This role offers significant ownership and immediate impact on the organization.

Qualifications

  • Proven experience building scalable data pipelines and infrastructure.
  • Strong background in big data ecosystems and distributed systems.
  • Experience in data lake architecture and table format management.
  • Comfort working across multiple types of databases (both SQL and NoSQL).
  • Familiarity with DevOps for data, including containerization (Docker), CI/CD for data pipelines.
  • Experience with cloud-native architectures (AWS, GCP or Azure is a plus).

Responsibilities

  • Design and build scalable, production-grade data pipelines using Airflow, dbt, and PySpark.
  • Architect and manage large-scale Data Lakes / Delta Lakes.
  • Work with big data processing frameworks to handle massive datasets efficiently.
  • Manage and optimize data storage across modern table formats (Iceberg, Delta) and warehouses (Snowflake, BigQuery, Redshift).
  • Operate across a wide range of NoSQL systems such as MongoDB, Redis, and Neo4j.
  • Deploy and scale data infrastructure in the cloud.

Skills

Building scalable data pipelines
Big data ecosystems
Data lake architecture
SQL and NoSQL databases
DevOps for data
Cloud-native architectures

Education

Bachelor’s degree in Computer Science, Engineering, or related field

Tools

Airflow
dbt
PySpark
Docker
AWS
GCP
Azure
MongoDB
Redis
Neo4j
Snowflake
BigQuery
Redshift
Job description
Overview

As a Big Data Engineer at Tavily, you’ll work at the heart of our systems, shaping the data backbone that powers real-time AI agents. From managing billions of records across NoSQL and SQL databases to optimizing high-throughput pipelines, your work will help our models think faster, our agents act smarter, and our users get answers in real time.

What You’ll Do
  • Design and build scalable, production-grade data pipelines using Airflow, dbt, and PySpark.
  • Architect and manage large-scale Data Lakes / Delta Lakes.
  • Work with big data processing frameworks to handle massive datasets efficiently.
  • Manage and optimize data storage across modern table formats (Iceberg, Delta) and warehouses (Snowflake, BigQuery, Redshift).
  • Operate across a wide range of NoSQL systems such as MongoDB, Redis, and Neo4j.
  • Deploy and scale data infrastructure in the cloud.
What We’re Looking For
  • Proven experience building scalable data pipelines and infrastructure.
  • Strong background in big data ecosystems and distributed systems.
  • Experience in data lake architecture and table format management.
  • Comfort working across multiple types of databases (both SQL and NoSQL).
  • Familiarity with DevOps for data, including containerization (Docker), CI/CD for data pipelines.
  • Experience with cloud-native architectures, (AWS , GCP or Azure is a plus).
  • Bonus: Bachelor’s degree in Computer Science, Engineering, or a related field.
What Kind of Engineer Thrives Here
  • Moves fast, breaks bottlenecks, and loves getting their hands dirty.
  • Embraces changing requirements and is motivated by real-world impact.
  • Thrives in high-ownership, zero-handoff environments. You build it, you run it.
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.