Data Engineer

Fluxon

Job Summary

Fluxon is a product development team founded by ex-Googlers and startup founders, partnering with visionary companies from fast-growing startups to tech leaders like Google and Stripe. As the first Data Engineer, you will lead the design, build, and maintenance of data infrastructure to power products and enable data-driven decision-making for clients. This remote role involves designing data models, building reliable data pipelines, ensuring data quality, optimizing query performance, and documenting processes.

Must Have

  • Design and implement data models and warehouse schemas
  • Build and maintain reliable data pipelines
  • Collaborate with product and engineering teams
  • Ensure data quality, integrity, and accessibility
  • Optimize query performance
  • Maintain comprehensive documentation for data models, pipelines, and processes
  • 3-5 years of industry experience in data engineering roles
  • Strong proficiency in SQL and experience with data warehousing concepts (dimensional modeling, star/snowflake schemas)
  • Experience building and maintaining ETL/ELT pipelines
  • Familiarity with cloud data platforms, preferably GCP and BigQuery
  • Understanding of data modeling best practices and data quality principles
  • Solid understanding of software development practices including version control (Git) and CI/CD

Good to Have

  • Experience with Python for data processing and automation
  • Experience with Apache Spark or similar distributed processing frameworks
  • Familiarity with workflow orchestration tools (Airflow, Prefect)
  • Exposure to dbt or similar transformation tools

Perks & Benefits

  • Exposure to high-profile SV startups and enterprise companies
  • Competitive salary
  • Fully remote work with flexible hours
  • Flexible paid time off
  • Profit-sharing program
  • Healthcare
  • Parental leave that supports all paths to parenthood, including fostering and adopting
  • Gym membership and tuition reimbursement
  • Hands-on career development

Job Description

Who we are

We are a product development team founded by ex-Googlers and startup founders. We offer full-cycle software development: from ideation and design to build and go-to-market. We partner with visionary companies, ranging from fast-growing startups to tech leaders like Google and Stripe, to turn bold ideas into products with the power to transform the world.

This is a remote position, with a preference for candidates located in Hyderabad, Bangalore, or Gurgaon, India.

About the role

As the first Data Engineer, you’ll take the lead in designing, building, and maintaining the data infrastructure that powers our products and enables data-driven decision-making for our clients.

You'll be responsible for:

  • Design and implement data models and warehouse schemas to support analytics and reporting needs
  • Build and maintain reliable data pipelines to ingest, transform, and load data from various sources
  • Collaborate with product and engineering teams to understand data requirements and deliver scalable solutions
  • Ensure data quality, integrity, and accessibility across the organization
  • Optimizing query performance and improving the efficiency of existing data infrastructure
  • Maintain comprehensive documentation for data models, pipelines, and processes for team reference

You'll work with technologies including:

Data & Analytics

  • Data Warehouse: Google BigQuery, Snowflake, AWS Redshift, Databricks
  • ETL/Pipeline Tools: Apache Spark, Apache Airflow, dbt
  • Streaming & Queuing: Apache Kafka, Pub/Sub, RabbitMQ

Languages

  • SQL
  • Python (good to have)

Cloud & Infrastructure

  • Platforms: Google Cloud Platform (GCP) or Amazon Web Services (AWS)
  • Storage: Google Cloud Storage (GCS) or AWS S3
  • Orchestration & Processing: Cloud Composer (Airflow), Dataflow, Dataproc

Data Stores

  • Relational: PostgreSQL

Monitoring & Observability

  • GCP Cloud Monitoring Suite

Qualifications

  • 3–5 years of industry experience in data engineering roles
  • Strong proficiency in SQL and experience with data warehousing concepts (dimensional modeling, star/snowflake schemas)
  • Experience building and maintaining ETL/ELT pipelines
  • Familiarity with cloud data platforms, preferably GCP and BigQuery
  • Understanding of data modeling best practices and data quality principles
  • Solid understanding of software development practices including version control (Git) and CI/CD

Nice to have:

  • Experience with Python for data processing and automation
  • Experience with Apache Spark or similar distributed processing frameworks
  • Familiarity with workflow orchestration tools (Airflow, Prefect)
  • Exposure to dbt or similar transformation tools

What we offer

  • Exposure to high-profile SV startups and enterprise companies
  • Competitive salary
  • Fully remote work with flexible hours
  • Flexible paid time off
  • Profit-sharing program
  • Healthcare
  • Parental leave that supports all paths to parenthood, including fostering and adopting
  • Gym membership and tuition reimbursement
  • Hands-on career development

15 Skills Required For This Role

Team Management Github Game Texts Postgresql Rabbitmq Aws Apache Kafka Spark Google Cloud Platform Amazon Web Services Ci Cd Git Python Sql Stripe

Similar Jobs