Data Engineer
Fluxon
Job Summary
Fluxon is a product development team founded by ex-Googlers and startup founders, partnering with visionary companies from fast-growing startups to tech leaders like Google and Stripe. As the first Data Engineer, you will lead the design, build, and maintenance of data infrastructure to power products and enable data-driven decision-making for clients. This remote role involves designing data models, building reliable data pipelines, ensuring data quality, optimizing query performance, and documenting processes.
Must Have
- Design and implement data models and warehouse schemas
- Build and maintain reliable data pipelines
- Collaborate with product and engineering teams
- Ensure data quality, integrity, and accessibility
- Optimize query performance
- Maintain comprehensive documentation for data models, pipelines, and processes
- 3-5 years of industry experience in data engineering roles
- Strong proficiency in SQL and experience with data warehousing concepts (dimensional modeling, star/snowflake schemas)
- Experience building and maintaining ETL/ELT pipelines
- Familiarity with cloud data platforms, preferably GCP and BigQuery
- Understanding of data modeling best practices and data quality principles
- Solid understanding of software development practices including version control (Git) and CI/CD
Good to Have
- Experience with Python for data processing and automation
- Experience with Apache Spark or similar distributed processing frameworks
- Familiarity with workflow orchestration tools (Airflow, Prefect)
- Exposure to dbt or similar transformation tools
Perks & Benefits
- Exposure to high-profile SV startups and enterprise companies
- Competitive salary
- Fully remote work with flexible hours
- Flexible paid time off
- Profit-sharing program
- Healthcare
- Parental leave that supports all paths to parenthood, including fostering and adopting
- Gym membership and tuition reimbursement
- Hands-on career development
Job Description
Who we are
We are a product development team founded by ex-Googlers and startup founders. We offer full-cycle software development: from ideation and design to build and go-to-market. We partner with visionary companies, ranging from fast-growing startups to tech leaders like Google and Stripe, to turn bold ideas into products with the power to transform the world.
This is a remote position, with a preference for candidates located in Hyderabad, Bangalore, or Gurgaon, India.
About the role
As the first Data Engineer, you’ll take the lead in designing, building, and maintaining the data infrastructure that powers our products and enables data-driven decision-making for our clients.
You'll be responsible for:
- Design and implement data models and warehouse schemas to support analytics and reporting needs
- Build and maintain reliable data pipelines to ingest, transform, and load data from various sources
- Collaborate with product and engineering teams to understand data requirements and deliver scalable solutions
- Ensure data quality, integrity, and accessibility across the organization
- Optimizing query performance and improving the efficiency of existing data infrastructure
- Maintain comprehensive documentation for data models, pipelines, and processes for team reference
You'll work with technologies including:
Data & Analytics
- Data Warehouse: Google BigQuery, Snowflake, AWS Redshift, Databricks
- ETL/Pipeline Tools: Apache Spark, Apache Airflow, dbt
- Streaming & Queuing: Apache Kafka, Pub/Sub, RabbitMQ
Languages
- SQL
- Python (good to have)
Cloud & Infrastructure
- Platforms: Google Cloud Platform (GCP) or Amazon Web Services (AWS)
- Storage: Google Cloud Storage (GCS) or AWS S3
- Orchestration & Processing: Cloud Composer (Airflow), Dataflow, Dataproc
Data Stores
- Relational: PostgreSQL
Monitoring & Observability
- GCP Cloud Monitoring Suite
Qualifications
- 3–5 years of industry experience in data engineering roles
- Strong proficiency in SQL and experience with data warehousing concepts (dimensional modeling, star/snowflake schemas)
- Experience building and maintaining ETL/ELT pipelines
- Familiarity with cloud data platforms, preferably GCP and BigQuery
- Understanding of data modeling best practices and data quality principles
- Solid understanding of software development practices including version control (Git) and CI/CD
Nice to have:
- Experience with Python for data processing and automation
- Experience with Apache Spark or similar distributed processing frameworks
- Familiarity with workflow orchestration tools (Airflow, Prefect)
- Exposure to dbt or similar transformation tools
What we offer
- Exposure to high-profile SV startups and enterprise companies
- Competitive salary
- Fully remote work with flexible hours
- Flexible paid time off
- Profit-sharing program
- Healthcare
- Parental leave that supports all paths to parenthood, including fostering and adopting
- Gym membership and tuition reimbursement
- Hands-on career development