Senior Data Engineer

Sumo logic

Job Summary

As a Sr Data Engineer at Sumo Logic, you will design and develop data engineering solutions, focusing on complex data integrity, security, process, and sanitization problems across Sumo Logic’s cloud services, data tooling, and business applications. You will work with a team to deliver high-impact engagements, ensuring scalable, resilient, secure, and cost-efficient data solutions. This role involves improving existing data warehouse systems and ensuring global delivery and 24/7 uptime.

Must Have

  • Designing and implementing complex and well-architected enterprise data solutions.
  • Work to improve existing data warehouse systems, solutions, and process.
  • Experience in DevOps methodologies and practices as applied to infrastructure, cloud, and business application data.
  • Improve systems to provide strict data compliance adherence, security guardrails, and cost-optimization.
  • Strong understanding and hands-on experience with AWS data infrastructure and compute services.
  • Design and manage data schemas and the flow of data through corporate systems.
  • Deliver well-architected, end-to-end data solutions to a growing enterprise organization.
  • Build strong partnerships and engage with other teams at Sumo Logic.
  • Ensuring global delivery and alignment on all data initiatives and maintaining 24/7 uptime.
  • Extensive experience in Databricks, Spark, and AWS (EC2, RDS, Aurora, DynamoDB, S3 and Kinesis primarily).
  • Experience developing scalable, secure, and resilient data architectures and implementations.
  • 6+ years of industry experience with a proven track record of ownership and delivery.
  • Experience with Python scripting, PySpark, and other data frameworks or tools.
  • Experience with SQL and data schemas.
  • Experience with API calls and API-based ingestion.
  • Hands-on and deep experience with Git and GitHub.
  • Proven experience and success in AWS infrastructure management and deployment for data platforms.
  • Experience working with the data ingestion, data storage, and data consumption layers.
  • Experience with both structured and unstructured data.
  • Agile development experience and a familiarity with Jira, sprints, and pointing.
  • Experience building robust and well-architected designs for enterprise scale data architectures and workflows.
  • Excellent verbal and written communication skills.
  • Experience and comfort with an on-call schedule for enterprise systems.

Good to Have

  • Experience in AI/ML, Data Science, LLMs, Contextualization, Amazon Bedrock, and Amazon Nova, SageMaker, Iceberg.
  • Experience in Terraform.
  • Experience in Tableau, Looker, and AWS Quicksight.
  • Experience with Big Data services such as HDFS, Spark, Hive, HBase, Yarn, and Oozie.

Job Description

The proliferation of machine log data has the potential to give organizations unprecedented real-time visibility into their infrastructure and operations. With this opportunity comes tremendous technical challenges around ingesting, managing, and understanding high-volume streams of heterogeneous data.

As a Sr Data Engineer, you will actively contribute to, execute for, and lead IT in the design and development of data engineering solutions. You will be instrumental in helping us solve complex data integrity, security, process, and sanitization problems among Sumo Logic’s robust inventory of cloud services, data tooling, and business applications. You will work closely with a team of highly skilled individuals to deliver on high impact engagements at Sumo Logic, rolling up to the Global Head of IT and reporting to our Senior Manager of Data Engineering.

You are a strong data engineer who is passionate about scalable, resilient, secure, and cost-efficient data solutions. You care about producing clean, elegant, maintainable, robust, well-tested code; you do this as a member of a team, helping the group come up with a better solution than you would as individuals and providing experienced guidance and insight. Ideally, you have experience with performance, scalability, and reliability issues of 24x7 uptime systems and solutions.

Responsibilities:

  • Designing and implementing complex and well-architected enterprise data solutions that support mid-to-large size organizations
  • Work to improve existing data warehouse systems, solutions, and process through architectural reviews and design enhancement initiatives
  • Experience in devops methodologies and practices as applied to infrastructure, cloud, and business application data
  • Improve systems to provide strict data compliance adherence, security guardrails, and cost-optimization
  • Strong understanding and hands-on experience with data infrastructure and compute services in AWS that our data platform is deployed on top of
  • Design and manage data schemas and the flow of data through corporate systems and applications to ensure compliance, data integrity, and data security
  • Deliver well-architected, end-to-end data solutions to a growing enterprise organization across multiple infrastructure environments, data sources, and business applications
  • Build strong partnerships and engage with other teams at Sumo Logic to create solutions that enable and benefit all parties and increased derived value
  • Ensuring global delivery and alignment on all data initiatives and maintaining 24/7 uptime

Requirements:

  • Extensive experience in Databricks, Spark, and AWS (EC2, RDS, Aurora, DynamoDB, S3 and Kinesis primarily)
  • Experience developing scalable, secure, and resilient data architectures and implementations
  • 6+ of industry experience with a proven track record of ownership and delivery
  • Experience with Python scripting, PySpark, and other data frameworks or tools
  • Experience with SQL and data schemas
  • Experience with API calls and API-based ingestion
  • Hands-on and deep experience with git and GitHub
  • Proven experience and success in AWS infrastructure management and deployment for data platforms to leverage and run on top of (EC2, S3, RDS, VPC, Basic Network Adjustments, KMS, PrivateLink all a bonus)
  • Experience working with the data ingestion, data storage, and data consumption layers
  • Experience with both structured and unstructured data
  • Agile development experience and a familiarity with Jira, sprints, and pointing
  • Experience building robust and well-architected designs for enterprise scale data architectures and workflows
  • You should have a passion for continuous learning and deep technological curiosity
  • Excellent verbal and written communication skills
  • Experience and comfort with an on-call schedule for enterprise systems

Desirable:

  • Experience in AI/ML, Data Science, LLMs, Contextualization, Amazon Bedrock, and Amazon Nova, SageMaker, Iceberg
  • Experience in Terraform is a major positive
  • Experience in the following specific technologies is a plus:
  • Tableau, Looker, and AWS Quicksight
  • Big Data services such as HDFS, Spark, Hive, HBase, Yarn, and Oozie

17 Skills Required For This Role

Communication Data Analytics Github Game Texts Agile Development Hbase Aws Terraform Looker Tableau Spark Data Science Yarn Git Python Jira Sql

Similar Jobs