SummaryBy Outscal
OKX seeks a Data Engineer to design, build, and maintain resilient data pipelines. You'll work with industry-standard tools like Spark and Flink on cloud infrastructure, collaborate with various teams, and ensure data integrity. Must have experience with data processing tools, batch & streaming pipelines, Python/Go/Scala/Java, SQL/NoSQL databases, and DevOps tools.
OKX will be prioritising applicants who have a current right to work in Singapore, and do not require OKX's sponsorship of a visa.
Who We Are
At OKX, we believe that the future will be reshaped by Crypto, ultimately contributing to every individual's freedom. OKX began as a crypto exchange giving millions of people access to crypto trading and over time becoming among the largest platforms in the world. In recent years, we have developed one of the most connected Web3 wallets used by millions to access decentralized crypto applications (dApps). OKX is a trusted brand by hundreds of large institutions seeking access to crypto markets on a reliable platform that seamlessly connects with global banking and payments. In the last year, OKX has expanded into new markets including Australia, Brazil, Netherlands, Singapore and Turkey, with plans to launch in the US, Belgium and the UAE.
We are deeply committed to shaping a fairer, more transparent and accessible society through blockchain technology. This is why we publish proof of reserves monthly, and continue to ship new innovative security features.
About the team:
OKX data team is responsible for the whole data scope of OKG, from techincal selection, architecture design, data ingestion, data storage, ETL, data visualization to business intelligence and data science. We are data engineers, data analysts and data scientists. The team has end-to-end ownership of most of the data at OKx throughout the whole data lifecycle including data ingestion, data ETL, data warehouse and data services. As a data engineer of the team, you will work with the team to leverage data technologies to empower evidence-based decision-making and improve the quality of the company's products and services.
Responsibilities:
- Design and build resilient and efficient data pipelines for both batch and real-time streaming data
- Architect and design data infrastructure on cloud using industry standard tools
- Execute projects with an Agile mindset
- Build software frameworks to solve data problems at scale
- Collaborate with product managers, software engineers, data analysts and data scientists to build scalable and data-driven platforms and tools
- Ensure data integrity and scalability through enforcement of data standards. Improve data validation and monitoring processes to proactively prevent issues and quickly identify issues. Drive resolution on the issues.
- Define, understand, and test external/internal opportunities to improve our products and services.
Requirements:
- Bachelor’s Degree in Computer Science or have equivalent professional experience
- Solid Experience with data processing tools such as Spark, Flink
- Solid Experience implementing batch and streaming data pipelines
- Solid experiences in Python/Go/Scala/Java.
- In-depth knowledge of both SQL and NoSQL databases, including performance tuning and troubleshooting
- Familiar with DevOps tools such as Git, Docker, k8s
- Experience with the cloud (e.g. AWS, Ali Cloud, GCP, Azure)
- Be proficient in SQL, familiar with advanced SQL features such as window functions, aggregate functions and creating scalar functions/user-defined functions.
- Proven successful and trackable experience in full end-to-end data solutions involving data ingestion, data persistence, data extraction and data analysis.
- Self-driven, innovative, collaborative, with good communication and presentation skills
- Fluent in English, both written and spoken.
Preferred Qualifications:
- Experience in FinTech, eCommerce, SaaS, AdTech, or Digital Wallet business industries.
- Experience in working with teams across offices and timezones is a plus.
- Experience in big data tools such as Amplitude/Tableau/QlikView, Ali Cloud DataWorks, MaxCompute, Hadoop, Hive, Spark and HBase is a big plus.
More that we love to tell you along the process!