Data Warehouse Engineer

See more jobs from Binance

5 months old

Apply Now

Binance is the leading global blockchain ecosystem and cryptocurrency infrastructure provider whose suite of financial products includes the world’s largest digital-asset exchange.
Our mission is to accelerate cryptocurrency adoption and increase the freedom of money.

If you’re looking for a fast-paced, mission-driven organization where opportunities to learn and excel are endless, then Binance is the place for you.

Requirements

  • According to the company's data warehouse specifications and business understanding, build a universal and flexible data warehouse system that can quickly support the needs and reduce repetitive development work efforts.
  • Data model design, development, testing, deployment, online data job monitoring, and the ability to quickly solve complex problems, especially the optimization of complex calculation logic and performance tuning, etc.
  • Participate in Data governance, including the construction of the company’s metadata management system and data quality monitoring system. 
  • Design and implement a data platform integrated with data lake warehouse to support real-time data processing and analysis requirements.
  • Build knowledge graph, and provide in-depth business insight.
  • Participate in technical team building and learning growth, and contribute to the team’s overall knowledge accumulation and skill improvement.
  • Responsibilities

  • 5+ years experiences of data lake and data warehouse design and development experience.
  • Deeply understanding of data warehouse modeling and data governance. Solid knowledge of data warehouse development methodology, including dimensional modeling, information factory etc.
  • Proficient in Java / Scala / Python (at least one language) and Hive & Spark SQL programming languages.
  • Familiar with OLAP technology (such as: kylin, impala, presto, druid, etc.).
  • Proficient in Big Data batch pipeline development.
  • Familiar with Big Data components including but not limited to Hadoop, Hive, Spark, Delta lake, Hudi, Presto, Hbase, Kafka, Zookeeper, Airflow, Elastic search, Redis, etc.
  • Experiences with AWS Big Data services are a plus.
  • Have a strong team collaboration attitude and develop partnerships with other teams and businesses.
  • Rich experience in real-time data processing, familiar with stream processing frameworks such as Apache Kafka, Apache Flink, in-depth knowledge of Lakehouse technology, practical project experience, proficiency in StarRocks, including its data model design, query optimization and performance tuning.
  • Experience in knowledge graph construction and application, and knowledge of graph databases such as Nebula, etc.