
Data Warehouse Engineer, AI Data Services
BinancePosted 5/18/2025

Data Warehouse Engineer, AI Data Services
Binance
Job Location
Job Summary
We are seeking a Senior Data Warehouse Engineer at Binance to design and maintain a flexible, enterprise-scale data warehouse. The ideal candidate will have 5+ years of experience in designing and developing data lakes and data warehouses, with expertise in data modeling, governance, and Big Data technologies such as Hadoop, Spark, Hive, etc. They will collaborate closely across business and technical teams to accelerate insights and minimize redundant work. With a strong analytical mindset and ability to abstract and architect solutions, the candidate will lead data governance initiatives and foster technical team growth through mentorship and knowledge sharing. Binance offers a competitive salary, company benefits, and a user-centric global organization with a flat structure.
Job Description
Responsibilities
- Architect and implement a flexible, scalable data warehouse aligned with company specifications and business requirements, accelerating delivery and reducing redundant development.
- Design, develop, test, deploy and monitor data models and ETL jobs; rapidly troubleshoot complex issues and optimize calculation logic and pipeline performance.
- Lead data governance initiatives by building and maintaining metadata management and data quality monitoring systems.
- Foster technical team growth through mentorship, knowledge sharing and continuous improvement of collective skills.
Requirements
- 5+ years of hands-on experience designing and developing data lakes and data warehouse solutions.
- Deep expertise in data warehouse modeling and governance, including dimensional modeling, information factory (data vault) methodologies and “one data” principles.
- Proficiency in at least one of Java, Scala or Python, plus strong Hive & Spark SQL programming skills.
- Practical experience with OLAP engines (e.g., Apache Kylin, Impala, Presto, Druid).
- Proven track record in building high-throughput batch pipelines on Big Data platforms.
- Familiarity with core Big Data technologies (Hadoop, Hive, Spark, Delta Lake, Hudi, Presto, HBase, Kafka, Zookeeper, Airflow, Elasticsearch, Redis).
- AWS Big Data service experience is a plus.
- Strong analytical and system-design capabilities, with a clear understanding of business requirements and ability to abstract and architect solutions.
- Collaborative mindset, skilled at building partnerships across teams and stakeholders.
- Preferred: Experience managing petabyte-scale data in Internet-scale environments and resolving critical production incidents.