Responsibilities:
- Work on next generation UDW data platform which is built with distributed SQL engine upon Hadoop eco-system
- Improve the daily ETL loading process and make it more automated and error prove to handle the increasing daily data volume of our system
- Design the new programming interface of our Data Warehouse, which will be used to connect to its upstream and downstream applications
- Work with the globally distributed team members and counterparts, understand the users ’ requirements, provide production support, and implement solutions for our projects
- Oversee the data quality issues of the daily risk reports, carry the data reconciliation process, explain the differences, and make updates/adjustments to the data as needed
Requirements (indicate mandatory or/and preferred):
Mandatory:
- Bachelor Degree in Computer Science
- Experience with Hadoop ecosystem (Spark/Drill/Kylin and other distributed SQL engines) over 4 years
- Experience with Java core and advanced programming
- Strong Unix shell scripting skills
- Able to communicate verbally in English
- Experience in coordinating work on global scale
- Familiar with testing and development for server side applications, experience with using Spring
- Diligent with unit testing and quality-assurance;
- Able to quickly research and understand new concepts and technologies; and apply to new development
- Excellent communication skill to work with application users and team members
Preferred:
- Experience with Spark/Kafka
- Experience with GemFire
- Experience with database, including UDB, Oracle SQL Server
- Experience with Microsoft Analysis Server and OLAP technology
- Understanding of financial products, specifically risk associated with fixed income and Equities products