Job Description
- Design and develop scalable ETL packages from the business source systems and the development of ETL routines to populate databases from sources and also to create aggregates
- Manage end-to-end data processing framework of Data Lake from Data Source to Data Mart and Data Exchange with peripheral systems, with process alignment to company policies
- Work collaboratively with the Data and Analytics team, provide support to the users for its data centric needs
- Perform thorough testing, validation, and deployment in order to deliver the accuracy of data transformations and data verification
- Maintain proficiency in a variety of computer programming languages, such as SQL, PL/SQL, Unix shell scripting, Spark, Python in order to successfully complete the tasks of the role
Qualification
- Bachelor's / Master's degree in computer Engineer, Computer Science, Information Technology, or related fields
- At least 1 year working experience in programming/engineer field; Fresh graduate are welcome
- Ability to work in a fast-moving environment and to deliver the project on time
- Ability to communicate clearly and professionally, both verbally and in writing
- Excellent analysis and reasoning of system behaviors
- Experience in big data and streaming data frameworks (Hadoop/ Spark/ Kafka) is preferable.
- Experience with AWS cloud services e.g., EC2, RDS, Redshift, etc. and pipeline orchestration services is preferable