JOB DESCRIPTION
• Design, develop and maintain an infrastructure for streaming, processing and storage of data. Build tools for effective maintenance and monitoring of the data infrastructure.
• Contribute to key data pipeline architecture decisions and lead the implementation of major initiatives.
• Work closely with stakeholders to develop scalable and performant solutions for their data requirements, including extraction, transformation and loading of data from a range of data sources.
• Develop the team’s data capabilities - share knowledge, enforce best practices and encourage data-driven decisions.
• Develop data retention policies, backup strategies and ensure that the firm’s data is stored redundantly and securely.
JOB REQUIREMENTS
• Solid Computer Science fundamentals, excellent problem-solving skills and a strong understanding of distributed computing principles.
• At least 3 years of experience in a similar role, with a proven track record of building scalable and performant data infrastructure.
• Expert SQL knowledge and deep experience working with relational and NoSQL databases.
• Advanced knowledge of Apache Kafka and demonstrated proficiency in Hadoop v2, HDFS, and MapReduce.
• Experience with stream-processing systems (e.g. Storm, Spark Streaming), big data querying tools (e.g. Pig, Hive, Spark) and data serialization frameworks (e.g. Protobuf, Thrift, Avro).
• Bachelor’s or Master’s degree in Computer Science or related field from a top university.
Job Type: Contract
Contract length: 6 months
Salary: Rp10,000,000 - Rp20,000,000 per month
Experience:
• SQL: 3 years (Preferred)
• Hadoop v2, HDFS, and MapReduce.: 3 years (Preferred)
• Storm, Spark Streaming: 3 years (Preferred)
• Pig, Hive, Spark: 3 years (Preferred)
• Protobuf, Thrift, Avro: 3 years (Preferred