Participate in the development of data products which consist of system design, coding review and continuous delivery.
Be able to create and / or maintain optimal data pipeline and provide solutions with good design and code quality.
To build an analytics tool that utilizes the data pipeline to provide insights into customer needs, operational efficiency and other key business performance metrics.
Keep learning emerging technologies and share the knowledge you have the team to improve team skills.
At least 3 years of hands-on software development experience with Python, Java or Golang.
Proficient in SQL and other query DSL.
Experience in working with cloud service such as AWS, GCP.
Experience with batch / streaming processing system, such as Hadoop, Spark, Spark Structured Streaming, Kafka, NiFi etc.
Experience in custom ETL design, implementation and maintenance.
Good understanding of distributed system, basic mathematics such as statistics and probability.
Nice to have
Experience with machine learning and deep learning framework, such as Scikit-learn, TensorFlow, PyTorch, etc. Applied experience preferred.
Experience in working with designing with RESTful APIs.
Experience with massive volume data version control use Delta Lake, Iceberg or Hudi.