Mô Tả Công Việc
- Works on the data pipeline infrastructure that is veritably the backbone of our business
- Writing elegant functional Scala code to crunch TBs of data on Hadoop clusters, mostly using Spark
- Be owning a data pipeline deployment to clusters: on-prem or on-cloud (AWS or GCP or more).
- Be managing Hadoop clusters right from security to reliability to HA.
- Building a pluggable, unified data lake from scratch.
- Automating and scaling tasks for the Data Science team.
- Constantly look to improve framework and pipelines, hence learning on the job is sort of a given.
- Our expertise and requirements include but are not limited to Spark, Scala, HDFS, Yarn, Hive, Kafka, Distributed Systems, Python, Datastore (Relational and NoSql) and Airflow.
Yêu Cầu Công Việc
- Experienced with Big Data stacks: Spark, Hadoop (Must have).
- Experienced with Java/Scala (Python is a plus).
- Knowledge of various ETL techniques and frameworks, such as Spark, Yarn, Airflow, Oozie, etc.
- Understanding of distributed computing principles
- Experience with NoSQL databases.
- Experience with various messaging systems, such as Kafka or RabbitMQ
Hình thức
Quyền Lợi
- Fixed 13th-month salary
- Performance bonus 1-2 months salary (excluding 13th month)
- Free Grab for work for employees (even during the probationary period - from the first day of joining the company)
- Lunch for employees (even during probation period - 1 day 5 dishes)
- 100% salary on probation
- 22 days of paid leave
- Premium health care for employees and 3 relatives (medical treatment at international standard hospitals)
- Provide MacbookPro and 27" monitor as well as necessary equipment for work
- Happy hour every Thursday (pizza, fried chicken, sushi...)
- Free coffee, tea, snacks, and fruit...all-day
- Comfortable working hours, no fingerprint timekeeping or card swiping.
- The company sponsors 100% of expenses for sports activities: badminton, football...
- Budget for monthly team building/ team bonding
- Yearly company trip...