Job Description

Overseeing the collection, storage, management, quality, and protection of data.

Gather and process raw data at scale (including writing scripts, web scraping, calling APIs, writing SQL queries, etc.)

Engineering costs optimization

Lead, motivate, and manage data engineers teams.

Effectively communicate the status, value, and importance of data collection to executive members and staff.

Target: - Client data monitoring and notifications

  • Crawler stability > 90%
  • Mean time to repair crawler incident
Min 4 years programming experience in python, NodeJs and PHP

Min 2 years experience for web crawling and data cleaning

Min 2 years experience in queuing and message broker (Kafka, RabbitMQ, Google Cloud Pub/Sub, NATS). min. 1 platform.

Min 2 years experience with the UNIX environment.

Min 1 year experience in GCP Product like Bigquery, Dataflow, PubSub, GCS

Min 1 year experience in protecting web scrapers against site ban, IP leak, browser crash, CAPTCHA and proxy failure.

Min 1+ years experience with container technology (Kubernetes, Docker)

Bachelor's degree in information technology or related fields

Related Jobs

Job Detail

  • Job Id
    13507e7f503233df
  • Location
    id
  • Company
  • Type
    Private
  • Employment Status
    Permanent
  • Positions
    Available
  • Career Level
    Experience
  • Gender
    Male/Female

Contact

Sponsored by

https://halokerja.id connects jobseekers and recruiters by accurately matching candidate profiles to the relevant job openings through an advanced 2-way matching technology. While most job portals only focus on getting candidates the next job, Shine focuses on the entire career growth of candidates.

Latest Job