Key Responsibilities:
- Collaborate with our data engineering team to design and develop data pipelines.
- Assist in the extraction, transformation, and loading (ETL) of data from various sources.
- Work with SQL, NoSQL databases to store and manage data efficiently.
- Use big data technologies like Apache Hive, Spark, and cloud big data platform.
- Operate Apache Airflow for orchestrating and scheduling large-scale data workflows.
- Support data modeling efforts to ensure data quality and usability.
- Contribute to the development of data warehousing solutions.
- Explore data stack cloud computing platforms.
- Troubleshoot and optimize data pipelines for performance.
- Collaborate with cross-functional teams on data projects.
Requirements:
- Currently pursuing or recently completed a degree in Computer Science, Information Technology, or a related field.
- Basic programming skills in languages like Python, Scala, or Golang.
- Basic knowledge of SQL and databases.
- Ability to operate Linux commands.
- Familiarity with containerization technologies (e.g., Docker, Kubernetes) would be great.
- Eagerness to learn and adapt to new technologies and tools.
- Excellent problem-solving and critical-thinking abilities.
- Effective communication and teamwork skills.
- Detail-oriented with a commitment to data accuracy.
- Self-motivated and able to work independently.
Laporkan lowongan