We are looking for several talented Data Engineer interns to join our Global Central Data Services team. He/She will have the chance to be involved in the development of the Data Engineering lifecycle implementations and enhancement over cloud-based large-scale distributed architectures and work in a cross-national team with various stakeholders including but not limited to software developers (Backend, Data Analyst, Data Scientist) and fellow Sr. Data engineers.
Responsibilities
Work as part of the Data Services team to maintain and level up the Data infrastructure/Pipeline/Portal for extending its capability, improving robustness and quality, and optimizing for performance
Develop solutions to drive optimization of data-related operations around cloud services management, time to market, security, privacy, and delivery of worldwide capability
Troubleshoot and resolve defects
Requirements
Must be currently enrolled in an undergraduate, master's, or Ph.D. program, ideally in Computer Science or Engineering-related
Experienced in at least one software programming language- Python, Java, Scala, Golang, or other OOP languages
Good computer knowledge, systematic understanding of the operating system, databases, data structures, etc
Concrete SQL skills, such as the principle of SQL execution under different frameworks, and familiarity with structured and unstructured analysis tools of big data
Fundamental experience in big data ecological technology stacks (HDFS, Hive, Elastic-search, HBase, Impala, Spark/Flink, Kafka, Airflow, Sqoop, etc.)
Experience in extracurricular activities such as hackathons, and open-source projects is a big plus