Offer summary
Qualifications:
5+ years experience in Hadoop and GCP, Proficiency in Apache Spark, Python, and cloud platforms.Key responsabilities:
- Develop data processing tasks using pySpark
- Design, build and test applications in Spark
Role: Hadoop+GCP
Necessary to have:
• Professional experience with a cloud platform
• Developers must have sound knowledge in Apache Spark and Python programming.
• Deep experience in developing data processing tasks using pySpark such as reading data from external sources, merge data, perform data enrichment and load in to target data destinations.
• Ability to design, build and unit test the application in Spark/Pyspark.
• In-depth knowledge of Hadoop, Spark, and similar frameworks.
• Ability to understand existing ETL & logic to convert into Spark/PySpark/ Spark SQL.
• Knowledge of Unix shell scripting, RDBMS, Hive, HDFS File System, HDFS File Types, HDFS compression codec.
• Experience in processing large amounts of structured and unstructured data, including integrating data from multiple sources.
ServiceNow
TalentKube
Kyndryl Japan
AppLab Systems, Inc
ServiceNow