Apache Spark including an understanding of optimization techniques
Experience in Python
Experience in SQL
Knowledge of Hadoop architecture
Experience in working with AWS (S3, EMR cluser,Lambda, Kinesis)
Experience in Linux
Good understanding of software development methodologies
Being result-oriented, ability to make things done in a highly dynamic and stressful environment
Good communication skills and good English
Being a good team player
Maintain and support Information Management Cluster based in AWS
Develop and support existing end-to-end data solutions for structured and unstructured data including, but not limited to, ingestion, parsing, integration, auditing, logging, aggregation, normalization, and error handling
Resolve failed or hung jobs
Create jobs and queries to perform auditing and error handling
Our client is a US company that is specialized in delivering reliable Internet connectivity and entertainment multimedia to aircrafts worldwide, enhancing the experience for both passengers and crew. At this moment we starting our cooperation with the Business Intelligence, Data Analysis and BigData solutions support direction and looking for talents who can contribute to the complex data management and analysis projects.