Descripción de la oferta
Scala/Spark Developer with working knowledge in AWS Cloud technology Job Responsibilities / Role \"The successful candidate will work on the AWS infrastructure. They will need to assimilate information from many sources, therefore require excellent data analytical skills. They will build ETL transformations using Scala/Spark. They will create Step functions, work with EMR instances, execute Scala/Spark codes within created EMR instances, lambda functions.\" Mandatory skills Programming Languages - Scala Spark Cloud Technology Services in AWS: EMR, S3, Step Functions Database – SQL Server, Oracle Experience with the following technologies: GitHub Working experience with Terraform modules to create and handle the infrastructure (Glue catalog, S3, code artifacts, etc).\" Nice to have skills Experience with the following technologies: JIRA and Confluence Excellent problem solving, analytical skills and technical troubleshooting skills. Experience in presenting and explaining complex issues and solutions to senior stakeholders in a clear and concise way. Strong teamwork and interpersonal skills; ability to communicate and persuade at all levels; and able to establish positive working relationships. Strong experience in IT build development and support.\" Qualifications : University degree in technology, preferably in computer science.