Job Description
To excel in this role, you should satisfy the following prerequisites: (Essential Requirements)
- Proficiency in Scala development and design utilizing Scala 2.10+ or Java development and design with Java 1.8+.
- Familiarity with a range of technologies, including Apache Hadoop, Scala, Apache Spark, Spark Streaming, YARN, Kafka, Hive, Python, ETL frameworks, MapReduce, SQL, and RESTful services.
- Solid understanding of Unix/Linux platforms.
- Hands-on experience constructing data pipelines using Hadoop components such as Hive, Spark, and Spark SQL.
- Proficiency with industry-standard version control tools like Git and GitHub, automated deployment tools like Ansible and Jenkins, and requirement management in JIRA.
- Knowledge of big data modeling techniques employing both relational and non-relational approaches.
- Capability to debug code issues and effectively communicate highlighted differences to the development team and architects.
Additionally, the ideal candidate will possess the following qualifications: (Desirable Requirements)
- Experience with time-series/analytics databases such as Elasticsearch.
- Familiarity with scheduling tools like Airflow and Control-M.
- Understanding or exposure to Cloud design patterns.
- Experience with DevOps and Agile project methodologies, including Scrum and Kanban.
- Proficiency in developing Hive QL and UDFs for analyzing semi-structured/structured datasets.
More Information
- Experience 5-10 Years