Jobs Description
RTH: Yes. Total exp: 6 years, Rel exp: 4 to 6 yrs, Mandatory skill: Python, Java, Spark JD: Hands on experience on Python, Java, Spark Programming and Streaming datasets in Big Data platform ? Should have extensive working experience in Hive and other components of Hadoop eco system ( Job Server and Kafka) ? Monitor and optimize data pipelines and data processing systems ? Work with other engineers to ensure data quality and data security ? Should be able to understand the complex transformation logic and translate them to Spark-script. ? Unix Shell Scripting and Airflow scheduling. ? Should have worked on Cloudera distribution framework, Airflow Workflow (or any Scheduler), Jenkins (or any version controller) ? Prior experience in Banking Domain is an advantage. ? Prior experience in agile delivery method is an advantage. ? Excellent understanding of technology life cycles and the concepts and practices required to build big data solutions. ? Familiar with Data Warehouse concepts andChange Data Capture (CDC and SCD Types) ? Python skillset, Micro services understanding is an added advantage. ? Good Knowledge and experience in any RDBMS and Data Lake is must. ? Proficiency in a programming language such as Java, Python, Spark Programming. ? Proficiency with notebook like Jupyter. ? Experience with cloud computing platforms (e.g., AWS, GCP) ? Experience with machine learning and artificial intelligence ? Hive and other components of Hadoop eco system. ? Knowledge of Clouderadistribution framework, Airflow Workflow. ? Familiarity with Data warehouse concepts and Database. ? understanding of technology life cycles and the concepts and practices required to build big data solutions. ? Knowledge of CDSW will be Plus.