Big Data Pyspark Developer Infosys

  • company name Infosys
  • working location Office Location
  • job type Full Time

Experience: 2 - 9 years required

Pay:

Salary Information not included

Type: Full Time

Location: BANGALORE null, undefined

Skills: General Skills, Communication, Teamwork

About Infosys

Job Description

Big data-Pyspark Developer-Hadoop

"Skillset required: 1. Excellent knowledge of UNIX/LINUX OS. 2. Knowing of core java is a plus but not mandatory. 3. Good understanding of OS concepts, process management and resource scheduling. 4. Basics of networking, CPU, memory and storage. 5. Good hold of shell scripting" "1. Deploying a hadoop cluster, maintaining a hadoop cluster, adding and removing nodes using cluster monitoring tools like Cloudera Manager, configuring the NameNode high availability and keeping a track of all the running hadoop jobs. 2. Implementing, managing and administering the overall hadoop infrastructure. 3. Knowledge all the components in the Hadoop ecosystem like Apache Spark, Apache Hive, HBase, Kafka, Sqoop, Yarn, Zookeeper etc. 4. Takes care of the day-to-day running of Hadoop clusters. 5. Candidate will have to work closely with the database team, network team, BI team and application teams to make sure that all the big data applications are highly available and performing as expected. 6. Candidate is responsible for capacity planning and estimating the requirements for lowering or increasing the capacity of the hadoop cluster. 7. Ensure that the hadoop cluster is up and running all the time. 8. Monitoring the cluster connectivity and performance. 9. Manage and review Hadoop log files. 10. Backup and recovery tasks 11. Resource and security management 12. Troubleshooting application errors and ensuring that they do not occur again."