MyPage is a personalized page based on your interests.The page is customized to help you to find content that matters you the most.


I'm not curious
2

Systems Integration Advisor

Location Noida, India
Posted 19-August-2021
Description
Looking for a Hadoop Admin with strong experience in Hadoop Administration. Should have excellent customer management and communications skills. Strong experience in Agile / Scrum and ADO is a must.

Role Responsibilities

Responsible for implementation and ongoing administration of Hadoop infrastructure.
Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users.
Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise, Dell Open Manage and other tools.
Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
Screen Hadoop cluster job performances and capacity planning
Monitor Hadoop cluster connectivity and security
Manage and review Hadoop log files.
File system management and monitoring.
HDFS support and maintenance.
Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
Point of Contact for Vendor escalation
Data modelling, design & implementation based on recognized standards.
Software installation and configuration.
Database backup and recovery.
Database connectivity and security.
Performance monitoring and tuning.
Disk space management.
Software patches and upgrades.
Automate manual tasks.

Technical experience and skills

General operational expertise such as good troubleshooting skills, understanding of systems capacity, bottlenecks, basics of memory, CPU, OS, storage, and networks.
Hadoop skillslike HBase, Hive, Pig, Mahout, etc.
The most essential requirements are: They should be able to deploy Hadoop cluster, add and remove nodes, keep track of jobs, monitor critical parts of the cluster, configure name-node high availability, schedule and configure it and take backups.
Good knowledge of Linux as Hadoop runs on Linux.
Familiarity with open source configuration management and deployment tools such as Puppet or Chef and Linux scripting.
Knowledge of Troubleshooting Core Java Applications is a plus.


Qualifications and other skills/ competencies:


Degree in Computer Science / Engineering
Excellent communication and interaction skills
Self-initiator, interact, coordinate and collaborate with teams


Job Segment: Computer Science, Database, Developer, Java, Linux, Technology

 
Awards & Accolades for MyTechLogy
Winner of
REDHERRING
Top 100 Asia
Finalist at SiTF Awards 2014 under the category Best Social & Community Product
Finalist at HR Vendor of the Year 2015 Awards under the category Best Learning Management System
Finalist at HR Vendor of the Year 2015 Awards under the category Best Talent Management Software
Hidden Image Url