Previous Job
Hadoop Administrator - Sunnyvale, CA
Ref No.: 16-04269
Location: Sunnyvale, California
Job Title                    :  Hadoop Administrator
Location                   :  Sunnyvale, CA
Position Type         : Full Time
Salary + benefits (Health Care and PTO Etc.)
US Citizen, Green Card and GC EAD only.
No sponsorship available for this job, direct hire and NO 1099 or C2C
(Please include the expected Salary Range)
Job Description:
Mandatory Technical / Functional Skills
  • BigData domain knowledge
  • Should have good work Experience in HADOOP Administration (3+ Years)
  • Should Have Good Knowledge of Scripting Language like Ruby or Python and Shell Scripting(TCL)
  • Should have knowledge of Linux Administration and Architecture
  • Hadoop Hortonworks upgrade experience
  • Hive,Oozie,Spark,Nagio,Ambari and all BigData related tools expertise
Roles and responsibility
  • Responsible for implementation and ongoing administration of Hadoop infrastructure.
  • Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
  • Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users.
  • Cluster maintenance as well as creation and removal of nodes
  • Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
  • Screen Hadoop cluster job performances and capacity planning
  • Monitor Hadoop cluster connectivity and security
  • Manage and review Hadoop log files.
  • File system management and monitoring.
  • HDFS support and maintenance.
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
  • Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
  • Point of Contact for Vendor escalation
Desirable Technical / Functional Skills
  • General operational expertise such as good troubleshooting skills, understanding of system's capacity, bottlenecks, basics of memory, CPU, OS, storage, and networks.
  • Hadoop skillslike HBase, Hive, Pig, Mahout, etc.
  • The most essential requirements are: They should be able to deploy Hadoop cluster, add and remove nodes, keep track of jobs, monitor critical parts of the cluster, configure name-node high availability, schedule and configure it and take backups.