Manages and prioritizes multiple tasks and assignments for the team including tracking tasks and tickets in system(s) of record, ensuring enterprise service level agreements are met, and participating in defect, incident, problem, and change review meetings and release activities
Provides point of escalation for team members, peers, and partner teams, and shares the leadership Point of Contact for triage of high priority incidents or consulting engagements
Performs all responsibilities of team members from applicable teams
Works with appropriate teams to ensure alerts, monitoring, dashboards, and processes are established appropriately for new projects and initiatives
Participates and helps with resource interviews and onboarding processes
Performs access reviews, conducts compliance activities, and serves as a delegate for administrative processes
Creates the on call rotation schedule and ensures support for off hours, weekends, and release windows
Skills:
Collaboration
Influence
Production Support
Risk Management
Solution Design
Analytical Thinking
Architecture
Innovative Thinking
Result Orientation
Stakeholder Management
Adaptability
Automation
DevOps Practices
Project Management
Solution Delivery Process
Required Qualifications:
10+ years in technology experience.
Hadoop System Administration skills with at least 8+ years of experience of handling Cloudera distribution of Hadoop or Horton works.
Experience on evaluating the Proof of Concept on new Hadoop tools and related technologies. Formulate and design systems scope and objectives for applications and the development of information technology projects.
Extensive experience in big data tools: Hadoop, Hive, Impala and Spark.
Proficiency in Scala, Python, SQL, and PySpark. Strong experience and triaging skills on the hive, Spark data analysis & Impala.
Experience with stream-processing systems: Kafka, Spark-Streaming, etc.
Experience on configuring high availability of the Cloudera Hadoop services to achieve never down situation on Hadoop components.
Good understanding of Operating System (OS) concepts, process management, capacity planning and resource scheduling.
Basics of Infrastructure networking, Memory, Racks and storage to help building the clusters for new client onboarding.
Proficiency in managing and communicating data warehouse plans to internal clients.
Great knowledge on Resiliency management for data and metadata on the Hadoop cluster via tools like Back-up & Disaster Recovery (BDR) & any other real time replication mechanism.
Experience with building processes supporting data transformation, data structures, metadata, dependency, and workload management.
Strong analytic skills related to working with structured & unstructured dataset. A successful history of manipulating, processing, and extracting value from large, disconnected datasets.
Team Leadership: Lead and mentor a team of Operational engineers. Coordinate with other teams to manage and prioritize projects. Drive strategy, roadmap, and execution of key data engineering initiatives.
Stakeholder Management: Collaborate with stakeholders across the organization to understand their data needs and deliver solution.
Desired Skills
Self-Motivated; ability to multi-task efficiently.
Strong communication skills, backed up with strong skills on data lake concepts.
Ability to handle production triages coordinating with all the required parties like support groups, users, vendor etc.
Must have idea on change management routines to implement solutions calculating the risks.
Strong SQL skills on any RDBMS like Oracle, Teradata, MySQL
Job Scheduling tools : Autosys.
Business Intelligence (BI) tools : Tableau, Statistical Analysis System (SAS)