Apply now »

TechOps-DE-CloudOps-CLOUD COMPUTING -TECHNICAL SUPPORT - Kafka

Location:  Pune
Other locations:  Primary Location Only
Salary: Competitive
Date:  Mar 11, 2025

Job description

Requisition ID:  1589600

At EY, you’ll have the chance to build a career as unique as you are, with the global scale, support, inclusive culture and technology to become the best version of you. And we’re counting on your unique voice and perspective to help EY become even better, too. Join us and build an exceptional experience for yourself, and a better working world for all. 

 

 

 

 

Career Family:TechOps -CloudOps

Role Type:Aws Event Streaming Engineer

 

The opportunity

You will support, monitor, and troubleshoot Kafka-based event streaming platforms in a 24x7 operational environment. You will handle incident response, request fulfillment, log analysis, and execute standard operating procedures (SOPs) to maintain operational stability. Collaboration with L3 teams to escalate and resolve critical issues while ensuring compliance with established support workflows will be a key part of your role.

 

Your key responsibilities

  • Provide L1/L2 operations support for event streaming platforms, ensuring Kafka cluster health and stability
  • Handle incident management, request fulfilment, and escalation through ITSM tools.
  • Analyse logs to debug issues and performance bottlenecks
  • Execute SOPs and runbooks for Event streaming platform topic management, ACL updates, and quota adjustments
  • Collaborate with L3 teams to troubleshoot Event streaming platform issues, consumer lag, and data replication inconsistencies
  • Perform routine maintenance tasks, including log retention management and partition rebalancing
  • Participate in shift handovers, documenting ongoing incidents and pending actions
  • Assist in weekly service review meetings and contribute to monthly governance discussions.

 

Skills and attributes for success

  • Strong understanding of Kafka components, including brokers, producers, consumers, and partitions
  • Familiarity with ITSM processes for incident and request management.
  • Basic scripting skills in Bash or Python for automation and log analysis.
  • Strong analytical and problem-solving skills for diagnosing Kafka-related issues
  • Effective communication and documentation abilities for incident handover and resolution tracking

 

To qualify for the role, you must have

  • 1-3 years of experience in Kafka operations, event streaming, or infrastructure support
  • Experience with monitoring tools like Datadog, OpenTelemetry, or Splunk for Kafka performance analysis
  • Hands-on experience with ServiceNow or similar ITSM tools for managing incidents and service requests
  • Ability to follow structured SOPs and runbooks for Kafka troubleshooting and maintenance
  • Willingness to work in a 24x7 rotational support environment
  • No location constraints

 

Technologies and Tools

Must haves

  • Event Streaming & Messaging: Kafka (Confluent Platform & Confluent Cloud)
  • Monitoring & Observability: OpenTelemetry (OTEL), Datadog
  • Log Analysis & Troubleshooting: Splunk
  • Incident & Request Management: Any ITSM tool (Preference for ServiceNow)
  • Scripting & Automation: Python (for log parsing & troubleshooting automation) or Bash
  • Kafka Performance Monitoring: Understanding Kafka consumer groups, partitioning, offsets, and lag monitoring
     

Good to have

  • Certification – Any Kafka certification , relevant AWS certification 
  • Advanced Log Processing & Event Handling: Cribl (Log processing tool for data streaming)
  • Security & Compliance: IAM role-based authentication & TLS encryption for Kafka
  • Performance Tuning & Debugging: Experience analyzing Kafka producer/consumer throughput, partitioning strategies, tuning batch sizes, and optimizing retention settings
  • Schema Management & Data Governance: Avro, Schema Registry, data serialization/deserialization techniques
  • CI/CD for Streaming Pipelines: Experience with Kafka connectors, deploying stream processing applications using ArgoCD, Jenkins, or GitHub Actions

 

What we look for

  • Problem solvers with a proactive approach to troubleshooting and optimization.
  • Team players who can collaborate effectively in a remote or hybrid work environment.
  • Detail-oriented professionals with strong documentation skills.

 

What we offer

EY Global Delivery Services (GDS) is a dynamic and truly global delivery network. We work across six locations – Argentina, China, India, the Philippines, Poland and the UK – and with teams from all EY service lines, geographies and sectors, playing a vital role in the delivery of the EY growth strategy. From accountants to coders to advisory consultants, we offer a wide variety of fulfilling career opportunities that span all business disciplines. In GDS, you will collaborate with EY teams on exciting projects and work with well-known brands from across the globe. We’ll introduce you to an ever-expanding ecosystem of people, learning, skills and insights that will stay with you throughout your career.

  • Continuous learning: You’ll develop the mindset and skills to navigate whatever comes next.
  • Success as defined by youWe’ll provide the tools and flexibility, so you can make a meaningful impact, your way.
  • Transformative leadershipWe’ll give you the insights, coaching and confidence to be the leader the world needs.
  • Diverse and inclusive culture: You’ll be embraced for who you are and empowered to use your voice to help others find theirs.

 

EY | Building a better working world 


 
EY exists to build a better working world, helping to create long-term value for clients, people and society and build trust in the capital markets.  


 
Enabled by data and technology, diverse EY teams in over 150 countries provide trust through assurance and help clients grow, transform and operate.  


 
Working across assurance, consulting, law, strategy, tax and transactions, EY teams ask better questions to find new answers for the complex issues facing our world today.  

Apply now »