
Senior Confluent Kafka Lead
Sanrnd, Raleigh, NC, United States
Overview
San R&D Business Solutions LLC | Full time
Columbus, United States | Posted on 04/26/2026
Job Title:
Senior Confluent Kafka Lead / Python Developer
Employment Type:
Contract (C2C)
The Confluent Kafka Lead / Python Developer is responsible for designing, building, and operating enterprise-grade event streaming solutions using Confluent Kafka while developing Python-based producers, consumers, and streaming applications. This role blends deep hands on development with technical leadership, ensuring scalable, reliable, and secure real-time data flows across distributed systems. The position plays a key role in event-driven architecture (EDA), data platform modernization, and real-time analytics initiatives.
Key Responsibilities
Kafka & Streaming Platform Leadership:
Lead the design and implementation of enterprise Kafka and Confluent Platform solutions (Kafka, Schema Registry, Connect, ksqlDB). Define and enforce topic design, partitioning, retention, and schema evolution standards. Act as technical owner for Kafka clusters across dev, test, and production environments. Drive best practices for high availability, fault tolerance, and scalability.
Python Development (Producers & Consumers):
Design and develop Python-based Kafka producers and consumers using Confluent Kafka Python APIs. Build event-driven microservices and streaming applications in Python. Implement message serialization and schema validation (Avro, JSON, Protobuf). Handle idempotency, retries, back-pressure, and error handling patterns.
Event-Driven Architecture & Integration:
Design event-driven integration patterns bridging microservices, data stores, APIs, and third party systems. Integrate Kafka with downstream consumers such as databases, data lakes, analytics platforms, and search systems. Support real-time pipelines for transactions, telemetry, customer events, and analytics. Collaborate with API, data, and application teams to align event contracts.
Security & Governance:
Implement Kafka security controls including TLS encryption, SASL / OAuth authentication, and ACL-based authorization. Enforce data governance, schema compatibility rules, and event ownership models. Ensure compliance with enterprise security and regulatory standards.
DevOps, Automation & Observability:
Build and maintain CI/CD pipelines for Kafka-related applications and configurations. Use Infrastructure as Code to provision and manage Kafka infrastructure. Implement monitoring and alerting using tools such as Confluent Control Center, Prometheus, Grafana, or cloud-native equivalents. Troubleshoot production streaming issues related to latency, lag, throughput, or data loss.
Technical Leadership:
Serve as Kafka subject matter expert and technical lead. Mentor developers on event-driven design and streaming best practices. Review designs and code for Kafka and Python-based streaming solutions. Partner with architects, SREs, and platform teams on roadmap and capacity planning.
Required Qualifications
Kafka & Streaming Experience:
6–10+ years of experience in software or data engineering. 4+ years of hands-on experience with Apache Kafka and/or Confluent Platform. Strong knowledge of Kafka internals (brokers, partitions, offsets, consumer groups), Schema Registry and schema evolution, and Kafka Connect architectures and connectors.
Python Development:
Strong proficiency in Python for backend and streaming development. Experience building production-grade services using Python frameworks and libraries. Familiarity with async processing, multithreading, or stream processing patterns is a plus.
Cloud & DevOps:
Experience deploying Kafka and applications in cloud or hybrid environments (AWS, Azure, GCP). CI/CD pipeline experience (GitHub Actions, Jenkins, GitLab, Azure DevOps). Infrastructure-as-Code experience (Terraform, CloudFormation, ARM/Bicep). Containerization experience (Docker, Kubernetes) preferred.
Preferred Qualifications
Experience with ksqlDB, Kafka Streams, or stream processing frameworks (Flink, Spark Streaming).
Exposure to event sourcing or CQRS patterns.
Integration of Kafka with data lakes, warehouses, and analytics platforms.
Confluent or cloud platform certifications.
Experience supporting high-throughput, low-latency systems.
Soft Skills & Leadership
Strong communication skills across engineering and stakeholder teams.
Ability to translate business use cases into event-driven technical solutions.
Comfortable acting as both hands-on developer and technical lead.
Experience influencing architecture and standards across teams.
#J-18808-Ljbffr
San R&D Business Solutions LLC | Full time
Columbus, United States | Posted on 04/26/2026
Job Title:
Senior Confluent Kafka Lead / Python Developer
Employment Type:
Contract (C2C)
The Confluent Kafka Lead / Python Developer is responsible for designing, building, and operating enterprise-grade event streaming solutions using Confluent Kafka while developing Python-based producers, consumers, and streaming applications. This role blends deep hands on development with technical leadership, ensuring scalable, reliable, and secure real-time data flows across distributed systems. The position plays a key role in event-driven architecture (EDA), data platform modernization, and real-time analytics initiatives.
Key Responsibilities
Kafka & Streaming Platform Leadership:
Lead the design and implementation of enterprise Kafka and Confluent Platform solutions (Kafka, Schema Registry, Connect, ksqlDB). Define and enforce topic design, partitioning, retention, and schema evolution standards. Act as technical owner for Kafka clusters across dev, test, and production environments. Drive best practices for high availability, fault tolerance, and scalability.
Python Development (Producers & Consumers):
Design and develop Python-based Kafka producers and consumers using Confluent Kafka Python APIs. Build event-driven microservices and streaming applications in Python. Implement message serialization and schema validation (Avro, JSON, Protobuf). Handle idempotency, retries, back-pressure, and error handling patterns.
Event-Driven Architecture & Integration:
Design event-driven integration patterns bridging microservices, data stores, APIs, and third party systems. Integrate Kafka with downstream consumers such as databases, data lakes, analytics platforms, and search systems. Support real-time pipelines for transactions, telemetry, customer events, and analytics. Collaborate with API, data, and application teams to align event contracts.
Security & Governance:
Implement Kafka security controls including TLS encryption, SASL / OAuth authentication, and ACL-based authorization. Enforce data governance, schema compatibility rules, and event ownership models. Ensure compliance with enterprise security and regulatory standards.
DevOps, Automation & Observability:
Build and maintain CI/CD pipelines for Kafka-related applications and configurations. Use Infrastructure as Code to provision and manage Kafka infrastructure. Implement monitoring and alerting using tools such as Confluent Control Center, Prometheus, Grafana, or cloud-native equivalents. Troubleshoot production streaming issues related to latency, lag, throughput, or data loss.
Technical Leadership:
Serve as Kafka subject matter expert and technical lead. Mentor developers on event-driven design and streaming best practices. Review designs and code for Kafka and Python-based streaming solutions. Partner with architects, SREs, and platform teams on roadmap and capacity planning.
Required Qualifications
Kafka & Streaming Experience:
6–10+ years of experience in software or data engineering. 4+ years of hands-on experience with Apache Kafka and/or Confluent Platform. Strong knowledge of Kafka internals (brokers, partitions, offsets, consumer groups), Schema Registry and schema evolution, and Kafka Connect architectures and connectors.
Python Development:
Strong proficiency in Python for backend and streaming development. Experience building production-grade services using Python frameworks and libraries. Familiarity with async processing, multithreading, or stream processing patterns is a plus.
Cloud & DevOps:
Experience deploying Kafka and applications in cloud or hybrid environments (AWS, Azure, GCP). CI/CD pipeline experience (GitHub Actions, Jenkins, GitLab, Azure DevOps). Infrastructure-as-Code experience (Terraform, CloudFormation, ARM/Bicep). Containerization experience (Docker, Kubernetes) preferred.
Preferred Qualifications
Experience with ksqlDB, Kafka Streams, or stream processing frameworks (Flink, Spark Streaming).
Exposure to event sourcing or CQRS patterns.
Integration of Kafka with data lakes, warehouses, and analytics platforms.
Confluent or cloud platform certifications.
Experience supporting high-throughput, low-latency systems.
Soft Skills & Leadership
Strong communication skills across engineering and stakeholder teams.
Ability to translate business use cases into event-driven technical solutions.
Comfortable acting as both hands-on developer and technical lead.
Experience influencing architecture and standards across teams.
#J-18808-Ljbffr