
Job Title: Senior Confluent Kafka Lead / Python Developer
Location: Columbus, OH / Nashville, TN Work Type: Onsite Employment Type: Contract (C2C)Role Overview:
The Confluent Kafka Lead / Python Developer is responsible for designing, building, and operating enterprise-grade event streaming solutions using Confluent Kafka while developing Python-based producers, consumers, and streaming applications. This role blends deep hands on development with technical leadership, ensuring scalable, reliable, and secure real-time data flows across distributed systems. The position plays a key role in event-driven architecture (EDA), data platform modernization, and real-time analytics initiatives. Key Responsibilities: Kafka & Streaming Platform Leadership: • Lead the design and implementation of enterprise Kafka and Confluent Platform solutions (Kafka, Schema Registry, Connect, ksqlDB). • Define and enforce topic design, partitioning, retention, and schema evolution standards. • Act as technical owner for Kafka clusters across dev, test, and production environments. • Drive best practices for high availability, fault tolerance, and scalability. Python Development (Producers & Consumers): • Design and develop Python-based Kafka producers and consumers using Confluent Kafka Python APIs. • Build event-driven microservices and streaming applications in Python. • Implement message serialization and schema validation (Avro, JSON, Protobuf). • Handle idempotency, retries, back-pressure, and error handling patterns. Event-Driven Architecture & Integration: • Design event-driven integration patterns bridging microservices, data stores, APIs, and third party systems. • Integrate Kafka with downstream consumers such as databases, data lakes, analytics platforms, and search systems. • Support real-time pipelines for transactions, telemetry, customer events, and analytics. • Collaborate with API, data, and application teams to align event contracts. Security & Governance: • Implement Kafka security controls: - TLS encryption - SASL / OAuth authentication - ACL-based authorization • Enforce data governance, schema compatibility rules, and event ownership models. • Ensure compliance with enterprise security and regulatory standards. DevOps, Automation & Observability • Build and maintain CI/CD pipelines for Kafka-related applications and configurations. • Use Infrastructure as Code to provision and manage Kafka infrastructure. • Implement monitoring and alerting using tools such as Confluent Control Center, Prometheus, Grafana, or cloud-native equivalents. • Troubleshoot production streaming issues related to latency, lag, throughput, or data loss. Technical Leadership: • Serve as Kafka subject matter expert and technical lead. • Mentor developers on event-driven design and streaming best practices. • Review designs and code for Kafka and Python-based streaming solutions. • Partner with architects, SREs, and platform teams on roadmap and capacity planning. Required Qualifications: Kafka & Streaming Experience: • 6–10+ years of experience in software or data engineering. • 4+ years of hands-on experience with Apache Kafka and/or Confluent Platform. • Strong knowledge of: - Kafka internals (brokers, partitions, offsets, consumer groups) - Schema Registry and schema evolution - Kafka Connect architectures and connectors Python Development: • Strong proficiency in Python for backend and streaming development. • Experience building production-grade services using Python frameworks and libraries. • Familiarity with async processing, multithreading, or stream processing patterns is a plus. Cloud & DevOps: • Experience deploying Kafka and applications in cloud or hybrid environments (AWS, Azure, GCP). • CI/CD pipeline experience (GitHub Actions, Jenkins, GitLab, Azure DevOps). • Infrastructure-as-Code experience (Terraform, CloudFormation, ARM/Bicep). • Containerization experience (Docker, Kubernetes) preferred. Preferred Qualifications: • Experience with ksqlDB, Kafka Streams, or stream processing frameworks (Flink, Spark Streaming). • Exposure to event sourcing or CQRS patterns. • Integration of Kafka with data lakes, warehouses, and analytics platforms. • Confluent or cloud platform certifications. • Experience supporting high-throughput, low-latency systems. Soft Skills & Leadership: • Strong communication skills across engineering and stakeholder teams. • Ability to translate business use cases into event-driven technical solutions. • Comfortable acting as both hands-on developer and technical lead. • Experience influencing architecture and standards across teams.