Accepting Applications
Full-time
On-site
Posted 2 weeks, 6 days ago
3 views
0 applications
Job Description
**Job Title:**
**Location: Remote (Pak) / Hybrid (Riyadh, KSA)**
**Employment Type: Full\-Time / Contract\-to\-Hire**
**Experience Level: 6\+ years (Senior)**
**Role Summary**
As a **Senior Confluent Kafka Engineer**, you'll architect, design, setup, install, implement and optimize high\-throughput data streaming solutions using Confluent Platform and Apache Kafka. You'll lead a team of engineers in delivering production\-grade pipelines, ensuring scalability, reliability, and security. This role involves hands\-on development, mentoring, and collaborating with data architects, DevOps, and stakeholders to implement event\-driven architectures. You'll champion best practices in real\-time data processing, from proof\-of\-concepts to enterprise deployments, including full lifecycle management from installation to optimization.
**Key Responsibilities*** **Architecture \& Design:** Lead the design of scalable Kafka clusters and Confluent\-based ecosystems (e.g., Kafka Streams, ksqlDB, Schema Registry, Connect) for on\-prem, hybrid, and multi\-cloud (GCP) environments.
* **Implementation \& Development:** Build and maintain real\-time data pipelines, integrations, and microservices using Kafka producers/consumers; integrate with tools like Flink, Spark, or ML frameworks for advanced analytics.
* **Installation \& Setup:** Oversee the end\-to\-end installation and initial configuration of Confluent Platform and Apache Kafka clusters, including:
+ Deploying Confluent Enterprise/Community editions on Kubernetes (via Helm/Operator), bare\-metal servers, or managed cloud services (e.g., Confluent Cloud, GCP).
+ Configuring brokers, ZooKeeper/KRaft mode, topics, partitions, replication factors, and security settings (e.g., SSL/TLS, SASL, ACLs) using Ansible, Terraform, or Confluent CLI.
+ Setting up auxiliary components like Schema Registry, Kafka Connect clusters, and monitoring agents (e.g., JMX exporters) with automated scripts for reproducible environments.
+ Performing initial health checks, load testing (e.g., with Kafka's performance tools), and integration with existing infrastructure (e.g., VPC peering, load balancers).
* **Operations \& Maintenance:** Oversee monitoring, troubleshooting, performance tuning, and lifecycle management (upgrades, patching) of Kafka/Confluent instances; implement DevSecOps practices for CI/CD pipelines.
* **Team Leadership:** Mentor junior engineers, conduct code reviews, and drive technical proofs\-of\-concept (POCs); gather requirements and define standards for Kafka as a managed service (e.g., access controls, documentation).
* **Optimization \& Innovation:** Ensure high availability (\>99\.99%), fault tolerance, and cost\-efficiency; explore emerging features like Kafka Tiered Storage or Confluent Cloud integrations for AI workloads.
* **Collaboration \& Delivery:** Partner with cross\-functional teams (data engineers, architects, product owners) to align streaming solutions with business goals; provide thought leadership on event\-driven patterns.
* **Security \& Compliance:** Implement RBAC, encryption, and auditing; conduct root\-cause analysis for incidents and ensure GDPR/HIPAA compliance in data flows.
**Required Qualifications \& Skills*** Bachelor's/Master's in Computer Science, Engineering, or related; certifications like Confluent Developer/Administrator a plus.
* 5\+ years hands\-on with Apache Kafka \& Confluent Platform (Cloud/Enterprise editions).
* Proficiency in Java/Scala/Python; Kafka Streams/Connect/ksqlDB; Schema Registry; REST/gRPC APIs.
* Event\-driven/microservices design; data pipeline optimization; handling high\-volume streams (TB/day scale).
* Expertise in containerization (Docker/Kubernetes); CI/CD (Jenkins/GitHub Actions); Terraform/Ansible for IaC.
* Multi\-cloud experience (AWS, GCP, Azure); monitoring tools (Prometheus, Grafana, Confluent Control Center).
* Experience with streaming integrations (e.g., Flink, Spark Streaming for CDC).
* Contributions to open\-source Kafka projects or publications on streaming architectures.
* Knowledge of AI/ML data pipelines (e.g., Kafka \+ TensorFlow/PyTorch).
* Familiarity with observability tools and security (OAuth, Kerberos).
* Strong problem\-solving, communication, and leadership; experience leading POCs and cross\-team projects.
* Agile/Scrum leadership in fast\-paced environments.
* Experience in client facing roles and leading teams.
9KCrLpPJlE
Login to Apply
Don't have an account? Register