Turn Your Vehicle Into a Smart Earning Asset

While you’re not driving your car or bike, it can still be working for you. MOTOSHARE helps you earn passive income by connecting your vehicle with trusted renters in your city.

🚗 You set the rental price
🔐 Secure bookings with verified renters
📍 Track your vehicle with GPS integration
💰 Start earning within 48 hours

Join as a Partner Today

It’s simple, safe, and rewarding. Your vehicle. Your rules. Your earnings.

Kafka & Confluent Terminology – Complete Glossary


I’ll cover Apache Kafka core terms, Confluent Platform extensions, and Confluent Cloud additions.


Kafka & Confluent Terminology – Complete Glossary

🔹 Core Kafka Concepts

  1. Kafka Cluster
    A group of servers (called brokers) working together to store, process, and stream data.
  2. Broker
    A single Kafka server that stores data and serves client requests (produce/consume).
  3. Producer
    An application that sends data (messages) into Kafka topics.
  4. Consumer
    An application that reads data (messages) from Kafka topics.
  5. Consumer Group
    A group of consumers working together to read data from a topic. Kafka ensures each message is processed by only one consumer within the group.
  6. Topic
    A named channel where producers send messages and consumers read messages (like a folder or queue).
  7. Partition
    A topic is divided into slices called partitions. Messages in a partition are ordered. Partitions allow parallelism and scalability.
  8. Offset
    The position of a message in a partition (like a bookmark). Consumers use offsets to track what they’ve read.
  9. Record / Message
    A single unit of data in Kafka. It has:
    • Key (optional, used for partitioning/order)
    • Value (the actual payload)
    • Headers (extra metadata)
  10. Log
    A partition is stored as an append-only log (new messages are always written at the end).
  11. Replication
    Kafka keeps copies of partitions across multiple brokers for fault tolerance.
  12. Leader & Follower
    • Leader: The main replica of a partition that handles all reads/writes.
    • Follower: Copies the leader’s data for backup.
  13. ISR (In-Sync Replicas)
    A set of replicas that are fully caught up with the leader.
  14. Retention Policy
    Defines how long Kafka keeps data (e.g., 7 days, forever, or until size limit).
  15. Compaction
    A cleanup policy that keeps only the latest value per key, deleting older duplicates.
  16. Throughput
    The rate at which Kafka processes messages (messages per second).
  17. Latency
    The time it takes for a message to travel from producer → broker → consumer.

🔹 Kafka Internals

  1. ZooKeeper (Legacy)
    Used in older Kafka versions to manage cluster metadata and leader election. (Being replaced by KRaft).
  2. KRaft (Kafka Raft Metadata mode)
    New architecture where Kafka itself manages metadata, removing the need for ZooKeeper.
  3. Controller
    A special broker responsible for managing partition leaders.
  4. Rebalancing
    When consumers join/leave a group, Kafka redistributes partitions among them.
  5. Coordinator
    The broker responsible for managing a consumer group.
  6. ACL (Access Control List)
    Security rules defining which user/app can access which topic or resource.
  7. Quotas
    Limits on how much data a client can produce/consume to prevent abuse.
  8. Idempotent Producer
    Ensures no duplicate messages are produced even if retries happen.
  9. Exactly-Once Semantics (EOS)
    Guarantee that messages are processed only once, even during failures.
  10. Transactions
    A way to group multiple messages into an atomic unit of work.

🔹 Confluent-Specific Terms

  1. Confluent Platform
    An enterprise distribution of Kafka with additional tools for management, monitoring, and integration.
  2. Confluent Cloud
    A fully managed Kafka service hosted by Confluent on AWS, Azure, or GCP.
  3. Schema Registry
    Stores and enforces schemas (data formats) for messages (e.g., Avro, JSON, Protobuf) to ensure compatibility.
  4. kSQL / ksqlDB
    A SQL-like engine to query, process, and transform Kafka streams in real-time.
  5. Kafka Connect
    A framework to move data in/out of Kafka using connectors (e.g., JDBC, S3, Elasticsearch).
  6. Connector
    A plugin used with Kafka Connect to integrate Kafka with external systems.
    • Source Connector: Pulls data into Kafka.
    • Sink Connector: Pushes data out of Kafka.
  7. Confluent Hub
    A marketplace of prebuilt Kafka connectors.
  8. Confluent Control Center
    A GUI tool for monitoring Kafka clusters, topics, connectors, and schemas.
  9. Replicator
    A Confluent tool to copy topics from one Kafka cluster to another (useful for multi-region).
  10. Confluent REST Proxy
    Allows producing/consuming data using REST APIs instead of Kafka clients.
  11. Confluent RBAC (Role-Based Access Control)
    Fine-grained access control for Kafka resources.
  12. Confluent CLI
    A command-line tool for managing Confluent Cloud clusters, topics, and connectors.
  13. Tiered Storage
    A Confluent feature that offloads older Kafka data to cheaper cloud storage (e.g., S3, GCS).
  14. Cluster Linking
    A Confluent Cloud feature to link clusters across regions/clouds for data replication.
  15. Confluent Cloud Metrics API
    Provides usage and performance metrics for monitoring clusters.

🔹 Stream Processing Terms

  1. Kafka Streams
    A Java library for building real-time streaming applications on top of Kafka.
  2. Stream
    A continuous flow of data records in Kafka.
  3. Stream Processor
    An application that transforms or processes Kafka data in real-time.
  4. Topology
    The workflow (graph of processors) that defines how streams are processed.
  5. State Store
    Local storage used by stream processing apps to maintain state (e.g., counts, aggregations).
  6. Global Store
    A replicated state store available to all stream tasks.
  7. Windowing
    Grouping data by time intervals (e.g., 5-minute sales totals).

🔹 Advanced Kafka Concepts

  1. Reassignment
    Moving partitions across brokers for load balancing.
  2. Throttling
    Slowing down producers/consumers to avoid overwhelming the cluster.
  3. Backpressure
    When consumers can’t keep up with producers, causing slowdowns.
  4. Dead Letter Queue (DLQ)
    A special topic where failed or invalid messages are sent for later debugging.
  5. MirrorMaker 2.0
    Kafka’s built-in tool for replicating data across clusters (open-source equivalent of Confluent Replicator).
  6. Metrics & JMX
    Kafka exposes metrics via JMX for monitoring cluster health.
  7. Log Segment
    Each partition’s log is broken into smaller files called log segments.
  8. Message Key Partitioning
    The method Kafka uses to decide which partition a message goes to (based on key hash).
  9. Rack Awareness
    Kafka spreads replicas across different racks/data centers for reliability.

Subscribe
Notify of
guest
0 Comments
Newest
Oldest Most Voted
Inline Feedbacks
View all comments

Certification Courses

DevOpsSchool has introduced a series of professional certification courses designed to enhance your skills and expertise in cutting-edge technologies and methodologies. Whether you are aiming to excel in development, security, or operations, these certifications provide a comprehensive learning experience. Explore the following programs:

DevOps Certification, SRE Certification, and DevSecOps Certification by DevOpsSchool

Explore our DevOps Certification, SRE Certification, and DevSecOps Certification programs at DevOpsSchool. Gain the expertise needed to excel in your career with hands-on training and globally recognized certifications.

0
Would love your thoughts, please comment.x
()
x