← Home

šŸ“Ø

⌘K
šŸ¤–
Claude Code AI Tools
šŸ¤—
Hugging Face AI Tools
🦜
LangChain AI Tools
🧠
Keras AI Tools
šŸ¦™
Ollama AI Tools
šŸ
Python Programming Languages
🟨
JavaScript Programming Languages
šŸ”·
TypeScript Programming Languages
āš›ļø
React Programming Languages
🐹
Go Programming Languages
šŸ¦€
Rust Programming Languages
šŸ“Š
MATLAB Programming Languages
šŸ—„ļø
SQL Programming Languages
āš™ļø
C/C++ Programming Languages
ā˜•
Java Programming Languages
🟣
C# Programming Languages
šŸŽ
Swift Programming Languages
🟠
Kotlin Programming Languages
ā–²
Next.js Programming Languages
šŸ’š
Vue.js Programming Languages
šŸ”„
Svelte Programming Languages
šŸŽØ
Tailwind CSS Programming Languages
šŸ’š
Node.js Programming Languages
🌐
HTML Programming Languages
šŸŽØ
CSS/SCSS Programming Languages
🐘
PHP Programming Languages
šŸ’Ž
Ruby Programming Languages
šŸ”“
Scala Programming Languages
šŸ“Š
R Programming Languages
šŸŽÆ
Dart Programming Languages
šŸ’§
Elixir Programming Languages
šŸŒ™
Lua Programming Languages
🐪
Perl Programming Languages
šŸ…°ļø
Angular Programming Languages
šŸš‚
Express.js Programming Languages
🐱
NestJS Programming Languages
šŸ›¤ļø
Ruby on Rails Programming Languages
ā—¼ļø
GraphQL Programming Languages
🟪
Haskell Programming Languages
šŸ’š
Nuxt.js Programming Languages
šŸ”·
SolidJS Programming Languages
⚔
htmx Programming Languages
šŸ’»
VS Code Development Tools
🧠
PyCharm Development Tools
šŸ““
Jupyter Development Tools
🧠
IntelliJ IDEA Development Tools
šŸ’š
Neovim Development Tools
šŸ”®
Emacs Development Tools
šŸ”€
Git DevOps & CLI
🐳
Docker DevOps & CLI
ā˜øļø
Kubernetes DevOps & CLI
ā˜ļø
AWS CLI DevOps & CLI
šŸ”„
GitHub Actions DevOps & CLI
🐧
Linux Commands DevOps & CLI
šŸ’»
Bash Scripting DevOps & CLI
🌐
Nginx DevOps & CLI
šŸ“
Vim DevOps & CLI
šŸ”Ø
Makefile DevOps & CLI
🧪
Pytest DevOps & CLI
🪟
Windows DevOps & CLI
šŸ“¦
Package Managers DevOps & CLI
šŸŽ
macOS DevOps & CLI
šŸ—ļø
Terraform DevOps & CLI
šŸ”§
Ansible DevOps & CLI
āŽˆ
Helm DevOps & CLI
šŸ”Ø
Jenkins DevOps & CLI
šŸ”„
Prometheus DevOps & CLI
šŸ“Š
Grafana DevOps & CLI
šŸ’»
Zsh DevOps & CLI
🐟
Fish Shell DevOps & CLI
šŸ’™
PowerShell DevOps & CLI
šŸ”„
Argo CD DevOps & CLI
šŸ”€
Traefik DevOps & CLI
ā˜ļø
Azure CLI DevOps & CLI
ā˜ļø
Google Cloud CLI DevOps & CLI
šŸ“Ÿ
tmux DevOps & CLI
šŸ”§
jq DevOps & CLI
āœ‚ļø
sed DevOps & CLI
šŸ“Š
awk DevOps & CLI
🌊
Apache Airflow DevOps & CLI
šŸ”¢
NumPy Databases & Data
🐼
Pandas Databases & Data
šŸ”„
PyTorch Databases & Data
🧠
TensorFlow Databases & Data
šŸ“ˆ
Matplotlib Databases & Data
🐘
PostgreSQL Databases & Data
🐬
MySQL Databases & Data
šŸƒ
MongoDB Databases & Data
šŸ”“
Redis Databases & Data
šŸ”
Elasticsearch Databases & Data
šŸ¤–
Scikit-learn Databases & Data
šŸ‘ļø
OpenCV Databases & Data
⚔
Apache Spark Databases & Data
🪶
SQLite Databases & Data
⚔
Supabase Databases & Data
šŸ”µ
Neo4j Databases & Data
šŸ“Ø
Apache Kafka Databases & Data
🐰
RabbitMQ Databases & Data
šŸ”¤
Regex Utilities
šŸ“
Markdown Utilities
šŸ“„
LaTeX Utilities
šŸ”
SSH & GPG Utilities
🌐
curl & HTTP Utilities
šŸ“œ
reStructuredText Utilities
šŸš€
Postman Utilities
šŸŽ¬
FFmpeg Utilities
šŸ–¼ļø
ImageMagick Utilities
šŸ”
ripgrep Utilities
šŸ”
fzf Utilities
šŸ“—
Microsoft Excel Office Applications
šŸ“˜
Microsoft Word Office Applications
šŸ“™
Microsoft PowerPoint Office Applications
šŸ“
Hancom Hangul Hancom Office
šŸ“½ļø
Hancom Hanshow Hancom Office
šŸ“Š
Hancom Hancell Hancom Office
šŸ“„
Google Docs Google Workspace
šŸ“Š
Google Sheets Google Workspace
šŸ“½ļø
Google Slides Google Workspace
šŸ”Œ
Cadence Virtuoso EDA & Hardware
āš™ļø
Synopsys EDA EDA & Hardware
šŸ’Ž
Verilog & VHDL EDA & Hardware
⚔
LTSpice EDA & Hardware
šŸ”§
KiCad EDA & Hardware
šŸ“
Notion Productivity
šŸ’Ž
Obsidian Productivity
šŸ’¬
Slack Productivity
šŸŽ®
Discord Productivity
šŸŽØ
Figma Design Tools
šŸ“˜
Confluence Atlassian
šŸ“‹
Jira Atlassian
šŸƒ
Jest Testing
⚔
Vitest Testing
šŸŽ­
Playwright Testing
🌲
Cypress Testing
🌐
Selenium Testing
šŸ’™
Flutter Mobile Development
šŸ“±
React Native Mobile Development
šŸŽ
SwiftUI Mobile Development
šŸ“±
Expo Mobile Development
šŸ
Django Web Frameworks
⚔
FastAPI Web Frameworks
šŸŒ¶ļø
Flask Web Frameworks
šŸƒ
Spring Boot Web Frameworks
šŸø
Gin Web Frameworks
⚔
Vite Build Tools
šŸ“¦
Webpack Build Tools
⚔
esbuild Build Tools
🐘
Gradle Build Tools
🪶
Maven Build Tools
šŸ”§
CMake Build Tools
šŸŽ®
Unity Game Development
šŸ¤–
Godot Game Development
šŸ”Œ
Arduino Embedded & IoT
šŸ”
Nmap Security
šŸ•
Datadog Monitoring
šŸ“–
Swagger/OpenAPI Documentation
No results found
EN KO

CLI Commands

Topics

kafka-topics.sh --list --bootstrap-server localhost:9092 List topics
kafka-topics.sh --create --topic my-topic --partitions 3 --replication-factor 1 --bootstrap-server localhost:9092 Create topic
kafka-topics.sh --describe --topic my-topic --bootstrap-server localhost:9092 Describe topic
kafka-topics.sh --delete --topic my-topic --bootstrap-server localhost:9092 Delete topic
kafka-topics.sh --alter --topic my-topic --partitions 5 --bootstrap-server localhost:9092 Add partitions

Producer & Consumer

kafka-console-producer.sh --topic my-topic --bootstrap-server localhost:9092 Console producer
kafka-console-consumer.sh --topic my-topic --from-beginning --bootstrap-server localhost:9092 Console consumer
kafka-console-consumer.sh --topic my-topic --group my-group --bootstrap-server localhost:9092 Consumer with group
kafka-consumer-groups.sh --list --bootstrap-server localhost:9092 List consumer groups
kafka-consumer-groups.sh --describe --group my-group --bootstrap-server localhost:9092 Describe group

Producer API

Java Producer

Basic producer
Properties props = new Properties();
props.put("bootstrap.servers", "localhost:9092");
props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer");
props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer");

KafkaProducer<String, String> producer = new KafkaProducer<>(props);

// Send message
ProducerRecord<String, String> record = new ProducerRecord<>("my-topic", "key", "value");
producer.send(record);

// With callback
producer.send(record, (metadata, exception) -> {
    if (exception == null) {
        System.out.println("Sent to: " + metadata.partition());
    }
});

producer.close();
Producer config
// Reliability
props.put("acks", "all");                    // Wait for all replicas
props.put("retries", 3);                     // Retry on failure
props.put("enable.idempotence", true);       // Exactly-once

// Performance
props.put("batch.size", 16384);              // Batch size bytes
props.put("linger.ms", 5);                   // Wait before sending
props.put("buffer.memory", 33554432);        // Buffer size
props.put("compression.type", "snappy");     // Compression

Consumer API

Java Consumer

Basic consumer
Properties props = new Properties();
props.put("bootstrap.servers", "localhost:9092");
props.put("group.id", "my-group");
props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");

KafkaConsumer<String, String> consumer = new KafkaConsumer<>(props);
consumer.subscribe(Arrays.asList("my-topic"));

while (true) {
    ConsumerRecords<String, String> records = consumer.poll(Duration.ofMillis(100));
    for (ConsumerRecord<String, String> record : records) {
        System.out.printf("offset=%d, key=%s, value=%s%n",
            record.offset(), record.key(), record.value());
    }
}
Consumer config
// Offset management
props.put("auto.offset.reset", "earliest");  // earliest, latest
props.put("enable.auto.commit", false);      // Manual commit

// Manual commit
consumer.commitSync();
consumer.commitAsync();

// Seek to offset
consumer.seek(partition, offset);
consumer.seekToBeginning(partitions);
consumer.seekToEnd(partitions);

// Assign specific partitions
consumer.assign(Arrays.asList(new TopicPartition("topic", 0)));

Kafka Streams

Streams API

Stream processing
Properties props = new Properties();
props.put(StreamsConfig.APPLICATION_ID_CONFIG, "my-stream-app");
props.put(StreamsConfig.BOOTSTRAP_SERVERS_CONFIG, "localhost:9092");

StreamsBuilder builder = new StreamsBuilder();

// Read from topic
KStream<String, String> source = builder.stream("input-topic");

// Transform
KStream<String, String> transformed = source
    .filter((key, value) -> value.length() > 5)
    .mapValues(value -> value.toUpperCase());

// Write to topic
transformed.to("output-topic");

// Start
KafkaStreams streams = new KafkaStreams(builder.build(), props);
streams.start();
Aggregations
// Group and count
KTable<String, Long> counts = source
    .groupByKey()
    .count();

// Windowed aggregation
KTable<Windowed<String>, Long> windowedCounts = source
    .groupByKey()
    .windowedBy(TimeWindows.of(Duration.ofMinutes(5)))
    .count();

// Join streams
KStream<String, String> joined = stream1.join(
    stream2,
    (v1, v2) -> v1 + "-" + v2,
    JoinWindows.of(Duration.ofMinutes(5))
);

Docker Setup

Docker Compose

docker-compose.yml
version: "3"
services:
  zookeeper:
    image: confluentinc/cp-zookeeper:latest
    environment:
      ZOOKEEPER_CLIENT_PORT: 2181

  kafka:
    image: confluentinc/cp-kafka:latest
    depends_on:
      - zookeeper
    ports:
      - "9092:9092"
    environment:
      KAFKA_BROKER_ID: 1
      KAFKA_ZOOKEEPER_CONNECT: zookeeper:2181
      KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://localhost:9092
      KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 1