From 4c04f4f0ef92c41ef3e14e7ef519785cbca8579e Mon Sep 17 00:00:00 2001 From: AnirudhPI Date: Wed, 20 Nov 2024 18:21:53 -0700 Subject: [PATCH] docker-compose.yml fix --- preprocessing/docker-compose.yml | 39 ++++++++++++++++++-------------- preprocessing/pcap_processor.py | 5 ++-- 2 files changed, 25 insertions(+), 19 deletions(-) diff --git a/preprocessing/docker-compose.yml b/preprocessing/docker-compose.yml index 008ded4..89b39e5 100644 --- a/preprocessing/docker-compose.yml +++ b/preprocessing/docker-compose.yml @@ -4,7 +4,9 @@ services: zookeeper: image: confluentinc/cp-zookeeper:latest networks: - - kafka_network + kafka_network: + aliases: + - zookeeper deploy: replicas: 1 restart_policy: @@ -13,11 +15,6 @@ services: ZOOKEEPER_CLIENT_PORT: 2181 ports: - "2181:2181" - healthcheck: - test: ["CMD", "nc", "-z", "localhost", "2181"] - interval: 10s - timeout: 5s - retries: 5 kafka: image: confluentinc/cp-kafka:latest @@ -26,34 +23,37 @@ services: environment: KAFKA_ZOOKEEPER_CONNECT: zookeeper:2181 KAFKA_LISTENERS: PLAINTEXT://0.0.0.0:9092 - KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://kafka:9092,PLAINTEXT://localhost:9092 - KAFKA_LISTENER_SECURITY_PROTOCOL: PLAINTEXT + KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://kafka:9092 + KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: PLAINTEXT:PLAINTEXT KAFKA_BROKER_ID: 1 + KAFKA_MESSAGE_MAX_BYTES: 200000000 + KAFKA_REPLICA_FETCH_MAX_BYTES: 200000000 networks: - - kafka_network + kafka_network: + aliases: + - kafka ports: - "9092:9092" + volumes: + - kafka_data:/var/lib/kafka/data deploy: replicas: 1 restart_policy: condition: on-failure - healthcheck: - test: ["CMD", "nc", "-z", "localhost", "9092"] - interval: 10s - timeout: 5s - retries: 5 pcap_streamer: image: levenshtein/streamer_test3:latest depends_on: - kafka networks: - - kafka_network + kafka_network: + aliases: + - pcap_streamer volumes: - - "/host_mnt/c/Users/akash/storage/Asu/sem3/dds/project:/data/pcap" + - "/host_mnt/c/Users/akash/storage/Asu/sem3/dds/project:/data/pcap" environment: PCAP_FILE: /data/pcap/202310081400.pcap - command: ["sh", "-c", "sleep 10 && python /app/pcap_processor.py -f /data/pcap/202310081400.pcap -s --stream_size 1000"] + command: ["sh", "-c", "sleep 30 && python /app/pcap_processor.py -f /data/pcap/202310081400.pcap -s --stream_size 1000"] deploy: replicas: 1 restart_policy: @@ -62,3 +62,8 @@ services: networks: kafka_network: driver: overlay + attachable: true + +volumes: + kafka_data: + driver: local \ No newline at end of file diff --git a/preprocessing/pcap_processor.py b/preprocessing/pcap_processor.py index 8e5cdaa..09c556a 100644 --- a/preprocessing/pcap_processor.py +++ b/preprocessing/pcap_processor.py @@ -17,7 +17,8 @@ class KafkaClient: self.topic_name = topic_name if mode == 'producer': self.client = KafkaProducer( - bootstrap_servers=['localhost:9092'], + bootstrap_servers=['kafka:9092'], + max_request_size = 200000000, api_version=(0,11,5), value_serializer=lambda x: json.dumps(x).encode('utf-8')) elif mode == 'consumer' and topic_name is not None: @@ -31,7 +32,7 @@ class KafkaClient: # Kafka Configuration KAFKA_TOPIC = 'pcap_stream' -KAFKA_SERVER = 'localhost:9092' # Adjust to your Kafka server +KAFKA_SERVER = 'kafka:9092' # Adjust to your Kafka server #KAFKA_SERVER = 'kafka_service:9092' # Initialize Kafka Producer