initial commit

This commit is contained in:
2025-08-20 17:34:20 +02:00
commit ca4aa96565
7 changed files with 2450 additions and 0 deletions

1
.gitignore vendored Normal file
View File

@@ -0,0 +1 @@
/target

2246
Cargo.lock generated Normal file

File diff suppressed because it is too large Load Diff

19
Cargo.toml Normal file
View File

@@ -0,0 +1,19 @@
[package]
name = "kafka-producer"
version = "0.1.0"
edition = "2024"
[dependencies]
clap = { version = "4", features = ["derive"] }
csv = "1"
# Serialization / parsing
serde = { version = "1", features = ["derive"] }
serde_json = "1"
serde_yaml2 = "0.1"
rdkafka = { version = "0.38", features = ["tokio"] }
schema_registry_converter = { version = "4", features = ["avro"] }
apache-avro = "0.19"
anyhow = "1"
tokio = { version = "1", features = ["full"] }

11
config.yaml Normal file
View File

@@ -0,0 +1,11 @@
kafka:
brokers: "localhost:9092"
security:
protocol: "PLAINTEXT"
schemas:
users:
file: "schemas/stores.avsc"
orders:
file: "schemas/link.avsc"

19
src/config.rs Normal file
View File

@@ -0,0 +1,19 @@
use serde::Deserialize;
use std::collections::HashMap;
#[derive(Debug, Deserialize)]
pub struct AppConfig {
pub kafka: KafkaConfig,
pub schemas: HashMap<String, SchemaConfig>,
}
#[derive(Debug, Deserialize)]
pub struct KafkaConfig {
pub brokers: String,
pub schema_registry: String,
}
#[derive(Debug, Deserialize)]
pub struct SchemaConfig {
pub file: String,
}

69
src/main.rs Normal file
View File

@@ -0,0 +1,69 @@
mod registry;
mod config;
use clap::Parser;
use csv::ReaderBuilder;
use schema_registry_converter::async_impl::avro::AvroEncoder;
use schema_registry_converter::async_impl::schema_registry::SrSettings;
use serde::Deserialize;
use std::fs;
use crate::config::AppConfig;
use crate::registry::{ChatMessage, KafkaProducer};
/// CLI tool to send CSV rows as Avro messages to Kafka
#[derive(Parser, Debug)]
#[command(author, version, about, long_about = None)]
struct Args {
/// Path to the config file (YAML)
#[arg(short, long, default_value = "config.yaml")]
config: String,
/// Path to the CSV file
#[arg(short, long)]
csv: String,
/// Kafka topic to produce to
#[arg(short, long)]
topic: String,
/// Schema Registry URL (Apicurio or Confluent compatible)
#[arg(short = 'r', long, default_value = "http://localhost:8080/apis/registry/v2/groups/default/subjects")]
registry_url: String,
}
#[derive(Debug, Deserialize)]
struct User {
id: i32,
name: String,
}
#[tokio::main]
async fn main() -> anyhow::Result<()> {
let args = Args::parse();
// Charger la configuration
let cfg_str = fs::read_to_string(&args.config).expect("Empty config");
let cfg: AppConfig = serde_yaml2::from_str(&cfg_str).unwrap();
// 1. Kafka Producer
let producer: KafkaProducer = KafkaProducer::new(cfg.kafka.brokers, cfg.kafka.schema_registry, args.topic);
// 2. Lecture CSV
let mut rdr = ReaderBuilder::new()
.has_headers(true)
.from_path(&args.csv)?;
for result in rdr.deserialize() {
let chat: ChatMessage = result.unwrap();
// 3. Envoi Kafka
if producer.produce(chat.user.to_string(), chat.clone()).await {
println!("✅ Sent chat {:?}", chat);
}
}
Ok(())
}

85
src/registry.rs Normal file
View File

@@ -0,0 +1,85 @@
use apache_avro::{AvroSchema, Schema};
use rdkafka::producer::{FutureProducer, FutureRecord};
use rdkafka::ClientConfig;
use schema_registry_converter::async_impl::avro::AvroEncoder;
use schema_registry_converter::async_impl::schema_registry::SrSettings;
use schema_registry_converter::avro_common::get_supplied_schema;
use schema_registry_converter::schema_registry_common::SubjectNameStrategy;
use serde::{Deserialize, Serialize};
use std::sync::Arc;
use std::time::Duration;
#[derive(Clone, Serialize, Deserialize, Debug)]
pub struct ChatMessage {
pub user: String,
pub message: String,
}
impl AvroSchema for ChatMessage {
fn get_schema() -> Schema {
// Define the Avro schema for ChatMessage
Schema::parse_str(
r#"{
"type": "record",
"name": "ChatMessage",
"fields": [
{"name": "user", "type": "string"},
{"name": "message", "type": "string"}
]
}"#
).unwrap()
}
}
pub struct KafkaProducer<'a> {
producer: FutureProducer,
avro_encoder: Arc<AvroEncoder<'a>>,
topic: String,
}
impl KafkaProducer<'_> {
pub fn new(bootstrap_servers: String, schema_registry_url: String, topic: String) -> Self {
let producer: FutureProducer = ClientConfig::new()
.set("bootstrap.servers", bootstrap_servers)
.set("produce.offset.report", "true")
.set("message.timeout.ms", "5000")
.set("queue.buffering.max.messages", "10")
.create()
.expect("Producer creation error");
let sr_settings = SrSettings::new(schema_registry_url);
let avro_encoder = AvroEncoder::new(sr_settings);
Self {
producer,
topic,
avro_encoder: Arc::new(avro_encoder),
}
}
pub async fn produce<T: Serialize + apache_avro::AvroSchema>(&self, key: String, payload: T) -> bool {
let value_strategy = SubjectNameStrategy::TopicNameStrategyWithSchema(
self.topic.clone(),
true,
get_supplied_schema(&T::get_schema()),
);
let payload_bytes = match self
.avro_encoder
.encode_struct(payload, &value_strategy)
.await
{
Ok(bytes) => bytes,
Err(e) => panic!("Error encoding payload: {}", e),
};
let record = FutureRecord::to(&self.topic)
.key(&key)
.payload(&payload_bytes);
match self.producer.send(record, Duration::from_secs(10)).await {
Ok(_) => true,
Err(_) => false,
}
}
}