With a rich technical background, I am currently working on leveraging Big Data, Apache Spark, and Databricks to deliver value from data using the combination of Big Data, Machine Learning and Agile principles.
My professional journey reflects a commitment to continuous improvement and strategic transformation within the tech landscape. Among others, I focus on Domain Driven Design, Clean Code, MLOps, DevOps and Developer Experience (DUX), to develop high-quality, resilient systems.
French session - Intermediate
On dit que les données sont l’or noir du 21e siècle. Cependant, la qualité des données dans les organisations est toujours pointée du doigt comme étant insuffisante pour en tirer une valeur. Dans cette conférence, nous allons passer en revue les techniques peuvent aider à améliorer la qualité et la valeur des données, comme les quality checks, le data observability, les data contract et le data mesh. Le tout accompagné d'exemples pratiques.
English session - Intermediate
Real-time data processing and analytics is more and more a requirement for modern applications. But how do you choose between Apache Spark and Apache Flink for streaming workloads? This talk dives into their architectures, trade-offs, and real-world use cases. We compare Spark Structured Streaming (micro-batching, ease of use) vs. Flink’s native streaming (low latency, stateful processing), looking at their features, performance and ecosystem.