- Using PySpark, write code that will transform and load the data from the data lake - By using Kafka as an input source for Spark Structured Streaming and Delta Lake as a storage layer, build a complete streaming data pipeline to consolidate our data - you should read [From Kafka to Delta Lake using Apache Spark Structured Streaming (michelin.io)](https://blogit.michelin.io/kafka-to-delta-lake-using-apache-spark-streaming-avro/)