Find everything you need to get certified on Fabric—skills challenges, live sessions, exam prep, role guidance, and a 50 percent discount on exams.
Get startedEarn a 50% discount on the DP-600 certification exam by completing the Fabric 30 Days to Learn It challenge.
Greetings, community. I have a bunch of files I'm planning to load into a Lakehouse in CSV format. From there, I have considered loading them as Delta tables where possible. Does this duplicate the data, however, as the data is persisted in "raw" file format and then generated in Parquet format for the Delta table?
Solved! Go to Solution.
Hi @arpost yes this will duplicate the data but you are transforming into a far better and more efficient format when saving as Delta, plus the underlying parquet will be compressed and likely smaller size than the source CSVs
Hi @arpost yes this will duplicate the data but you are transforming into a far better and more efficient format when saving as Delta, plus the underlying parquet will be compressed and likely smaller size than the source CSVs
Ask questions in Eventhouse and KQL, Eventstream, and Reflex.
Ask questions in Data Engineering, Data Science, Data Warehouse and General Discussion.