Click any tag below to further narrow down your results
Links
This article explains Spark Declarative Pipelines (SDP), a framework for creating data pipelines in Spark. It covers key concepts like flows, datasets, and pipelines, along with how to implement them in Python and SQL. The guide also includes installation instructions and usage of the command line interface.
To efficiently insert large datasets into a Postgres database, combining Spark's parallel processing with Python's COPY command can significantly enhance performance. By repartitioning the data and utilizing multiple writers, the author was able to insert 22 million records in under 14 minutes, leveraging Postgres's bulk-loading capabilities over traditional JDBC methods.