powered by
Compute the number of records within each partition of a Spark DataFrame
sdf_partition_sizes(x)
A spark_connection, ml_pipeline, or a tbl_spark.
spark_connection
ml_pipeline
tbl_spark
if (FALSE) { library(sparklyr) sc <- spark_connect(master = "spark://HOST:PORT") example_sdf <- sdf_len(sc, 100L, repartition = 10L) example_sdf %>% sdf_partition_sizes() %>% print() }
Run the code above in your browser using DataLab