@@ -79,7 +79,7 @@ inline fun <reified T> List<T>.toDS(spark: SparkSession): Dataset<T> =
79
79
spark.createDataset(this , encoder<T >())
80
80
81
81
/* *
82
- * Main method of API, which gives you seamless integraion with Spark:
82
+ * Main method of API, which gives you seamless integration with Spark:
83
83
* It creates encoder for any given supported type T
84
84
*
85
85
* Supported types are data classes, primitives, and Lists, Maps and Arrays containing them
@@ -220,7 +220,7 @@ inline fun <reified L : Any?, reified R : Any?> Dataset<L>.fullJoin(right: Datas
220
220
}
221
221
222
222
/* *
223
- * Alias for [Dataset.sort] which forces user to provide sortedcolumns from source dataset
223
+ * Alias for [Dataset.sort] which forces user to provide sorted columns from the source dataset
224
224
*
225
225
* @receiver source [Dataset]
226
226
* @param columns producer of sort columns
@@ -232,7 +232,7 @@ inline fun <reified T> Dataset<T>.sort(columns: (Dataset<T>) -> Array<Column>) =
232
232
* This function creates block, where one can call any further computations on already cached dataset
233
233
* Data will be unpersisted automatically at the end of computation
234
234
*
235
- * it may be useful in many situatiions , for example when one needs to write data to several targets
235
+ * it may be useful in many situations , for example, when one needs to write data to several targets
236
236
* ```kotlin
237
237
* ds.withCached {
238
238
* write()
@@ -241,7 +241,7 @@ inline fun <reified T> Dataset<T>.sort(columns: (Dataset<T>) -> Array<Column>) =
241
241
* }
242
242
* ```
243
243
*
244
- * @param blockingUnpersist if execution should be blocked until everyting persisted will be deleted
244
+ * @param blockingUnpersist if execution should be blocked until everything persisted will be deleted
245
245
* @param executeOnCached Block which should be executed on cached dataset.
246
246
* @return result of block execution for further usage. It may be anything including source or new dataset
247
247
*/
@@ -254,8 +254,8 @@ inline fun <reified T> Dataset<Row>.toList() = KSparkExtensions.collectAsList(to
254
254
inline fun <reified R > Dataset <* >.toArray (): Array <R > = to<R >().collect() as Array <R >
255
255
256
256
/* *
257
- * Alternative to [Dataset.show] which returns surce dataset.
258
- * Useful in debug purposes when you need to view contant of dataset as intermediate operation
257
+ * Alternative to [Dataset.show] which returns source dataset.
258
+ * Useful for debug purposes when you need to view content of a dataset as an intermediate operation
259
259
*/
260
260
fun <T > Dataset<T>.showDS (numRows : Int = 20, truncate : Boolean = true) = apply { show(numRows, truncate) }
261
261
0 commit comments