Codec value indicating that a text file should be written using the BZip2 codec
An option key with which to specify the compression codec with which to write text data
Codec value indicating that a text file should be written using the GZip codec
An option key with which to specify the minimum number of partitions into which to read an input file
Simple text format, one line per record.
Stub object necessary due to https://issues.scala-lang.org/browse/SI-8124
Stub object necessary due to https://issues.scala-lang.org/browse/SI-8124
Documentation for ops.core.rdd.io
can be found at software.uncharted.sparkpipe.ops.core.rdd.io
Translates a SparkSession into a SparkContext, so that RDD operations can be called with either
Translates a SparkSession into a SparkContext, so that RDD operations can be called with either
A SparkSession in which to run operations
The spark context from which the SQL context was created
Traslate a function from a SparkContext into a function from a SparkSession, so that RDD operations can be run off a Pipe[SparkSession]
Traslate a function from a SparkContext into a function from a SparkSession, so that RDD operations can be run off a Pipe[SparkSession]
The return type of the function
The SparkContext-based function
The same function, but working on a SparkSession.
Reads a file into an RDD
Reads a file into an RDD
The location of the source data
The format in which to read the data. Currently, only "text" is supported.
A Map[String, String] of options. Currently, the only supported option is "minPartitions", which will set the minimum number of partitions into which the data is read.
The spark context in which to read the data
An RDD of the text of the source data, line by line
Write an RDD
Write an RDD
The type of data contained in the RDD
The location to which to write the data
The format in which to write the data. Currently, only "text" is supported.
A Map[String, String] of options. Currently, only the "codec" option is supported, for which valid values are "bzip2", and "gzip"; any other value will result in the default codec.
The RDD to write
The input RDD
Input/output operations for RDDs, based on the
SparkContext.textFile
API