implicit class SparkeyPairSCollection[K, V] extends Serializable
- Alphabetic
- By Inheritance
- SparkeyPairSCollection
- Serializable
- AnyRef
- Any
- Hide All
- Show All
- Public
- Protected
Instance Constructors
- new SparkeyPairSCollection(self: SCollection[(K, V)])
Value Members
- final def !=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- final def ##: Int
- Definition Classes
- AnyRef → Any
- final def ==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- def asCachedStringSparkeySideInput(cache: Cache[String, String], numShards: Short = SparkeyIO.DefaultSideInputNumShards, compressionType: CompressionType = SparkeyIO.DefaultCompressionType, compressionBlockSize: Int = SparkeyIO.DefaultCompressionBlockSize)(implicit w: SparkeyWritable[K, V]): SideInput[CachedStringSparkeyReader]
Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
CachedStringSparkeyReader
, to be used with SCollection.withSideInputs.Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
CachedStringSparkeyReader
, to be used with SCollection.withSideInputs.- Annotations
- @experimental()
- final def asInstanceOf[T0]: T0
- Definition Classes
- Any
- def asLargeMapSideInput(numShards: Short = SparkeyIO.DefaultSideInputNumShards, compressionType: CompressionType = SparkeyIO.DefaultCompressionType, compressionBlockSize: Int = SparkeyIO.DefaultCompressionBlockSize): SideInput[SparkeyMap[K, V]]
Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyMap
, to be used with SCollection.withSideInputs.Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyMap
, to be used with SCollection.withSideInputs. It is required that each key of the input be associated with a single value. The resulting SideInput must fit on disk on each worker that reads it. This is strongly recommended over a regular MapSideInput if the data in the side input exceeds 100MB.- Annotations
- @experimental()
- def asLargeMapSideInput: SideInput[SparkeyMap[K, V]]
Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyMap
, to be used with SCollection.withSideInputs.Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyMap
, to be used with SCollection.withSideInputs. It is required that each key of the input be associated with a single value. The resulting SideInput must fit on disk on each worker that reads it. This is strongly recommended over a regular MapSideInput if the data in the side input exceeds 100MB.- Annotations
- @experimental()
- def asLargeMultiMapSideInput(numShards: Short = SparkeyIO.DefaultSideInputNumShards, compressionType: CompressionType = SparkeyIO.DefaultCompressionType, compressionBlockSize: Int = SparkeyIO.DefaultCompressionBlockSize): SideInput[SparkeyMap[K, Iterable[V]]]
Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
Map[key, Iterable[value]]
, to be used with SCollection.withSideInputs.Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
Map[key, Iterable[value]]
, to be used with SCollection.withSideInputs. In contrast to asLargeMapSideInput, it is not required that the keys in the input collection be unique. The resulting map is required to fit on disk on each worker. This is strongly recommended over a regular MultiMapSideInput if the data in the side input exceeds 100MB. - def asLargeMultiMapSideInput: SideInput[SparkeyMap[K, Iterable[V]]]
Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
Map[key, Iterable[value]]
, to be used with SCollection.withSideInputs.Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
Map[key, Iterable[value]]
, to be used with SCollection.withSideInputs. In contrast to asLargeMapSideInput, it is not required that the keys in the input collection be unique. The resulting map is required to fit on disk on each worker. This is strongly recommended over a regular MultiMapSideInput if the data in the side input exceeds 100MB.- Annotations
- @experimental()
- def asSparkeySideInput(implicit w: SparkeyWritable[K, V]): SideInput[SparkeyReader]
Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyReader
, to be used with SCollection.withSideInputs.Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyReader
, to be used with SCollection.withSideInputs. It is required that each key of the input be associated with a single value.- Annotations
- @experimental()
- def asSparkeySideInput(numShards: Short = SparkeyIO.DefaultSideInputNumShards, compressionType: CompressionType = SparkeyIO.DefaultCompressionType, compressionBlockSize: Int = SparkeyIO.DefaultCompressionBlockSize)(implicit w: SparkeyWritable[K, V]): SideInput[SparkeyReader]
Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyReader
, to be used with SCollection.withSideInputs.Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyReader
, to be used with SCollection.withSideInputs. It is required that each key of the input be associated with a single value.- numShards
the number of shards to use when writing the Sparkey file(s).
- Annotations
- @experimental()
- def asTypedSparkeySideInput[T](cache: Cache[String, T], numShards: Short = SparkeyIO.DefaultSideInputNumShards, compressionType: CompressionType = SparkeyIO.DefaultCompressionType, compressionBlockSize: Int = SparkeyIO.DefaultCompressionBlockSize)(decoder: (Array[Byte]) => T)(implicit w: SparkeyWritable[K, V]): SideInput[TypedSparkeyReader[T]]
Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyReader
, to be used with SCollection.withSideInputs.Convert this SCollection to a SideInput, mapping key-value pairs of each window to a
SparkeyReader
, to be used with SCollection.withSideInputs. It is required that each key of the input be associated with a single value. The provided Cache will be used to cache reads from the resulting SparkeyReader.- Annotations
- @experimental()
- def clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.CloneNotSupportedException]) @native()
- final def eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- def equals(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef → Any
- def finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.Throwable])
- final def getClass(): Class[_ <: AnyRef]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
- def hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
- final def isInstanceOf[T0]: Boolean
- Definition Classes
- Any
- final def ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- final def notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
- final def notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
- def saveAsSparkey(path: String, maxMemoryUsage: Long = -1, numShards: Short = SparkeyIO.DefaultNumShards, compressionType: CompressionType = SparkeyIO.DefaultCompressionType, compressionBlockSize: Int = SparkeyIO.DefaultCompressionBlockSize)(implicit writable: SparkeyWritable[K, V]): ClosedTap[Nothing]
Write the key-value pairs of this SCollection as a Sparkey file to a specific location.
Write the key-value pairs of this SCollection as a Sparkey file to a specific location.
- path
where to write the sparkey files
- maxMemoryUsage
(optional) how much memory (in bytes) is allowed for writing the index file
- numShards
(optional) the number of shards to split this dataset into before writing. One pair of Sparkey files will be written for each shard, sharded by MurmurHash3 of the key mod the number of shards.
- final def synchronized[T0](arg0: => T0): T0
- Definition Classes
- AnyRef
- def toString(): String
- Definition Classes
- AnyRef → Any
- final def wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- final def wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- final def wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException]) @native()
Deprecated Value Members
- def asSparkey(implicit w: SparkeyWritable[K, V]): SCollection[SparkeyUri]
Write the key-value pairs of this SCollection as a Sparkey file to a temporary location.
Write the key-value pairs of this SCollection as a Sparkey file to a temporary location.
- returns
A singleton SCollection containing the SparkeyUri of the saved files.
- Annotations
- @deprecated
- Deprecated
(Since version 0.14.0) Use saveAsSparkey instead
- def asSparkey(path: String = null, maxMemoryUsage: Long = -1, numShards: Short = SparkeyIO.DefaultNumShards, compressionType: CompressionType = SparkeyIO.DefaultCompressionType, compressionBlockSize: Int = SparkeyIO.DefaultCompressionBlockSize)(implicit w: SparkeyWritable[K, V]): SCollection[SparkeyUri]
Write the key-value pairs of this SCollection as a Sparkey file to a specific location.
Write the key-value pairs of this SCollection as a Sparkey file to a specific location.
- path
where to write the sparkey files. Defaults to a temporary location.
- maxMemoryUsage
(optional) how much memory (in bytes) is allowed for writing the index file
- numShards
(optional) the number of shards to split this dataset into before writing. One pair of Sparkey files will be written for each shard, sharded by MurmurHash3 of the key mod the number of shards.
- returns
A singleton SCollection containing the SparkeyUri of the saved files.
- Annotations
- @deprecated
- Deprecated
(Since version 0.14.0) Use saveAsSparkey instead