Packages

c

org.apache.spark.sql.execution.streaming

DistributedHTTPSource

class DistributedHTTPSource extends Source with Logging with Serializable

Linear Supertypes
Serializable, Serializable, Logging, Source, SparkDataStream, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. DistributedHTTPSource
  2. Serializable
  3. Serializable
  4. Logging
  5. Source
  6. SparkDataStream
  7. AnyRef
  8. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new DistributedHTTPSource(name: String, host: String, port: Int, maxPortAttempts: Int, maxPartitions: Option[Int], handleResponseErrors: Boolean, sqlContext: SQLContext)

Value Members

  1. def commit(end: connector.read.streaming.Offset): Unit
    Definition Classes
    DistributedHTTPSource → Source → SparkDataStream
  2. def commit(end: Offset): Unit
    Definition Classes
    Source
  3. def deserializeOffset(json: String): connector.read.streaming.Offset
    Definition Classes
    Source → SparkDataStream
  4. def getBatch(start: Option[Offset], end: Offset): DataFrame

    Returns the data that is between the offsets (start, end].

    Returns the data that is between the offsets (start, end].

    Definition Classes
    DistributedHTTPSource → Source
  5. def getOffset: Option[Offset]
    Definition Classes
    DistributedHTTPSource → Source
  6. def initialOffset(): connector.read.streaming.Offset
    Definition Classes
    Source → SparkDataStream
  7. def schema: StructType

    Returns the schema of the data from this source

    Returns the schema of the data from this source

    Definition Classes
    DistributedHTTPSource → Source
  8. def stop(): Unit

    Stop this source.

    Stop this source.

    Definition Classes
    DistributedHTTPSource → SparkDataStream
  9. def toString(): String
    Definition Classes
    DistributedHTTPSource → AnyRef → Any