org.apache.spark.util.collection.OpenHashSet Scala Examples
The following examples show how to use org.apache.spark.util.collection.OpenHashSet.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
Example 1
Source File: GraphXUtils.scala From iolap with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import org.apache.spark.SparkConf import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.{OpenHashSet, BitSet} import org.apache.spark.util.BoundedPriorityQueue object GraphXUtils { def registerKryoClasses(conf: SparkConf) { conf.registerKryoClasses(Array( classOf[Edge[Object]], classOf[(VertexId, Object)], classOf[EdgePartition[Object, Object]], classOf[BitSet], classOf[VertexIdToIndexMap], classOf[VertexAttributeBlock[Object]], classOf[PartitionStrategy], classOf[BoundedPriorityQueue[Object]], classOf[EdgeDirection], classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]], classOf[OpenHashSet[Int]], classOf[OpenHashSet[Long]])) } }
Example 2
Source File: GraphXUtils.scala From zen with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx2 import org.apache.spark.SparkConf import org.apache.spark.graphx2.impl._ import org.apache.spark.graphx2.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.{OpenHashSet, BitSet} import org.apache.spark.util.BoundedPriorityQueue object GraphXUtils { def registerKryoClasses(conf: SparkConf) { conf.registerKryoClasses(Array( classOf[Edge[Object]], classOf[(VertexId, Object)], classOf[EdgePartition[Object, Object]], classOf[BitSet], classOf[VertexIdToIndexMap], classOf[VertexAttributeBlock[Object]], classOf[PartitionStrategy], classOf[BoundedPriorityQueue[Object]], classOf[EdgeDirection], classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]], classOf[OpenHashSet[Int]], classOf[OpenHashSet[Long]])) } }
Example 3
Source File: UIData.scala From BigDatalog with Apache License 2.0 | 5 votes |
package org.apache.spark.ui.jobs import org.apache.spark.JobExecutionStatus import org.apache.spark.executor.TaskMetrics import org.apache.spark.scheduler.{AccumulableInfo, TaskInfo} import org.apache.spark.util.collection.OpenHashSet import scala.collection.mutable import scala.collection.mutable.HashMap private[spark] object UIData { class ExecutorSummary { var taskTime : Long = 0 var failedTasks : Int = 0 var succeededTasks : Int = 0 var inputBytes : Long = 0 var inputRecords : Long = 0 var outputBytes : Long = 0 var outputRecords : Long = 0 var shuffleRead : Long = 0 var shuffleReadRecords : Long = 0 var shuffleWrite : Long = 0 var shuffleWriteRecords : Long = 0 var memoryBytesSpilled : Long = 0 var diskBytesSpilled : Long = 0 } class JobUIData( var jobId: Int = -1, var submissionTime: Option[Long] = None, var completionTime: Option[Long] = None, var stageIds: Seq[Int] = Seq.empty, var jobGroup: Option[String] = None, var status: JobExecutionStatus = JobExecutionStatus.UNKNOWN, case class TaskUIData( var taskInfo: TaskInfo, var taskMetrics: Option[TaskMetrics] = None, var errorMessage: Option[String] = None) case class ExecutorUIData( val startTime: Long, var finishTime: Option[Long] = None, var finishReason: Option[String] = None) }
Example 4
Source File: GraphXUtils.scala From BigDatalog with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import org.apache.spark.SparkConf import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.{OpenHashSet, BitSet} import org.apache.spark.util.BoundedPriorityQueue object GraphXUtils { def registerKryoClasses(conf: SparkConf) { conf.registerKryoClasses(Array( classOf[Edge[Object]], classOf[(VertexId, Object)], classOf[EdgePartition[Object, Object]], classOf[BitSet], classOf[VertexIdToIndexMap], classOf[VertexAttributeBlock[Object]], classOf[PartitionStrategy], classOf[BoundedPriorityQueue[Object]], classOf[EdgeDirection], classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]], classOf[OpenHashSet[Int]], classOf[OpenHashSet[Long]])) } }
Example 5
Source File: GraphKryoRegistrator.scala From BigDatalog with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import com.esotericsoftware.kryo.Kryo import org.apache.spark.serializer.KryoRegistrator import org.apache.spark.util.BoundedPriorityQueue import org.apache.spark.util.collection.BitSet import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.OpenHashSet @deprecated("Register GraphX classes with Kryo using GraphXUtils.registerKryoClasses", "1.2.0") class GraphKryoRegistrator extends KryoRegistrator { def registerClasses(kryo: Kryo) { kryo.register(classOf[Edge[Object]]) kryo.register(classOf[(VertexId, Object)]) kryo.register(classOf[EdgePartition[Object, Object]]) kryo.register(classOf[BitSet]) kryo.register(classOf[VertexIdToIndexMap]) kryo.register(classOf[VertexAttributeBlock[Object]]) kryo.register(classOf[PartitionStrategy]) kryo.register(classOf[BoundedPriorityQueue[Object]]) kryo.register(classOf[EdgeDirection]) kryo.register(classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]]) kryo.register(classOf[OpenHashSet[Int]]) kryo.register(classOf[OpenHashSet[Long]]) } }
Example 6
Source File: GraphXUtils.scala From Spark-2.3.1 with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import scala.reflect.ClassTag import org.apache.spark.SparkConf import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.BoundedPriorityQueue import org.apache.spark.util.collection.{BitSet, OpenHashSet} object GraphXUtils { private[graphx] def mapReduceTriplets[VD: ClassTag, ED: ClassTag, A: ClassTag]( g: Graph[VD, ED], mapFunc: EdgeTriplet[VD, ED] => Iterator[(VertexId, A)], reduceFunc: (A, A) => A, activeSetOpt: Option[(VertexRDD[_], EdgeDirection)] = None): VertexRDD[A] = { def sendMsg(ctx: EdgeContext[VD, ED, A]) { mapFunc(ctx.toEdgeTriplet).foreach { kv => val id = kv._1 val msg = kv._2 if (id == ctx.srcId) { ctx.sendToSrc(msg) } else { assert(id == ctx.dstId) ctx.sendToDst(msg) } } } g.aggregateMessagesWithActiveSet( sendMsg, reduceFunc, TripletFields.All, activeSetOpt) } }
Example 7
Source File: UIData.scala From spark1.52 with Apache License 2.0 | 5 votes |
package org.apache.spark.ui.jobs import org.apache.spark.JobExecutionStatus import org.apache.spark.executor.TaskMetrics import org.apache.spark.scheduler.{AccumulableInfo, TaskInfo} import org.apache.spark.util.collection.OpenHashSet import scala.collection.mutable import scala.collection.mutable.HashMap private[spark] object UIData { class ExecutorSummary { var taskTime : Long = 0//任务时间 var failedTasks : Int = 0//失败任务数 var succeededTasks : Int = 0//完成任务数 var inputBytes : Long = 0 var inputRecords : Long = 0 var outputBytes : Long = 0 var outputRecords : Long = 0 var shuffleRead : Long = 0 var shuffleReadRecords : Long = 0 var shuffleWrite : Long = 0 var shuffleWriteRecords : Long = 0 var memoryBytesSpilled : Long = 0 var diskBytesSpilled : Long = 0 } class JobUIData( var jobId: Int = -1, var submissionTime: Option[Long] = None,//提交时间 var completionTime: Option[Long] = None,//完成时间 var stageIds: Seq[Int] = Seq.empty, var jobGroup: Option[String] = None, var status: JobExecutionStatus = JobExecutionStatus.UNKNOWN, case class TaskUIData( var taskInfo: TaskInfo, var taskMetrics: Option[TaskMetrics] = None, var errorMessage: Option[String] = None) case class ExecutorUIData( val startTime: Long, var finishTime: Option[Long] = None, var finishReason: Option[String] = None) }
Example 8
Source File: GraphXUtils.scala From spark1.52 with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import org.apache.spark.SparkConf import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.{OpenHashSet, BitSet} import org.apache.spark.util.BoundedPriorityQueue object GraphXUtils { def registerKryoClasses(conf: SparkConf) { conf.registerKryoClasses(Array( classOf[Edge[Object]], classOf[(VertexId, Object)], classOf[EdgePartition[Object, Object]], classOf[BitSet], classOf[VertexIdToIndexMap], classOf[VertexAttributeBlock[Object]], classOf[PartitionStrategy], classOf[BoundedPriorityQueue[Object]], classOf[EdgeDirection], classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]], classOf[OpenHashSet[Int]], classOf[OpenHashSet[Long]])) } }
Example 9
Source File: GraphKryoRegistrator.scala From spark1.52 with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import com.esotericsoftware.kryo.Kryo import org.apache.spark.serializer.KryoRegistrator import org.apache.spark.util.BoundedPriorityQueue import org.apache.spark.util.collection.BitSet import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.OpenHashSet @deprecated("Register GraphX classes with Kryo using GraphXUtils.registerKryoClasses", "1.2.0") class GraphKryoRegistrator extends KryoRegistrator { def registerClasses(kryo: Kryo) { kryo.register(classOf[Edge[Object]]) kryo.register(classOf[(VertexId, Object)]) kryo.register(classOf[EdgePartition[Object, Object]]) kryo.register(classOf[BitSet]) kryo.register(classOf[VertexIdToIndexMap]) kryo.register(classOf[VertexAttributeBlock[Object]]) kryo.register(classOf[PartitionStrategy]) kryo.register(classOf[BoundedPriorityQueue[Object]]) kryo.register(classOf[EdgeDirection]) kryo.register(classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]]) kryo.register(classOf[OpenHashSet[Int]]) kryo.register(classOf[OpenHashSet[Long]]) } }
Example 10
Source File: UIData.scala From iolap with Apache License 2.0 | 5 votes |
package org.apache.spark.ui.jobs import org.apache.spark.JobExecutionStatus import org.apache.spark.executor.TaskMetrics import org.apache.spark.scheduler.{AccumulableInfo, TaskInfo} import org.apache.spark.util.collection.OpenHashSet import scala.collection.mutable.HashMap private[spark] object UIData { class ExecutorSummary { var taskTime : Long = 0 var failedTasks : Int = 0 var succeededTasks : Int = 0 var inputBytes : Long = 0 var inputRecords : Long = 0 var outputBytes : Long = 0 var outputRecords : Long = 0 var shuffleRead : Long = 0 var shuffleReadRecords : Long = 0 var shuffleWrite : Long = 0 var shuffleWriteRecords : Long = 0 var memoryBytesSpilled : Long = 0 var diskBytesSpilled : Long = 0 } class JobUIData( var jobId: Int = -1, var submissionTime: Option[Long] = None, var completionTime: Option[Long] = None, var stageIds: Seq[Int] = Seq.empty, var jobGroup: Option[String] = None, var status: JobExecutionStatus = JobExecutionStatus.UNKNOWN, case class TaskUIData( var taskInfo: TaskInfo, var taskMetrics: Option[TaskMetrics] = None, var errorMessage: Option[String] = None) case class ExecutorUIData( val startTime: Long, var finishTime: Option[Long] = None, var finishReason: Option[String] = None) }
Example 11
Source File: GraphXPrimitiveKeyOpenHashMap.scala From drizzle-spark with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx.util.collection import scala.reflect._ import org.apache.spark.util.collection.OpenHashSet def computeNextPair(): (K, V) = { pos = keySet.nextPos(pos) if (pos >= 0) { val ret = (keySet.getValue(pos), _values(pos)) pos += 1 ret } else { null } } def hasNext: Boolean = nextPair != null def next(): (K, V) = { val pair = nextPair nextPair = computeNextPair() pair } } // The following member variables are declared as protected instead of private for the // specialization to work (specialized class extends the unspecialized one and needs access // to the "private" variables). // They also should have been val's. We use var's because there is a Scala compiler bug that // would throw illegal access error at runtime if they are declared as val's. protected var grow = (newCapacity: Int) => { _oldValues = _values _values = new Array[V](newCapacity) } protected var move = (oldPos: Int, newPos: Int) => { _values(newPos) = _oldValues(oldPos) } }
Example 12
Source File: GraphKryoRegistrator.scala From iolap with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import com.esotericsoftware.kryo.Kryo import org.apache.spark.serializer.KryoRegistrator import org.apache.spark.util.BoundedPriorityQueue import org.apache.spark.util.collection.BitSet import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.OpenHashSet @deprecated("Register GraphX classes with Kryo using GraphXUtils.registerKryoClasses", "1.2.0") class GraphKryoRegistrator extends KryoRegistrator { def registerClasses(kryo: Kryo) { kryo.register(classOf[Edge[Object]]) kryo.register(classOf[(VertexId, Object)]) kryo.register(classOf[EdgePartition[Object, Object]]) kryo.register(classOf[BitSet]) kryo.register(classOf[VertexIdToIndexMap]) kryo.register(classOf[VertexAttributeBlock[Object]]) kryo.register(classOf[PartitionStrategy]) kryo.register(classOf[BoundedPriorityQueue[Object]]) kryo.register(classOf[EdgeDirection]) kryo.register(classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]]) kryo.register(classOf[OpenHashSet[Int]]) kryo.register(classOf[OpenHashSet[Long]]) } }
Example 13
Source File: GraphXUtils.scala From multi-tenancy-spark with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import scala.reflect.ClassTag import org.apache.spark.SparkConf import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.BoundedPriorityQueue import org.apache.spark.util.collection.{BitSet, OpenHashSet} object GraphXUtils { private[graphx] def mapReduceTriplets[VD: ClassTag, ED: ClassTag, A: ClassTag]( g: Graph[VD, ED], mapFunc: EdgeTriplet[VD, ED] => Iterator[(VertexId, A)], reduceFunc: (A, A) => A, activeSetOpt: Option[(VertexRDD[_], EdgeDirection)] = None): VertexRDD[A] = { def sendMsg(ctx: EdgeContext[VD, ED, A]) { mapFunc(ctx.toEdgeTriplet).foreach { kv => val id = kv._1 val msg = kv._2 if (id == ctx.srcId) { ctx.sendToSrc(msg) } else { assert(id == ctx.dstId) ctx.sendToDst(msg) } } } g.aggregateMessagesWithActiveSet( sendMsg, reduceFunc, TripletFields.All, activeSetOpt) } }
Example 14
Source File: UIData.scala From SparkCore with Apache License 2.0 | 5 votes |
package org.apache.spark.ui.jobs import org.apache.spark.JobExecutionStatus import org.apache.spark.executor.TaskMetrics import org.apache.spark.scheduler.{AccumulableInfo, TaskInfo} import org.apache.spark.util.collection.OpenHashSet import scala.collection.mutable.HashMap private[jobs] object UIData { class ExecutorSummary { var taskTime : Long = 0 var failedTasks : Int = 0 var succeededTasks : Int = 0 var inputBytes : Long = 0 var inputRecords : Long = 0 var outputBytes : Long = 0 var outputRecords : Long = 0 var shuffleRead : Long = 0 var shuffleReadRecords : Long = 0 var shuffleWrite : Long = 0 var shuffleWriteRecords : Long = 0 var memoryBytesSpilled : Long = 0 var diskBytesSpilled : Long = 0 } class JobUIData( var jobId: Int = -1, var submissionTime: Option[Long] = None, var completionTime: Option[Long] = None, var stageIds: Seq[Int] = Seq.empty, var jobGroup: Option[String] = None, var status: JobExecutionStatus = JobExecutionStatus.UNKNOWN, case class TaskUIData( var taskInfo: TaskInfo, var taskMetrics: Option[TaskMetrics] = None, var errorMessage: Option[String] = None) }
Example 15
Source File: GraphXUtils.scala From sparkoscope with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import scala.reflect.ClassTag import org.apache.spark.SparkConf import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.BoundedPriorityQueue import org.apache.spark.util.collection.{BitSet, OpenHashSet} object GraphXUtils { private[graphx] def mapReduceTriplets[VD: ClassTag, ED: ClassTag, A: ClassTag]( g: Graph[VD, ED], mapFunc: EdgeTriplet[VD, ED] => Iterator[(VertexId, A)], reduceFunc: (A, A) => A, activeSetOpt: Option[(VertexRDD[_], EdgeDirection)] = None): VertexRDD[A] = { def sendMsg(ctx: EdgeContext[VD, ED, A]) { mapFunc(ctx.toEdgeTriplet).foreach { kv => val id = kv._1 val msg = kv._2 if (id == ctx.srcId) { ctx.sendToSrc(msg) } else { assert(id == ctx.dstId) ctx.sendToDst(msg) } } } g.aggregateMessagesWithActiveSet( sendMsg, reduceFunc, TripletFields.All, activeSetOpt) } }
Example 16
Source File: GraphXUtils.scala From graphx-algorithm with GNU General Public License v2.0 | 5 votes |
package org.apache.spark.graphx import org.apache.spark.SparkConf import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.{OpenHashSet, BitSet} import org.apache.spark.util.BoundedPriorityQueue object GraphXUtils { def registerKryoClasses(conf: SparkConf) { conf.registerKryoClasses(Array( classOf[Edge[Object]], classOf[(VertexId, Object)], classOf[EdgePartition[Object, Object]], classOf[BitSet], classOf[VertexIdToIndexMap], classOf[VertexAttributeBlock[Object]], classOf[PartitionStrategy], classOf[BoundedPriorityQueue[Object]], classOf[EdgeDirection], classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]], classOf[OpenHashSet[Int]], classOf[OpenHashSet[Long]])) } }
Example 17
Source File: GraphXPrimitiveKeyOpenHashMap.scala From graphx-algorithm with GNU General Public License v2.0 | 5 votes |
package org.apache.spark.graphx.util.collection import org.apache.spark.util.collection.OpenHashSet import scala.reflect._ def computeNextPair(): (K, V) = { pos = keySet.nextPos(pos) if (pos >= 0) { val ret = (keySet.getValue(pos), _values(pos)) pos += 1 ret } else { null } } def hasNext = nextPair != null def next() = { val pair = nextPair nextPair = computeNextPair() pair } } // The following member variables are declared as protected instead of private for the // specialization to work (specialized class extends the unspecialized one and needs access // to the "private" variables). // They also should have been val's. We use var's because there is a Scala compiler bug that // would throw illegal access error at runtime if they are declared as val's. protected var grow = (newCapacity: Int) => { _oldValues = _values _values = new Array[V](newCapacity) } protected var move = (oldPos: Int, newPos: Int) => { _values(newPos) = _oldValues(oldPos) } }
Example 18
Source File: GraphKryoRegistrator.scala From graphx-algorithm with GNU General Public License v2.0 | 5 votes |
package org.apache.spark.graphx import com.esotericsoftware.kryo.Kryo import org.apache.spark.serializer.KryoRegistrator import org.apache.spark.util.BoundedPriorityQueue import org.apache.spark.util.collection.BitSet import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.collection.OpenHashSet @deprecated("Register GraphX classes with Kryo using GraphXUtils.registerKryoClasses", "1.2.0") class GraphKryoRegistrator extends KryoRegistrator { def registerClasses(kryo: Kryo) { kryo.register(classOf[Edge[Object]]) kryo.register(classOf[(VertexId, Object)]) kryo.register(classOf[EdgePartition[Object, Object]]) kryo.register(classOf[BitSet]) kryo.register(classOf[VertexIdToIndexMap]) kryo.register(classOf[VertexAttributeBlock[Object]]) kryo.register(classOf[PartitionStrategy]) kryo.register(classOf[BoundedPriorityQueue[Object]]) kryo.register(classOf[EdgeDirection]) kryo.register(classOf[GraphXPrimitiveKeyOpenHashMap[VertexId, Int]]) kryo.register(classOf[OpenHashSet[Int]]) kryo.register(classOf[OpenHashSet[Long]]) } }
Example 19
Source File: GraphXUtils.scala From drizzle-spark with Apache License 2.0 | 5 votes |
package org.apache.spark.graphx import scala.reflect.ClassTag import org.apache.spark.SparkConf import org.apache.spark.graphx.impl._ import org.apache.spark.graphx.util.collection.GraphXPrimitiveKeyOpenHashMap import org.apache.spark.util.BoundedPriorityQueue import org.apache.spark.util.collection.{BitSet, OpenHashSet} object GraphXUtils { private[graphx] def mapReduceTriplets[VD: ClassTag, ED: ClassTag, A: ClassTag]( g: Graph[VD, ED], mapFunc: EdgeTriplet[VD, ED] => Iterator[(VertexId, A)], reduceFunc: (A, A) => A, activeSetOpt: Option[(VertexRDD[_], EdgeDirection)] = None): VertexRDD[A] = { def sendMsg(ctx: EdgeContext[VD, ED, A]) { mapFunc(ctx.toEdgeTriplet).foreach { kv => val id = kv._1 val msg = kv._2 if (id == ctx.srcId) { ctx.sendToSrc(msg) } else { assert(id == ctx.dstId) ctx.sendToDst(msg) } } } g.aggregateMessagesWithActiveSet( sendMsg, reduceFunc, TripletFields.All, activeSetOpt) } }