scala.collection.generic.Growable Scala Examples
The following examples show how to use scala.collection.generic.Growable.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
Example 1
Source File: BoundedPriorityQueue.scala From mmlspark with MIT License | 5 votes |
// Copyright (C) Microsoft Corporation. All rights reserved. // Licensed under the MIT License. See LICENSE in project root for information. package com.microsoft.ml.spark.nn class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { import scala.collection.JavaConverters._ private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear(): Unit = { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 2
Source File: BoundedUniquePriorityQueue.scala From spark3D with Apache License 2.0 | 5 votes |
package com.astrolabsoftware.spark3d.utils import java.io.Serializable import com.astrolabsoftware.spark3d.geometryObjects.Shape3D.Shape3D import scala.collection.generic.Growable import scala.collection.mutable import collection.mutable.PriorityQueue class BoundedUniquePriorityQueue[A <: Shape3D](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { // underlying base priority queue private val underlying = new PriorityQueue[A]()(ord) // HashSet of elements contained in the priority queue used to ensure uniqueness of the elements // in the priority queue. private val containedElements = new mutable.HashSet[Int]() override def iterator: Iterator[A] = underlying.iterator override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { val elementHash = elem.center.getCoordinate.hashCode // check if element to be inserted is unique or not if (!containedElements.contains(elementHash)) { if (size < maxSize) { underlying.enqueue(elem) containedElements.add(elementHash) } else { maybeReplaceLowest(elem) } } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.head // Definition of scala.Ordering.get(x, y) is: // Returns true iff y comes before x in the ordering and is not the same as x. if (head != null && ord.gt(head, a)) { underlying.dequeue underlying.enqueue(a) containedElements.add(a.center.getCoordinate.hashCode) containedElements.remove(head.center.getCoordinate.hashCode) true } else { false } } }
Example 3
Source File: ItemFiles.scala From flamy with Apache License 2.0 | 5 votes |
package com.flaminem.flamy.model.files import com.flaminem.flamy.model.names.ItemName import com.flaminem.flamy.utils.logging.Logging import scala.collection.generic.Growable import scala.collection.mutable.ListBuffer import scala.collection.{TraversableLike, mutable} def filterFileTypes(predicate: (ItemName, FileType)=>Boolean): ItemFiles[T] = { this.filter{ case file => predicate(file.getItemName, file.fileType) } } override def toString: String = mm.values.toString override def +=(file: T): ItemFiles.this.type = { mm.addBinding(file.fileType, file) this } override def clear(): Unit = { mm.clear() } }
Example 4
Source File: BoundedPriorityQueue.scala From albedo with MIT License | 5 votes |
// COPY DIRECTLY FROM SPARK SOURCE CODE package ws.vinta.albedo.recommenders import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 5
Source File: BoundedPriorityQueue.scala From BigDatalog with Apache License 2.0 | 5 votes |
package org.apache.spark.util import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable private[spark] class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 6
Source File: BoundedPriorityQueue.scala From Spark-2.3.1 with Apache License 2.0 | 5 votes |
package org.apache.spark.util import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable private[spark] class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } def poll(): A = { underlying.poll() } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 7
Source File: BoundedPriorityQueue.scala From spark1.52 with Apache License 2.0 | 5 votes |
package org.apache.spark.util import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable private[spark] class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 8
Source File: BoundedPriorityQueue.scala From iolap with Apache License 2.0 | 5 votes |
package org.apache.spark.util import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable private[spark] class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 9
Source File: BoundedPriorityQueue.scala From multi-tenancy-spark with Apache License 2.0 | 5 votes |
package org.apache.spark.util import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable private[spark] class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 10
Source File: collect.scala From drizzle-spark with Apache License 2.0 | 5 votes |
package org.apache.spark.sql.catalyst.expressions.aggregate import scala.collection.generic.Growable import scala.collection.mutable import org.apache.spark.sql.catalyst.analysis.TypeCheckResult import org.apache.spark.sql.catalyst.expressions._ import org.apache.spark.sql.catalyst.util.GenericArrayData import org.apache.spark.sql.catalyst.InternalRow import org.apache.spark.sql.types._ @ExpressionDescription( usage = "_FUNC_(expr) - Collects and returns a set of unique elements.") case class CollectSet( child: Expression, mutableAggBufferOffset: Int = 0, inputAggBufferOffset: Int = 0) extends Collect { def this(child: Expression) = this(child, 0, 0) override def checkInputDataTypes(): TypeCheckResult = { if (!child.dataType.existsRecursively(_.isInstanceOf[MapType])) { TypeCheckResult.TypeCheckSuccess } else { TypeCheckResult.TypeCheckFailure("collect_set() cannot have map type data") } } override def withNewMutableAggBufferOffset(newMutableAggBufferOffset: Int): ImperativeAggregate = copy(mutableAggBufferOffset = newMutableAggBufferOffset) override def withNewInputAggBufferOffset(newInputAggBufferOffset: Int): ImperativeAggregate = copy(inputAggBufferOffset = newInputAggBufferOffset) override def prettyName: String = "collect_set" override protected[this] val buffer: mutable.HashSet[Any] = mutable.HashSet.empty }
Example 11
Source File: BoundedPriorityQueue.scala From sparker with GNU General Public License v3.0 | 5 votes |
package Utilities import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 12
Source File: BoundedPriorityQueue.scala From sparker with GNU General Public License v3.0 | 5 votes |
package Utilities import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 13
Source File: BoundedPriorityQueue.scala From sparker with GNU General Public License v3.0 | 5 votes |
package SparkER.Utilities import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 14
Source File: BoundedPriorityQueue.scala From SparkCore with Apache License 2.0 | 5 votes |
package org.apache.spark.util import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable private[spark] class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 15
Source File: BoundedPriorityQueue.scala From sparkoscope with Apache License 2.0 | 5 votes |
package org.apache.spark.util import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable private[spark] class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 16
Source File: collect.scala From XSQL with Apache License 2.0 | 5 votes |
package org.apache.spark.sql.catalyst.expressions.aggregate import scala.collection.generic.Growable import scala.collection.mutable import org.apache.spark.sql.catalyst.InternalRow import org.apache.spark.sql.catalyst.analysis.TypeCheckResult import org.apache.spark.sql.catalyst.expressions._ import org.apache.spark.sql.catalyst.util.GenericArrayData import org.apache.spark.sql.types._ @ExpressionDescription( usage = "_FUNC_(expr) - Collects and returns a set of unique elements.") case class CollectSet( child: Expression, mutableAggBufferOffset: Int = 0, inputAggBufferOffset: Int = 0) extends Collect[mutable.HashSet[Any]] { def this(child: Expression) = this(child, 0, 0) override def checkInputDataTypes(): TypeCheckResult = { if (!child.dataType.existsRecursively(_.isInstanceOf[MapType])) { TypeCheckResult.TypeCheckSuccess } else { TypeCheckResult.TypeCheckFailure("collect_set() cannot have map type data") } } override def withNewMutableAggBufferOffset(newMutableAggBufferOffset: Int): ImperativeAggregate = copy(mutableAggBufferOffset = newMutableAggBufferOffset) override def withNewInputAggBufferOffset(newInputAggBufferOffset: Int): ImperativeAggregate = copy(inputAggBufferOffset = newInputAggBufferOffset) override def prettyName: String = "collect_set" override def createAggregationBuffer(): mutable.HashSet[Any] = mutable.HashSet.empty }
Example 17
Source File: BoundedPriorityQueue.scala From incubator-s2graph with Apache License 2.0 | 5 votes |
package org.apache.s2graph.s2jobs.wal.utils import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }
Example 18
Source File: BoundedPriorityQueue.scala From drizzle-spark with Apache License 2.0 | 5 votes |
package org.apache.spark.util import java.io.Serializable import java.util.{PriorityQueue => JPriorityQueue} import scala.collection.JavaConverters._ import scala.collection.generic.Growable private[spark] class BoundedPriorityQueue[A](maxSize: Int)(implicit ord: Ordering[A]) extends Iterable[A] with Growable[A] with Serializable { private val underlying = new JPriorityQueue[A](maxSize, ord) override def iterator: Iterator[A] = underlying.iterator.asScala override def size: Int = underlying.size override def ++=(xs: TraversableOnce[A]): this.type = { xs.foreach { this += _ } this } override def +=(elem: A): this.type = { if (size < maxSize) { underlying.offer(elem) } else { maybeReplaceLowest(elem) } this } override def +=(elem1: A, elem2: A, elems: A*): this.type = { this += elem1 += elem2 ++= elems } override def clear() { underlying.clear() } private def maybeReplaceLowest(a: A): Boolean = { val head = underlying.peek() if (head != null && ord.gt(a, head)) { underlying.poll() underlying.offer(a) } else { false } } }