diff --git a/.travis.yml b/.travis.yml index 3b9388af..b236dd26 100644 --- a/.travis.yml +++ b/.travis.yml @@ -5,7 +5,7 @@ import: scala/scala-dev:travis/default.yml language: scala scala: - - 2.13.2 + - 2.13.3 env: - ADOPTOPENJDK=8 diff --git a/core/src/main/scala/scala/collection/generic/GenericTraversableTemplate.scala b/core/src/main/scala/scala/collection/generic/GenericTraversableTemplate.scala index b78e83d8..a1524eb1 100644 --- a/core/src/main/scala/scala/collection/generic/GenericTraversableTemplate.scala +++ b/core/src/main/scala/scala/collection/generic/GenericTraversableTemplate.scala @@ -226,7 +226,7 @@ trait GenericTraversableTemplate[+A, +CC[X] <: ParIterable[X]] extends HasNewBui fail } val bb = genericBuilder[CC[B]] - for (b <- bs) bb += b.result + for (b <- bs) bb += b.result() bb.result() } } diff --git a/core/src/main/scala/scala/collection/parallel/ParIterable.scala b/core/src/main/scala/scala/collection/parallel/ParIterable.scala index 8686d5cc..86ece1de 100644 --- a/core/src/main/scala/scala/collection/parallel/ParIterable.scala +++ b/core/src/main/scala/scala/collection/parallel/ParIterable.scala @@ -37,8 +37,8 @@ trait ParIterable[+T] object ParIterable extends ParFactory[ParIterable] { implicit def canBuildFrom[T]: CanCombineFrom[ParIterable[_], T, ParIterable[T]] = new GenericCanCombineFrom[T] - def newBuilder[T]: Combiner[T, ParIterable[T]] = ParArrayCombiner[T] + def newBuilder[T]: Combiner[T, ParIterable[T]] = ParArrayCombiner[T]() - def newCombiner[T]: Combiner[T, ParIterable[T]] = ParArrayCombiner[T] + def newCombiner[T]: Combiner[T, ParIterable[T]] = ParArrayCombiner[T]() } diff --git a/core/src/main/scala/scala/collection/parallel/ParIterableLike.scala b/core/src/main/scala/scala/collection/parallel/ParIterableLike.scala index 343fc0e4..d8d757d5 100644 --- a/core/src/main/scala/scala/collection/parallel/ParIterableLike.scala +++ b/core/src/main/scala/scala/collection/parallel/ParIterableLike.scala @@ -626,7 +626,7 @@ self => val it = splitter var left = n while (left > 0) { - cb += it.next + cb += it.next() left -= 1 } cb.resultWithTaskSupport @@ -642,7 +642,7 @@ self => val it = splitter drop n val cb = newCombiner cb.sizeHint(size - n) - while (it.hasNext) cb += it.next + while (it.hasNext) cb += it.next() cb.resultWithTaskSupport } @@ -658,7 +658,7 @@ self => var left = until - from val it = splitter drop from while (left > 0) { - cb += it.next + cb += it.next() left -= 1 } cb.resultWithTaskSupport @@ -1291,7 +1291,7 @@ self => @volatile var result: Result = null def leaf(prev: Option[Combiner[U, That]]): Unit = { result = cbf() - while (pit.hasNext) result += pit.next + while (pit.hasNext) result += pit.next() } protected[this] def newSubtask(p: IterableSplitter[T]) = new ToParCollection[U, That](cbf, p) override def merge(that: ToParCollection[U, That]) = result = result combine that.result @@ -1302,7 +1302,7 @@ self => @volatile var result: Result = null def leaf(prev: Option[Combiner[(K, V), That]]): Unit = { result = cbf() - while (pit.hasNext) result += pit.next + while (pit.hasNext) result += pit.next() } protected[this] def newSubtask(p: IterableSplitter[T]) = new ToParMap[K, V, That](cbf, p)(ev) override def merge(that: ToParMap[K, V, That]) = result = result combine that.result diff --git a/core/src/main/scala/scala/collection/parallel/ParSeq.scala b/core/src/main/scala/scala/collection/parallel/ParSeq.scala index bdd9f456..dba70f37 100644 --- a/core/src/main/scala/scala/collection/parallel/ParSeq.scala +++ b/core/src/main/scala/scala/collection/parallel/ParSeq.scala @@ -44,6 +44,6 @@ trait ParSeq[+T] extends ParIterable[T] object ParSeq extends ParFactory[ParSeq] { implicit def canBuildFrom[T]: CanCombineFrom[ParSeq[_], T, ParSeq[T]] = new GenericCanCombineFrom[T] - def newBuilder[T]: Combiner[T, ParSeq[T]] = ParArrayCombiner[T] - def newCombiner[T]: Combiner[T, ParSeq[T]] = ParArrayCombiner[T] + def newBuilder[T]: Combiner[T, ParSeq[T]] = ParArrayCombiner[T]() + def newCombiner[T]: Combiner[T, ParSeq[T]] = ParArrayCombiner[T]() } diff --git a/core/src/main/scala/scala/collection/parallel/RemainsIterator.scala b/core/src/main/scala/scala/collection/parallel/RemainsIterator.scala index 2a45d708..b852338e 100644 --- a/core/src/main/scala/scala/collection/parallel/RemainsIterator.scala +++ b/core/src/main/scala/scala/collection/parallel/RemainsIterator.scala @@ -126,7 +126,7 @@ private[collection] trait AugmentedIterableIterator[+T] extends RemainsIterator[ def copy2builder[U >: T, Coll, Bld <: Builder[U, Coll]](b: Bld): Bld = { if (isRemainingCheap) b.sizeHint(remaining) - while (hasNext) b += next + while (hasNext) b += next() b } @@ -159,7 +159,7 @@ private[collection] trait AugmentedIterableIterator[+T] extends RemainsIterator[ cb.sizeHint(n) var left = n while (left > 0) { - cb += next + cb += next() left -= 1 } cb @@ -168,7 +168,7 @@ private[collection] trait AugmentedIterableIterator[+T] extends RemainsIterator[ def drop2combiner[U >: T, This](n: Int, cb: Combiner[U, This]): Combiner[U, This] = { drop(n) if (isRemainingCheap) cb.sizeHint(remaining) - while (hasNext) cb += next + while (hasNext) cb += next() cb } @@ -177,7 +177,7 @@ private[collection] trait AugmentedIterableIterator[+T] extends RemainsIterator[ var left = scala.math.max(until - from, 0) cb.sizeHint(left) while (left > 0) { - cb += next + cb += next() left -= 1 } cb @@ -188,10 +188,10 @@ private[collection] trait AugmentedIterableIterator[+T] extends RemainsIterator[ if (isRemainingCheap) after.sizeHint(remaining - at) var left = at while (left > 0) { - before += next + before += next() left -= 1 } - while (hasNext) after += next + while (hasNext) after += next() (before, after) } @@ -216,7 +216,7 @@ private[collection] trait AugmentedIterableIterator[+T] extends RemainsIterator[ isBefore = false } } - while (hasNext) after += next + while (hasNext) after += next() (before, after) } @@ -320,7 +320,7 @@ private[collection] trait AugmentedSeqIterator[+T] extends AugmentedIterableIter def reverse2combiner[U >: T, This](cb: Combiner[U, This]): Combiner[U, This] = { if (isRemainingCheap) cb.sizeHint(remaining) var lst = List[T]() - while (hasNext) lst ::= next + while (hasNext) lst ::= next() while (lst != Nil) { cb += lst.head lst = lst.tail @@ -348,7 +348,7 @@ private[collection] trait AugmentedSeqIterator[+T] extends AugmentedIterableIter if (j == index) { cb += elem next() - } else cb += next + } else cb += next() j += 1 } cb @@ -439,7 +439,7 @@ self => private[collection] def newSliceInternal[U <: Taken](it: U, from1: Int): U = { var count = from1 while (count > 0 && it.hasNext) { - it.next + it.next() count -= 1 } it @@ -656,7 +656,7 @@ self => (pits(0).appendParSeq[U, SeqSplitter[U]](patch)) appendParSeq pits(2) } def hasNext = trio.hasNext - def next() = trio.next + def next() = trio.next() def remaining = trio.remaining def dup = self.dup.patchParSeq(from, patch, replaced) def split = trio.split diff --git a/core/src/main/scala/scala/collection/parallel/immutable/ParHashMap.scala b/core/src/main/scala/scala/collection/parallel/immutable/ParHashMap.scala index 3316051e..5a742084 100644 --- a/core/src/main/scala/scala/collection/parallel/immutable/ParHashMap.scala +++ b/core/src/main/scala/scala/collection/parallel/immutable/ParHashMap.scala @@ -306,10 +306,10 @@ extends scala.collection.parallel.BucketCombiner[(K, V), ParHashMap[K, V], (K, V } private def evaluateCombiners(trie: OldHashMap[K, Combiner[V, Repr]]): OldHashMap[K, Repr] = trie match { case hm1: OldHashMap.OldHashMap1[_, _] => - val evaledvalue = hm1.value.result + val evaledvalue = hm1.value.result() new OldHashMap.OldHashMap1[K, Repr](hm1.key, hm1.hash, evaledvalue, null) case hmc: OldHashMap.OldHashMapCollision1[_, _] => - val evaledkvs = hmc.kvs map { p => (p._1, p._2.result) } + val evaledkvs = hmc.kvs map { p => (p._1, p._2.result()) } new OldHashMap.OldHashMapCollision1[K, Repr](hmc.hash, evaledkvs) case htm: OldHashMap.HashTrieMap[k, v] => var i = 0 diff --git a/core/src/main/scala/scala/collection/parallel/immutable/ParRange.scala b/core/src/main/scala/scala/collection/parallel/immutable/ParRange.scala index 38445be8..fb28a325 100644 --- a/core/src/main/scala/scala/collection/parallel/immutable/ParRange.scala +++ b/core/src/main/scala/scala/collection/parallel/immutable/ParRange.scala @@ -104,7 +104,7 @@ self => override def map2combiner[S, That](f: Int => S, cb: Combiner[S, That]): Combiner[S, That] = { while (hasNext) { - cb += f(next) + cb += f(next()) } cb } diff --git a/core/src/main/scala/scala/collection/parallel/immutable/ParVector.scala b/core/src/main/scala/scala/collection/parallel/immutable/ParVector.scala index bed7e7ac..fcc6f2c9 100644 --- a/core/src/main/scala/scala/collection/parallel/immutable/ParVector.scala +++ b/core/src/main/scala/scala/collection/parallel/immutable/ParVector.scala @@ -122,9 +122,9 @@ private[immutable] class LazyParVectorCombiner[T] extends Combiner[T, ParVector[ def result(): ParVector[T] = { val rvb = new VectorBuilder[T] for (vb <- vectors) { - rvb ++= vb.result + rvb ++= vb.result() } - new ParVector(rvb.result) + new ParVector(rvb.result()) } def combine[U <: T, NewTo >: ParVector[T]](other: Combiner[U, NewTo]) = if (other eq this) this else { diff --git a/core/src/main/scala/scala/collection/parallel/mutable/ParArray.scala b/core/src/main/scala/scala/collection/parallel/mutable/ParArray.scala index 5df6da66..2050a184 100644 --- a/core/src/main/scala/scala/collection/parallel/mutable/ParArray.scala +++ b/core/src/main/scala/scala/collection/parallel/mutable/ParArray.scala @@ -375,7 +375,7 @@ self => var same = true val thatIt = that.iterator while (i < until && thatIt.hasNext) { - if (arr(i) != thatIt.next) { + if (arr(i) != thatIt.next()) { i = until same = false } @@ -693,7 +693,7 @@ self => object ParArray extends ParFactory[ParArray] { implicit def canBuildFrom[T]: CanCombineFrom[ParArray[_], T, ParArray[T]] = new GenericCanCombineFrom[T] def newBuilder[T]: Combiner[T, ParArray[T]] = newCombiner - def newCombiner[T]: Combiner[T, ParArray[T]] = ParArrayCombiner[T] + def newCombiner[T]: Combiner[T, ParArray[T]] = ParArrayCombiner[T]() /** Creates a new parallel array by wrapping the specified array. */ @@ -720,7 +720,7 @@ object ParArray extends ParFactory[ParArray] { for (xs <- xss) { cb ++= xs } - cb.result + cb.result() } } diff --git a/core/src/main/scala/scala/collection/parallel/mutable/ParIterable.scala b/core/src/main/scala/scala/collection/parallel/mutable/ParIterable.scala index cbf74fbc..b432e74d 100644 --- a/core/src/main/scala/scala/collection/parallel/mutable/ParIterable.scala +++ b/core/src/main/scala/scala/collection/parallel/mutable/ParIterable.scala @@ -44,6 +44,6 @@ trait ParIterable[T] extends scala.collection.parallel.ParIterable[T] object ParIterable extends ParFactory[ParIterable] { implicit def canBuildFrom[T]: CanCombineFrom[ParIterable[_], T, ParIterable[T]] = new GenericCanCombineFrom[T] - def newBuilder[T]: Combiner[T, ParIterable[T]] = ParArrayCombiner[T] - def newCombiner[T]: Combiner[T, ParIterable[T]] = ParArrayCombiner[T] + def newBuilder[T]: Combiner[T, ParIterable[T]] = ParArrayCombiner[T]() + def newCombiner[T]: Combiner[T, ParIterable[T]] = ParArrayCombiner[T]() } diff --git a/core/src/main/scala/scala/collection/parallel/mutable/ParSeq.scala b/core/src/main/scala/scala/collection/parallel/mutable/ParSeq.scala index afa04345..54d6eeda 100644 --- a/core/src/main/scala/scala/collection/parallel/mutable/ParSeq.scala +++ b/core/src/main/scala/scala/collection/parallel/mutable/ParSeq.scala @@ -48,7 +48,7 @@ self => object ParSeq extends ParFactory[ParSeq] { implicit def canBuildFrom[T]: CanCombineFrom[ParSeq[_], T, ParSeq[T]] = new GenericCanCombineFrom[T] - def newBuilder[T]: Combiner[T, ParSeq[T]] = ParArrayCombiner[T] + def newBuilder[T]: Combiner[T, ParSeq[T]] = ParArrayCombiner[T]() - def newCombiner[T]: Combiner[T, ParSeq[T]] = ParArrayCombiner[T] + def newCombiner[T]: Combiner[T, ParSeq[T]] = ParArrayCombiner[T]() } diff --git a/junit/src/test/scala/scala/SerializationStabilityTest.scala b/junit/src/test/scala/scala/SerializationStabilityTest.scala index ce26e82b..02c586f2 100644 --- a/junit/src/test/scala/scala/SerializationStabilityTest.scala +++ b/junit/src/test/scala/scala/SerializationStabilityTest.scala @@ -35,7 +35,7 @@ object SerializationStability extends App { } def amend(file: File)(f: String => String): Unit = { - file.writeAll(f(file.slurp)) + file.writeAll(f(file.slurp())) } def quote(s: String) = List("\"", s, "\"").mkString @@ -44,7 +44,7 @@ object SerializationStability extends App { content => content.linesIterator.toList.zipWithIndex.map { case (content, i) if i == line - 1 => - val newContent = content.replaceAllLiterally(quote(prevResult), quote(result)) + val newContent = content.replace(quote(prevResult), quote(result)) if (newContent != content) println(s"- $content\n+ $newContent\n") newContent @@ -113,4 +113,4 @@ object SerializationStability extends App { class SerializationStabilityTest { @Test def testAll: Unit = SerializationStability.main(new Array[String](0)) -} \ No newline at end of file +} diff --git a/junit/src/test/scala/scala/collection/concurrent/ctries_new/IteratorSpec.scala b/junit/src/test/scala/scala/collection/concurrent/ctries_new/IteratorSpec.scala index b425d6c2..58fe5dae 100644 --- a/junit/src/test/scala/scala/collection/concurrent/ctries_new/IteratorSpec.scala +++ b/junit/src/test/scala/scala/collection/concurrent/ctries_new/IteratorSpec.scala @@ -25,7 +25,7 @@ class IteratorSpec extends Spec { val it = ct.iterator it.hasNext shouldEqual (false) - evaluating { it.next() }.shouldProduce [NoSuchElementException] + evaluating { it.next() }.shouldProduce[NoSuchElementException]() } def nonEmptyIteratorCheck(sz: Int): Unit = { @@ -36,11 +36,11 @@ class IteratorSpec extends Spec { val tracker = mutable.Map[Wrap, Int]() for (i <- 0 until sz) { assert(it.hasNext == true) - tracker += it.next + tracker += it.next() } it.hasNext shouldEqual (false) - evaluating { it.next() }.shouldProduce [NoSuchElementException] + evaluating { it.next() }.shouldProduce[NoSuchElementException]() tracker.size shouldEqual (sz) tracker shouldEqual (ct) } @@ -101,11 +101,11 @@ class IteratorSpec extends Spec { val tracker = mutable.Map[DumbHash, Int]() for (i <- 0 until sz) { assert(it.hasNext == true) - tracker += it.next + tracker += it.next() } it.hasNext shouldEqual (false) - evaluating { it.next() }.shouldProduce [NoSuchElementException] + evaluating { it.next() }.shouldProduce[NoSuchElementException]() tracker.size shouldEqual (sz) tracker shouldEqual (ct) } diff --git a/junit/src/test/scala/scala/collection/concurrent/ctries_old/IteratorSpec.scala b/junit/src/test/scala/scala/collection/concurrent/ctries_old/IteratorSpec.scala index bf77df73..b46edfd3 100644 --- a/junit/src/test/scala/scala/collection/concurrent/ctries_old/IteratorSpec.scala +++ b/junit/src/test/scala/scala/collection/concurrent/ctries_old/IteratorSpec.scala @@ -26,7 +26,7 @@ class IteratorSpec extends Spec { val it = ct.iterator it.hasNext shouldEqual (false) - evaluating { it.next() }.shouldProduce [NoSuchElementException] + evaluating { it.next() }.shouldProduce[NoSuchElementException]() } def nonEmptyIteratorCheck(sz: Int): Unit = { @@ -37,11 +37,11 @@ class IteratorSpec extends Spec { val tracker = mutable.Map[Wrap, Int]() for (i <- 0 until sz) { assert(it.hasNext == true) - tracker += it.next + tracker += it.next() } it.hasNext shouldEqual (false) - evaluating { it.next() }.shouldProduce [NoSuchElementException] + evaluating { it.next() }.shouldProduce[NoSuchElementException]() tracker.size shouldEqual (sz) tracker shouldEqual (ct) } @@ -102,11 +102,11 @@ class IteratorSpec extends Spec { val tracker = mutable.Map[DumbHash, Int]() for (i <- 0 until sz) { assert(it.hasNext == true) - tracker += it.next + tracker += it.next() } it.hasNext shouldEqual (false) - evaluating { it.next() }.shouldProduce [NoSuchElementException] + evaluating { it.next() }.shouldProduce[NoSuchElementException]() tracker.size shouldEqual (sz) tracker shouldEqual (ct) } diff --git a/scalacheck/src/test/scala/ParallelHashTrieCheck.scala b/scalacheck/src/test/scala/ParallelHashTrieCheck.scala index a09712e8..bebb0a97 100644 --- a/scalacheck/src/test/scala/ParallelHashTrieCheck.scala +++ b/scalacheck/src/test/scala/ParallelHashTrieCheck.scala @@ -69,7 +69,7 @@ with PairValues[Int, Int] override def printDataStructureDebugInfo(ds: AnyRef) = ds match { case pm: ParHashMap[k, v] => - pm.printDebugInfo + pm.printDebugInfo() case _ => println("could not match data structure type: " + ds.getClass) } @@ -124,7 +124,7 @@ with IntValues override def printDataStructureDebugInfo(ds: AnyRef) = ds match { case pm: ParHashMap[k, v] => - pm.printDebugInfo + pm.printDebugInfo() case _ => println("could not match data structure type: " + ds.getClass) } diff --git a/scalacheck/src/test/scala/ParallelIterableCheck.scala b/scalacheck/src/test/scala/ParallelIterableCheck.scala index 25ab1ee8..14ab7611 100644 --- a/scalacheck/src/test/scala/ParallelIterableCheck.scala +++ b/scalacheck/src/test/scala/ParallelIterableCheck.scala @@ -94,7 +94,7 @@ abstract class ParallelIterableCheck[T](collName: String) extends Properties(col def printDebugInfo[A, CC[X] <: ParIterable[X], C <: ParIterable[A], S <: Iterable[A] with IterableOps[A, Iterable, S]](coll: ParIterableLike[A, CC, C, S]): Unit = { println("Collection debug info: ") - coll.printDebugBuffer + coll.printDebugBuffer() println("Task debug info: ") println(coll.tasksupport.debugMessages.mkString("\n")) } @@ -104,17 +104,17 @@ abstract class ParallelIterableCheck[T](collName: String) extends Properties(col println("Operator: " + ind) println("sz: " + t.size) println(t) - println + println() println("sz: " + coll.size) println(coll) println("transformed to:") - println + println() println("size: " + tf.size) println(tf) - println + println() println("size: " + cf.size) println(cf) - println + println() println("tf sameElements cf - " + (tf.iterator sameElements cf)) println("cf sameElements tf - " + (cf.iterator sameElements tf)) } @@ -226,16 +226,16 @@ abstract class ParallelIterableCheck[T](collName: String) extends Properties(col println("Operator: " + ind) println("sz: " + t.size) println(t) - println + println() println("sz: " + coll.size) println(coll) - println + println() println("filtered to:") - println + println() println(cf) - println + println() println(tf) - println + println() println("areEqual(tf, cf) - " + areEqual(tf, cf)) printDataStructureDebugInfo(cf) println("valid: " + invs) @@ -289,11 +289,11 @@ abstract class ParallelIterableCheck[T](collName: String) extends Properties(col println(collsl) println("as list: " + collsl.toList) println(collsl.iterator.hasNext) - println(collsl.iterator.next) + println(collsl.iterator.next()) println(collsl.iterator.hasNext) - println(collsl.iterator.next) + println(collsl.iterator.next()) println(collsl.iterator.hasNext) - println(collsl.iterator.next) + println(collsl.iterator.next()) println(collsl.iterator.hasNext) } ("slice from " + from + " until " + until) |: tsl.iterator.sameElements(collsl) diff --git a/scalacheck/src/test/scala/ParallelVectorCheck.scala b/scalacheck/src/test/scala/ParallelVectorCheck.scala index 14407024..ed60cf83 100644 --- a/scalacheck/src/test/scala/ParallelVectorCheck.scala +++ b/scalacheck/src/test/scala/ParallelVectorCheck.scala @@ -44,13 +44,13 @@ abstract class ParallelVectorCheck[T](tp: String) extends collection.parallel.Pa val vb = new immutable.VectorBuilder[T]() val gen = vals(rnd.nextInt(vals.size)) for (i <- 0 until sz) vb += sample(gen) - vb.result + vb.result() } def fromSeq(a: Seq[T]) = { val pc = ParVector.newCombiner[T] for (elem <- a.toList) pc += elem - val pv = pc.result + val pv = pc.result() pv.tasksupport = tasksupport pv }