Binary compatibility report for the sparkling-water-core_2.10-1.3.7 library  between 1.3.0 and 1.4.0 versions   (relating to the portability of client application sparkling-water-core_2.10-1.3.7.jar)

Test Info


Library Namesparkling-water-core_2.10-1.3.7
Version #11.3.0
Version #21.4.0
Java Version1.7.0_75

Test Results


Total Java ARchives3
Total Methods / Classes598 / 2606
VerdictIncompatible
(0.8%)

Problem Summary


SeverityCount
Added Methods-67
Removed MethodsHigh5
Problems with
Data Types
High0
Medium0
Low0
Problems with
Methods
High0
Medium0
Low0
Other Changes
in Data Types
-1

Added Methods (67)


spark-core_2.10-1.4.0.jar, SparkConf.class
package org.apache.spark
SparkConf.getDeprecatedConfig ( String p1, SparkConf p2 ) [static]  :  scala.Option<String>
SparkConf.getSizeAsBytes ( String key )  :  long
SparkConf.getSizeAsBytes ( String key, String defaultValue )  :  long
SparkConf.getSizeAsGb ( String key )  :  long
SparkConf.getSizeAsGb ( String key, String defaultValue )  :  long
SparkConf.getSizeAsKb ( String key )  :  long
SparkConf.getSizeAsKb ( String key, String defaultValue )  :  long
SparkConf.getSizeAsMb ( String key )  :  long
SparkConf.getSizeAsMb ( String key, String defaultValue )  :  long
SparkConf.getTimeAsMs ( String key )  :  long
SparkConf.getTimeAsMs ( String key, String defaultValue )  :  long
SparkConf.getTimeAsSeconds ( String key )  :  long
SparkConf.getTimeAsSeconds ( String key, String defaultValue )  :  long
SparkConf.logDeprecationWarning ( String p1 ) [static]  :  void

spark-core_2.10-1.4.0.jar, SparkContext.class
package org.apache.spark
SparkContext.applicationAttemptId ( )  :  scala.Option<String>
SparkContext.externalBlockStoreFolderName ( )  :  String
SparkContext.getOrCreate ( ) [static]  :  SparkContext
SparkContext.getOrCreate ( SparkConf p1 ) [static]  :  SparkContext
SparkContext.SparkContext.._conf ( )  :  SparkConf
SparkContext.SparkContext.._env ( )  :  SparkEnv
SparkContext.SparkContext..assertNotStopped ( )  :  void
SparkContext.range ( long start, long end, long step, int numSlices )  :  rdd.RDD<Object>
SparkContext.setLogLevel ( String logLevel )  :  void
SparkContext.supportDynamicAllocation ( )  :  boolean
SparkContext.withScope ( scala.Function0<U> body )  :  U

spark-core_2.10-1.4.0.jar, SparkEnv.class
package org.apache.spark
SparkEnv.executorMemoryManager ( )  :  unsafe.memory.ExecutorMemoryManager
SparkEnv.SparkEnv..driverTmpDirToDelete_.eq ( scala.Option<String> p1 )  :  void
SparkEnv.rpcEnv ( )  :  rpc.RpcEnv
SparkEnv.SparkEnv ( String executorId, rpc.RpcEnv rpcEnv, serializer.Serializer serializer, serializer.Serializer closureSerializer, CacheManager cacheManager, MapOutputTracker mapOutputTracker, shuffle.ShuffleManager shuffleManager, broadcast.BroadcastManager broadcastManager, network.BlockTransferService blockTransferService, storage.BlockManager blockManager, SecurityManager securityManager, HttpFileServer httpFileServer, String sparkFilesDir, metrics.MetricsSystem metricsSystem, shuffle.ShuffleMemoryManager shuffleMemoryManager, unsafe.memory.ExecutorMemoryManager executorMemoryManager, scheduler.OutputCommitCoordinator outputCommitCoordinator, SparkConf conf )

spark-core_2.10-1.4.0.jar, TaskContext.class
package org.apache.spark
TaskContext.taskMemoryManager ( ) [abstract]  :  unsafe.memory.TaskMemoryManager

spark-sql_2.10-1.4.0.jar, DataFrame.class
package org.apache.spark.sql
DataFrame.coalesce ( int numPartitions )  :  DataFrame
DataFrame.cube ( Column... cols )  :  GroupedData
DataFrame.cube ( scala.collection.Seq<Column> cols )  :  GroupedData
DataFrame.cube ( String col1, scala.collection.Seq<String> cols )  :  GroupedData
DataFrame.cube ( String col1, String... cols )  :  GroupedData
DataFrame.describe ( scala.collection.Seq<String> cols )  :  DataFrame
DataFrame.describe ( String... cols )  :  DataFrame
DataFrame.drop ( String colName )  :  DataFrame
DataFrame.dropDuplicates ( )  :  DataFrame
DataFrame.dropDuplicates ( scala.collection.Seq<String> colNames )  :  DataFrame
DataFrame.dropDuplicates ( String[ ] colNames )  :  DataFrame
DataFrame.join ( DataFrame right, String usingColumn )  :  DataFrame
DataFrame.na ( )  :  DataFrameNaFunctions
DataFrame.DataFrame..logicalPlanToDataFrame ( catalyst.plans.logical.LogicalPlan logicalPlan )  :  DataFrame
DataFrame.randomSplit ( double[ ] weights )  :  DataFrame[ ]
DataFrame.randomSplit ( double[ ] weights, long seed )  :  DataFrame[ ]
DataFrame.randomSplit ( scala.collection.immutable.List<Object> weights, long seed )  :  DataFrame[ ]
DataFrame.rollup ( Column... cols )  :  GroupedData
DataFrame.rollup ( scala.collection.Seq<Column> cols )  :  GroupedData
DataFrame.rollup ( String col1, scala.collection.Seq<String> cols )  :  GroupedData
DataFrame.rollup ( String col1, String... cols )  :  GroupedData
DataFrame.stat ( )  :  DataFrameStatFunctions
DataFrame.write ( )  :  DataFrameWriter

spark-sql_2.10-1.4.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.cacheManager ( )  :  execution.CacheManager
SQLContext.createDataFrame ( org.apache.spark.rdd.RDD<Row> rowRDD, types.StructType schema, boolean needsConversion )  :  DataFrame
SQLContext.createSession ( )  :  SQLContext.SQLSession
SQLContext.currentSession ( )  :  SQLContext.SQLSession
SQLContext.defaultSession ( )  :  SQLContext.SQLSession
SQLContext.detachSession ( )  :  void
SQLContext.dialectClassName ( )  :  String
SQLContext.getOrCreate ( org.apache.spark.SparkContext p1 ) [static]  :  SQLContext
SQLContext.getSQLDialect ( )  :  catalyst.ParserDialect
SQLContext.openSession ( )  :  SQLContext.SQLSession
SQLContext.range ( long start, long end )  :  DataFrame
SQLContext.range ( long start, long end, long step, int numPartitions )  :  DataFrame
SQLContext.read ( )  :  DataFrameReader
SQLContext.tlSession ( )  :  ThreadLocal<SQLContext.SQLSession>

to the top

Removed Methods (5)


spark-core_2.10-1.3.0.jar, SparkConf.class
package org.apache.spark
SparkConf.translateConfKey ( String p1, boolean p2 ) [static]  :  String

spark-core_2.10-1.3.0.jar, SparkEnv.class
package org.apache.spark
SparkEnv.SparkEnv ( String executorId, akka.actor.ActorSystem actorSystem, serializer.Serializer serializer, serializer.Serializer closureSerializer, CacheManager cacheManager, MapOutputTracker mapOutputTracker, shuffle.ShuffleManager shuffleManager, broadcast.BroadcastManager broadcastManager, network.BlockTransferService blockTransferService, storage.BlockManager blockManager, SecurityManager securityManager, HttpFileServer httpFileServer, String sparkFilesDir, metrics.MetricsSystem metricsSystem, shuffle.ShuffleMemoryManager shuffleMemoryManager, scheduler.OutputCommitCoordinator outputCommitCoordinator, SparkConf conf )

spark-sql_2.10-1.3.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.cacheManager ( )  :  CacheManager
SQLContext.checkAnalysis ( )  :  catalyst.analysis.CheckAnalysis
SQLContext.createDataFrame ( org.apache.spark.api.java.JavaRDD<Row> rowRDD, java.util.List<String> columns )  :  DataFrame

to the top

Other Changes in Data Types (1)


spark-core_2.10-1.3.0.jar
package org.apache.spark
[+] TaskContext (1)

to the top

Java ARchives (3)


spark-core_2.10-1.3.0.jar
spark-mllib_2.10-1.3.0.jar
spark-sql_2.10-1.3.0.jar

to the top




Generated on Mon Jul 6 17:29:06 2015 for sparkling-water-core_2.10-1.3.7 by Java API Compliance Checker 1.4.1  
A tool for checking backward compatibility of a Java library API