Binary compatibility report for the sparkling-water-core_2.10-1.4.3 library between 1.4.0 and 1.3.0 versions (relating to the portability of client application sparkling-water-core_2.10-1.4.3.jar)
Test Info
Library Name | sparkling-water-core_2.10-1.4.3 |
Version #1 | 1.4.0 |
Version #2 | 1.3.0 |
Java Version | 1.7.0_75 |
Test Results
Total Java ARchives | 3 |
---|
Total Methods / Classes | 699 / 3063 |
---|
Verdict | Incompatible (14.2%) |
Problem Summary
| Severity | Count |
---|
Added Methods | - | 5 |
---|
Removed Methods | High | 73 |
---|
Problems with Data Types | High | 4 |
---|
Medium | 1 |
Low | 0 |
Problems with Methods | High | 0 |
---|
Medium | 0 |
Low | 0 |
Added Methods (5)
spark-core_2.10-1.3.0.jar, SparkConf.class
package org.apache.spark
SparkConf.translateConfKey ( String p1, boolean p2 ) [static] : String
[mangled: org/apache/spark/SparkConf.translateConfKey:(Ljava/lang/String;Z)Ljava/lang/String;]
spark-core_2.10-1.3.0.jar, SparkEnv.class
package org.apache.spark
SparkEnv.SparkEnv ( String executorId, akka.actor.ActorSystem actorSystem, serializer.Serializer serializer, serializer.Serializer closureSerializer, CacheManager cacheManager, MapOutputTracker mapOutputTracker, shuffle.ShuffleManager shuffleManager, broadcast.BroadcastManager broadcastManager, network.BlockTransferService blockTransferService, storage.BlockManager blockManager, SecurityManager securityManager, HttpFileServer httpFileServer, String sparkFilesDir, metrics.MetricsSystem metricsSystem, shuffle.ShuffleMemoryManager shuffleMemoryManager, scheduler.OutputCommitCoordinator outputCommitCoordinator, SparkConf conf )
[mangled: org/apache/spark/SparkEnv."<init>":(Ljava/lang/String;Lakka/actor/ActorSystem;Lorg/apache/spark/serializer/Serializer;Lorg/apache/spark/serializer/Serializer;Lorg/apache/spark/CacheManager;Lorg/apache/spark/MapOutputTracker;Lorg/apache/spark/shuffle/ShuffleManager;Lorg/apache/spark/broadcast/BroadcastManager;Lorg/apache/spark/network/BlockTransferService;Lorg/apache/spark/storage/BlockManager;Lorg/apache/spark/SecurityManager;Lorg/apache/spark/HttpFileServer;Ljava/lang/String;Lorg/apache/spark/metrics/MetricsSystem;Lorg/apache/spark/shuffle/ShuffleMemoryManager;Lorg/apache/spark/scheduler/OutputCommitCoordinator;Lorg/apache/spark/SparkConf;)V]
spark-sql_2.10-1.3.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.cacheManager ( ) : CacheManager
[mangled: org/apache/spark/sql/SQLContext.cacheManager:()Lorg/apache/spark/sql/CacheManager;]
SQLContext.checkAnalysis ( ) : catalyst.analysis.CheckAnalysis
[mangled: org/apache/spark/sql/SQLContext.checkAnalysis:()Lorg/apache/spark/sql/catalyst/analysis/CheckAnalysis;]
SQLContext.createDataFrame ( org.apache.spark.api.java.JavaRDD<Row> rowRDD, java.util.List<String> columns ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/api/java/JavaRDD;Ljava/util/List;)Lorg/apache/spark/sql/DataFrame;]
to the top
Removed Methods (73)
spark-core_2.10-1.4.0.jar, RDD<T>.class
package org.apache.spark.rdd
RDD<T>.RDD..doCheckpointCalled ( ) : boolean
[mangled: org/apache/spark/rdd/RDD<T>.org.apache.spark.rdd.RDD..doCheckpointCalled:()Z]
RDD<T>.RDD..doCheckpointCalled_.eq ( boolean p1 ) : void
[mangled: org/apache/spark/rdd/RDD<T>.org.apache.spark.rdd.RDD..doCheckpointCalled_.eq:(Z)V]
RDD<T>.RDD..sc ( ) : org.apache.spark.SparkContext
[mangled: org/apache/spark/rdd/RDD<T>.org.apache.spark.rdd.RDD..sc:()Lorg/apache/spark/SparkContext;]
RDD<T>.randomSampleWithRange ( double lb, double ub, long seed ) : RDD<T>
[mangled: org/apache/spark/rdd/RDD<T>.randomSampleWithRange:(DDJ)Lorg/apache/spark/rdd/RDD;]
RDD<T>.scope ( ) : scala.Option<RDDOperationScope>
[mangled: org/apache/spark/rdd/RDD<T>.scope:()Lscala/Option;]
RDD<T>.withScope ( scala.Function0<U> body ) : U
[mangled: org/apache/spark/rdd/RDD<T>.withScope:(Lscala/Function0;)Ljava/lang/Object;]
spark-core_2.10-1.4.0.jar, SparkConf.class
package org.apache.spark
SparkConf.getDeprecatedConfig ( String p1, SparkConf p2 ) [static] : scala.Option<String>
[mangled: org/apache/spark/SparkConf.getDeprecatedConfig:(Ljava/lang/String;Lorg/apache/spark/SparkConf;)Lscala/Option;]
SparkConf.getSizeAsBytes ( String key ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsBytes:(Ljava/lang/String;)J]
SparkConf.getSizeAsBytes ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsBytes:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getSizeAsGb ( String key ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsGb:(Ljava/lang/String;)J]
SparkConf.getSizeAsGb ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsGb:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getSizeAsKb ( String key ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsKb:(Ljava/lang/String;)J]
SparkConf.getSizeAsKb ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsKb:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getSizeAsMb ( String key ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsMb:(Ljava/lang/String;)J]
SparkConf.getSizeAsMb ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsMb:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getTimeAsMs ( String key ) : long
[mangled: org/apache/spark/SparkConf.getTimeAsMs:(Ljava/lang/String;)J]
SparkConf.getTimeAsMs ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getTimeAsMs:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getTimeAsSeconds ( String key ) : long
[mangled: org/apache/spark/SparkConf.getTimeAsSeconds:(Ljava/lang/String;)J]
SparkConf.getTimeAsSeconds ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getTimeAsSeconds:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.logDeprecationWarning ( String p1 ) [static] : void
[mangled: org/apache/spark/SparkConf.logDeprecationWarning:(Ljava/lang/String;)V]
spark-core_2.10-1.4.0.jar, SparkContext.class
package org.apache.spark
SparkContext.applicationAttemptId ( ) : scala.Option<String>
[mangled: org/apache/spark/SparkContext.applicationAttemptId:()Lscala/Option;]
SparkContext.externalBlockStoreFolderName ( ) : String
[mangled: org/apache/spark/SparkContext.externalBlockStoreFolderName:()Ljava/lang/String;]
SparkContext.getOrCreate ( ) [static] : SparkContext
[mangled: org/apache/spark/SparkContext.getOrCreate:()Lorg/apache/spark/SparkContext;]
SparkContext.getOrCreate ( SparkConf p1 ) [static] : SparkContext
[mangled: org/apache/spark/SparkContext.getOrCreate:(Lorg/apache/spark/SparkConf;)Lorg/apache/spark/SparkContext;]
SparkContext.SparkContext.._conf ( ) : SparkConf
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext.._conf:()Lorg/apache/spark/SparkConf;]
SparkContext.SparkContext.._env ( ) : SparkEnv
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext.._env:()Lorg/apache/spark/SparkEnv;]
SparkContext.SparkContext..assertNotStopped ( ) : void
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext..assertNotStopped:()V]
SparkContext.range ( long start, long end, long step, int numSlices ) : rdd.RDD<Object>
[mangled: org/apache/spark/SparkContext.range:(JJJI)Lorg/apache/spark/rdd/RDD;]
SparkContext.setLogLevel ( String logLevel ) : void
[mangled: org/apache/spark/SparkContext.setLogLevel:(Ljava/lang/String;)V]
SparkContext.supportDynamicAllocation ( ) : boolean
[mangled: org/apache/spark/SparkContext.supportDynamicAllocation:()Z]
SparkContext.withScope ( scala.Function0<U> body ) : U
[mangled: org/apache/spark/SparkContext.withScope:(Lscala/Function0;)Ljava/lang/Object;]
spark-core_2.10-1.4.0.jar, SparkEnv.class
package org.apache.spark
SparkEnv.executorMemoryManager ( ) : unsafe.memory.ExecutorMemoryManager
[mangled: org/apache/spark/SparkEnv.executorMemoryManager:()Lorg/apache/spark/unsafe/memory/ExecutorMemoryManager;]
SparkEnv.SparkEnv..driverTmpDirToDelete_.eq ( scala.Option<String> p1 ) : void
[mangled: org/apache/spark/SparkEnv.org.apache.spark.SparkEnv..driverTmpDirToDelete_.eq:(Lscala/Option;)V]
SparkEnv.rpcEnv ( ) : rpc.RpcEnv
[mangled: org/apache/spark/SparkEnv.rpcEnv:()Lorg/apache/spark/rpc/RpcEnv;]
SparkEnv.SparkEnv ( String executorId, rpc.RpcEnv rpcEnv, serializer.Serializer serializer, serializer.Serializer closureSerializer, CacheManager cacheManager, MapOutputTracker mapOutputTracker, shuffle.ShuffleManager shuffleManager, broadcast.BroadcastManager broadcastManager, network.BlockTransferService blockTransferService, storage.BlockManager blockManager, SecurityManager securityManager, HttpFileServer httpFileServer, String sparkFilesDir, metrics.MetricsSystem metricsSystem, shuffle.ShuffleMemoryManager shuffleMemoryManager, unsafe.memory.ExecutorMemoryManager executorMemoryManager, scheduler.OutputCommitCoordinator outputCommitCoordinator, SparkConf conf )
[mangled: org/apache/spark/SparkEnv."<init>":(Ljava/lang/String;Lorg/apache/spark/rpc/RpcEnv;Lorg/apache/spark/serializer/Serializer;Lorg/apache/spark/serializer/Serializer;Lorg/apache/spark/CacheManager;Lorg/apache/spark/MapOutputTracker;Lorg/apache/spark/shuffle/ShuffleManager;Lorg/apache/spark/broadcast/BroadcastManager;Lorg/apache/spark/network/BlockTransferService;Lorg/apache/spark/storage/BlockManager;Lorg/apache/spark/SecurityManager;Lorg/apache/spark/HttpFileServer;Ljava/lang/String;Lorg/apache/spark/metrics/MetricsSystem;Lorg/apache/spark/shuffle/ShuffleMemoryManager;Lorg/apache/spark/unsafe/memory/ExecutorMemoryManager;Lorg/apache/spark/scheduler/OutputCommitCoordinator;Lorg/apache/spark/SparkConf;)V]
spark-core_2.10-1.4.0.jar, TaskContext.class
package org.apache.spark
TaskContext.taskMemoryManager ( ) [abstract] : unsafe.memory.TaskMemoryManager
[mangled: org/apache/spark/TaskContext.taskMemoryManager:()Lorg/apache/spark/unsafe/memory/TaskMemoryManager;]
spark-sql_2.10-1.4.0.jar, DataFrame.class
package org.apache.spark.sql
DataFrame.coalesce ( int numPartitions ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.coalesce:(I)Lorg/apache/spark/sql/DataFrame;]
DataFrame.cube ( Column... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.cube:([Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.cube ( scala.collection.Seq<Column> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.cube:(Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.cube ( String col1, scala.collection.Seq<String> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.cube:(Ljava/lang/String;Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.cube ( String col1, String... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.cube:(Ljava/lang/String;[Ljava/lang/String;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.describe ( scala.collection.Seq<String> cols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.describe:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.describe ( String... cols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.describe:([Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.drop ( String colName ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.drop:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.dropDuplicates ( ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.dropDuplicates:()Lorg/apache/spark/sql/DataFrame;]
DataFrame.dropDuplicates ( scala.collection.Seq<String> colNames ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.dropDuplicates:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.dropDuplicates ( String[ ] colNames ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.dropDuplicates:([Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.join ( DataFrame right, String usingColumn ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.join:(Lorg/apache/spark/sql/DataFrame;Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.na ( ) : DataFrameNaFunctions
[mangled: org/apache/spark/sql/DataFrame.na:()Lorg/apache/spark/sql/DataFrameNaFunctions;]
DataFrame.DataFrame..logicalPlanToDataFrame ( catalyst.plans.logical.LogicalPlan logicalPlan ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.org.apache.spark.sql.DataFrame..logicalPlanToDataFrame:(Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.randomSplit ( double[ ] weights ) : DataFrame[ ]
[mangled: org/apache/spark/sql/DataFrame.randomSplit:([D)[Lorg/apache/spark/sql/DataFrame;]
DataFrame.randomSplit ( double[ ] weights, long seed ) : DataFrame[ ]
[mangled: org/apache/spark/sql/DataFrame.randomSplit:([DJ)[Lorg/apache/spark/sql/DataFrame;]
DataFrame.randomSplit ( scala.collection.immutable.List<Object> weights, long seed ) : DataFrame[ ]
[mangled: org/apache/spark/sql/DataFrame.randomSplit:(Lscala/collection/immutable/List;J)[Lorg/apache/spark/sql/DataFrame;]
DataFrame.rollup ( Column... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.rollup:([Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.rollup ( scala.collection.Seq<Column> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.rollup:(Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.rollup ( String col1, scala.collection.Seq<String> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.rollup:(Ljava/lang/String;Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.rollup ( String col1, String... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.rollup:(Ljava/lang/String;[Ljava/lang/String;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.stat ( ) : DataFrameStatFunctions
[mangled: org/apache/spark/sql/DataFrame.stat:()Lorg/apache/spark/sql/DataFrameStatFunctions;]
DataFrame.write ( ) : DataFrameWriter
[mangled: org/apache/spark/sql/DataFrame.write:()Lorg/apache/spark/sql/DataFrameWriter;]
spark-sql_2.10-1.4.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.cacheManager ( ) : execution.CacheManager
[mangled: org/apache/spark/sql/SQLContext.cacheManager:()Lorg/apache/spark/sql/execution/CacheManager;]
SQLContext.createDataFrame ( org.apache.spark.rdd.RDD<Row> rowRDD, types.StructType schema, boolean needsConversion ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/rdd/RDD;Lorg/apache/spark/sql/types/StructType;Z)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.createSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.currentSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.currentSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.defaultSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.defaultSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.detachSession ( ) : void
[mangled: org/apache/spark/sql/SQLContext.detachSession:()V]
SQLContext.dialectClassName ( ) : String
[mangled: org/apache/spark/sql/SQLContext.dialectClassName:()Ljava/lang/String;]
SQLContext.getOrCreate ( org.apache.spark.SparkContext p1 ) [static] : SQLContext
[mangled: org/apache/spark/sql/SQLContext.getOrCreate:(Lorg/apache/spark/SparkContext;)Lorg/apache/spark/sql/SQLContext;]
SQLContext.getSQLDialect ( ) : catalyst.ParserDialect
[mangled: org/apache/spark/sql/SQLContext.getSQLDialect:()Lorg/apache/spark/sql/catalyst/ParserDialect;]
SQLContext.openSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.openSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.range ( long start, long end ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.range:(JJ)Lorg/apache/spark/sql/DataFrame;]
SQLContext.range ( long start, long end, long step, int numPartitions ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.range:(JJJI)Lorg/apache/spark/sql/DataFrame;]
SQLContext.read ( ) : DataFrameReader
[mangled: org/apache/spark/sql/SQLContext.read:()Lorg/apache/spark/sql/DataFrameReader;]
SQLContext.tlSession ( ) : ThreadLocal<SQLContext.SQLSession>
[mangled: org/apache/spark/sql/SQLContext.tlSession:()Ljava/lang/ThreadLocal;]
to the top
Problems with Data Types, High Severity (4)
spark-core_2.10-1.4.0.jar
package org.apache.spark
[+] TaskContext (1)
| Change | Effect |
---|
1 | Abstract method taskMemoryManager ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (15)
compute ( Partition, TaskContext )2nd parameter 'p2' of this abstract method has type 'TaskContext'.
computeOrReadCheckpoint ( Partition, TaskContext )2nd parameter 'context' of this method has type 'TaskContext'.
iterator ( Partition, TaskContext )2nd parameter 'context' of this method has type 'TaskContext'.
addTaskCompletionListener ( util.TaskCompletionListener )Return value of this abstract method has type 'TaskContext'.
addTaskCompletionListener ( scala.Function1<TaskContext,scala.runtime.BoxedUnit> )Return value of this abstract method has type 'TaskContext'.
attemptNumber ( )This abstract method is from 'TaskContext' abstract class.
get ( )Return value of this method has type 'TaskContext'.
isCompleted ( )This abstract method is from 'TaskContext' abstract class.
isInterrupted ( )This abstract method is from 'TaskContext' abstract class.
isRunningLocally ( )This abstract method is from 'TaskContext' abstract class.
partitionId ( )This abstract method is from 'TaskContext' abstract class.
stageId ( )This abstract method is from 'TaskContext' abstract class.
taskAttemptId ( )This abstract method is from 'TaskContext' abstract class.
TaskContext ( )This constructor is from 'TaskContext' abstract class.
taskMetrics ( )This abstract method is from 'TaskContext' abstract class.
spark-sql_2.10-1.4.0.jar
package org.apache.spark.sql
[+] Column (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.Logging. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (9)
agg ( Column, Column... )1st parameter 'expr' of this method has type 'Column'.
agg ( Column, scala.collection.Seq<Column> )1st parameter 'expr' of this method has type 'Column'.
apply ( java.lang.String )Return value of this method has type 'Column'.
col ( java.lang.String )Return value of this method has type 'Column'.
filter ( Column )1st parameter 'condition' of this method has type 'Column'.
join ( DataFrame, Column )2nd parameter 'joinExprs' of this method has type 'Column'.
join ( DataFrame, Column, java.lang.String )2nd parameter 'joinExprs' of this method has type 'Column'.
where ( Column )1st parameter 'condition' of this method has type 'Column'.
withColumn ( java.lang.String, Column )2nd parameter 'col' of this method has type 'Column'.
[+] SQLConf (1)
| Change | Effect |
---|
1 | Removed super-interface catalyst.CatalystConf. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (1)
conf ( )Return value of this method has type 'SQLConf'.
package org.apache.spark.sql.sources
[+] DDLParser (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.sql.types.DataTypeParser. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (1)
ddlParser ( )Return value of this method has type 'DDLParser'.
to the top
Problems with Data Types, Medium Severity (1)
spark-core_2.10-1.4.0.jar
package org.apache.spark.api.java
[+] JavaRDD<T> (1)
| Change | Effect |
---|
1 | Removed super-class AbstractJavaRDDLike<T,JavaRDD<T>>. | Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. |
[+] affected methods (1)
toJavaRDD ( )Return value of this method has type 'JavaRDD<T>'.
to the top
Java ARchives (3)
spark-core_2.10-1.4.0.jar
spark-mllib_2.10-1.4.0.jar
spark-sql_2.10-1.4.0.jar
to the top
Generated on Mon Jul 6 17:38:51 2015 for sparkling-water-core_2.10-1.4.3 by Java API Compliance Checker 1.4.1
A tool for checking backward compatibility of a Java library API