Binary compatibility report for the spark-ryft-connector-core_2.10-0.4.0 library between 1.4.0 and 1.3.0 versions (relating to the portability of client application spark-ryft-connector-core_2.10-0.4.0.jar)
Test Info
Library Name | spark-ryft-connector-core_2.10-0.4.0 |
Version #1 | 1.4.0 |
Version #2 | 1.3.0 |
Java Version | 1.7.0_85 |
Test Results
Total Java ARchives | 4 |
---|
Total Methods / Classes | 542 / 2724 |
---|
Verdict | Incompatible (16.1%) |
Problem Summary
| Severity | Count |
---|
Added Methods | - | 4 |
---|
Removed Methods | High | 70 |
---|
Problems with Data Types | High | 7 |
---|
Medium | 1 |
Low | 0 |
Problems with Methods | High | 0 |
---|
Medium | 0 |
Low | 0 |
Added Methods (4)
spark-core_2.10-1.3.0.jar, SparkConf.class
package org.apache.spark
SparkConf.translateConfKey ( String p1, boolean p2 ) [static] : String
[mangled: org/apache/spark/SparkConf.translateConfKey:(Ljava/lang/String;Z)Ljava/lang/String;]
spark-sql_2.10-1.3.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.cacheManager ( ) : CacheManager
[mangled: org/apache/spark/sql/SQLContext.cacheManager:()Lorg/apache/spark/sql/CacheManager;]
SQLContext.checkAnalysis ( ) : catalyst.analysis.CheckAnalysis
[mangled: org/apache/spark/sql/SQLContext.checkAnalysis:()Lorg/apache/spark/sql/catalyst/analysis/CheckAnalysis;]
SQLContext.createDataFrame ( org.apache.spark.api.java.JavaRDD<Row> rowRDD, java.util.List<String> columns ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/api/java/JavaRDD;Ljava/util/List;)Lorg/apache/spark/sql/DataFrame;]
to the top
Removed Methods (70)
spark-core_2.10-1.4.0.jar, RDD<T>.class
package org.apache.spark.rdd
RDD<T>.RDD..doCheckpointCalled ( ) : boolean
[mangled: org/apache/spark/rdd/RDD<T>.org.apache.spark.rdd.RDD..doCheckpointCalled:()Z]
RDD<T>.RDD..doCheckpointCalled_.eq ( boolean p1 ) : void
[mangled: org/apache/spark/rdd/RDD<T>.org.apache.spark.rdd.RDD..doCheckpointCalled_.eq:(Z)V]
RDD<T>.RDD..sc ( ) : org.apache.spark.SparkContext
[mangled: org/apache/spark/rdd/RDD<T>.org.apache.spark.rdd.RDD..sc:()Lorg/apache/spark/SparkContext;]
RDD<T>.randomSampleWithRange ( double lb, double ub, long seed ) : RDD<T>
[mangled: org/apache/spark/rdd/RDD<T>.randomSampleWithRange:(DDJ)Lorg/apache/spark/rdd/RDD;]
RDD<T>.scope ( ) : scala.Option<RDDOperationScope>
[mangled: org/apache/spark/rdd/RDD<T>.scope:()Lscala/Option;]
RDD<T>.withScope ( scala.Function0<U> body ) : U
[mangled: org/apache/spark/rdd/RDD<T>.withScope:(Lscala/Function0;)Ljava/lang/Object;]
spark-core_2.10-1.4.0.jar, SparkConf.class
package org.apache.spark
SparkConf.getDeprecatedConfig ( String p1, SparkConf p2 ) [static] : scala.Option<String>
[mangled: org/apache/spark/SparkConf.getDeprecatedConfig:(Ljava/lang/String;Lorg/apache/spark/SparkConf;)Lscala/Option;]
SparkConf.getSizeAsBytes ( String key ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsBytes:(Ljava/lang/String;)J]
SparkConf.getSizeAsBytes ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsBytes:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getSizeAsGb ( String key ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsGb:(Ljava/lang/String;)J]
SparkConf.getSizeAsGb ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsGb:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getSizeAsKb ( String key ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsKb:(Ljava/lang/String;)J]
SparkConf.getSizeAsKb ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsKb:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getSizeAsMb ( String key ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsMb:(Ljava/lang/String;)J]
SparkConf.getSizeAsMb ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getSizeAsMb:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getTimeAsMs ( String key ) : long
[mangled: org/apache/spark/SparkConf.getTimeAsMs:(Ljava/lang/String;)J]
SparkConf.getTimeAsMs ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getTimeAsMs:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.getTimeAsSeconds ( String key ) : long
[mangled: org/apache/spark/SparkConf.getTimeAsSeconds:(Ljava/lang/String;)J]
SparkConf.getTimeAsSeconds ( String key, String defaultValue ) : long
[mangled: org/apache/spark/SparkConf.getTimeAsSeconds:(Ljava/lang/String;Ljava/lang/String;)J]
SparkConf.logDeprecationWarning ( String p1 ) [static] : void
[mangled: org/apache/spark/SparkConf.logDeprecationWarning:(Ljava/lang/String;)V]
spark-core_2.10-1.4.0.jar, SparkContext.class
package org.apache.spark
SparkContext.applicationAttemptId ( ) : scala.Option<String>
[mangled: org/apache/spark/SparkContext.applicationAttemptId:()Lscala/Option;]
SparkContext.externalBlockStoreFolderName ( ) : String
[mangled: org/apache/spark/SparkContext.externalBlockStoreFolderName:()Ljava/lang/String;]
SparkContext.getOrCreate ( ) [static] : SparkContext
[mangled: org/apache/spark/SparkContext.getOrCreate:()Lorg/apache/spark/SparkContext;]
SparkContext.getOrCreate ( SparkConf p1 ) [static] : SparkContext
[mangled: org/apache/spark/SparkContext.getOrCreate:(Lorg/apache/spark/SparkConf;)Lorg/apache/spark/SparkContext;]
SparkContext.SparkContext.._conf ( ) : SparkConf
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext.._conf:()Lorg/apache/spark/SparkConf;]
SparkContext.SparkContext.._env ( ) : SparkEnv
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext.._env:()Lorg/apache/spark/SparkEnv;]
SparkContext.SparkContext..assertNotStopped ( ) : void
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext..assertNotStopped:()V]
SparkContext.range ( long start, long end, long step, int numSlices ) : rdd.RDD<Object>
[mangled: org/apache/spark/SparkContext.range:(JJJI)Lorg/apache/spark/rdd/RDD;]
SparkContext.setLogLevel ( String logLevel ) : void
[mangled: org/apache/spark/SparkContext.setLogLevel:(Ljava/lang/String;)V]
SparkContext.supportDynamicAllocation ( ) : boolean
[mangled: org/apache/spark/SparkContext.supportDynamicAllocation:()Z]
SparkContext.withScope ( scala.Function0<U> body ) : U
[mangled: org/apache/spark/SparkContext.withScope:(Lscala/Function0;)Ljava/lang/Object;]
spark-core_2.10-1.4.0.jar, TaskContext.class
package org.apache.spark
TaskContext.taskMemoryManager ( ) [abstract] : unsafe.memory.TaskMemoryManager
[mangled: org/apache/spark/TaskContext.taskMemoryManager:()Lorg/apache/spark/unsafe/memory/TaskMemoryManager;]
spark-sql_2.10-1.4.0.jar, BaseRelation.class
package org.apache.spark.sql.sources
BaseRelation.needConversion ( ) : boolean
[mangled: org/apache/spark/sql/sources/BaseRelation.needConversion:()Z]
spark-sql_2.10-1.4.0.jar, DataFrameReader.class
package org.apache.spark.sql
DataFrameReader.DataFrameReader ( SQLContext sqlContext )
[mangled: org/apache/spark/sql/DataFrameReader."<init>":(Lorg/apache/spark/sql/SQLContext;)V]
DataFrameReader.format ( String source ) : DataFrameReader
[mangled: org/apache/spark/sql/DataFrameReader.format:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrameReader;]
DataFrameReader.jdbc ( String url, String table, java.util.Properties properties ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.jdbc:(Ljava/lang/String;Ljava/lang/String;Ljava/util/Properties;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.jdbc ( String url, String table, String columnName, long lowerBound, long upperBound, int numPartitions, java.util.Properties connectionProperties ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.jdbc:(Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;JJILjava/util/Properties;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.jdbc ( String url, String table, String[ ] predicates, java.util.Properties connectionProperties ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.jdbc:(Ljava/lang/String;Ljava/lang/String;[Ljava/lang/String;Ljava/util/Properties;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.json ( org.apache.spark.api.java.JavaRDD<String> jsonRDD ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.json:(Lorg/apache/spark/api/java/JavaRDD;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.json ( org.apache.spark.rdd.RDD<String> jsonRDD ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.json:(Lorg/apache/spark/rdd/RDD;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.json ( String path ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.json:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.load ( ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.load:()Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.load ( String path ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.load:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.option ( String key, String value ) : DataFrameReader
[mangled: org/apache/spark/sql/DataFrameReader.option:(Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/sql/DataFrameReader;]
DataFrameReader.options ( java.util.Map<String,String> options ) : DataFrameReader
[mangled: org/apache/spark/sql/DataFrameReader.options:(Ljava/util/Map;)Lorg/apache/spark/sql/DataFrameReader;]
DataFrameReader.options ( scala.collection.Map<String,String> options ) : DataFrameReader
[mangled: org/apache/spark/sql/DataFrameReader.options:(Lscala/collection/Map;)Lorg/apache/spark/sql/DataFrameReader;]
DataFrameReader.parquet ( scala.collection.Seq<String> paths ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.parquet:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.parquet ( String... paths ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.parquet:([Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrameReader.schema ( types.StructType schema ) : DataFrameReader
[mangled: org/apache/spark/sql/DataFrameReader.schema:(Lorg/apache/spark/sql/types/StructType;)Lorg/apache/spark/sql/DataFrameReader;]
DataFrameReader.table ( String tableName ) : DataFrame
[mangled: org/apache/spark/sql/DataFrameReader.table:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
spark-sql_2.10-1.4.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.cacheManager ( ) : execution.CacheManager
[mangled: org/apache/spark/sql/SQLContext.cacheManager:()Lorg/apache/spark/sql/execution/CacheManager;]
SQLContext.createDataFrame ( org.apache.spark.rdd.RDD<Row> rowRDD, types.StructType schema, boolean needsConversion ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/rdd/RDD;Lorg/apache/spark/sql/types/StructType;Z)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.createSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.currentSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.currentSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.defaultSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.defaultSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.detachSession ( ) : void
[mangled: org/apache/spark/sql/SQLContext.detachSession:()V]
SQLContext.dialectClassName ( ) : String
[mangled: org/apache/spark/sql/SQLContext.dialectClassName:()Ljava/lang/String;]
SQLContext.getOrCreate ( org.apache.spark.SparkContext p1 ) [static] : SQLContext
[mangled: org/apache/spark/sql/SQLContext.getOrCreate:(Lorg/apache/spark/SparkContext;)Lorg/apache/spark/sql/SQLContext;]
SQLContext.getSQLDialect ( ) : catalyst.ParserDialect
[mangled: org/apache/spark/sql/SQLContext.getSQLDialect:()Lorg/apache/spark/sql/catalyst/ParserDialect;]
SQLContext.openSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.openSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.range ( long start, long end ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.range:(JJ)Lorg/apache/spark/sql/DataFrame;]
SQLContext.range ( long start, long end, long step, int numPartitions ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.range:(JJJI)Lorg/apache/spark/sql/DataFrame;]
SQLContext.read ( ) : DataFrameReader
[mangled: org/apache/spark/sql/SQLContext.read:()Lorg/apache/spark/sql/DataFrameReader;]
SQLContext.tlSession ( ) : ThreadLocal<SQLContext.SQLSession>
[mangled: org/apache/spark/sql/SQLContext.tlSession:()Ljava/lang/ThreadLocal;]
spark-sql_2.10-1.4.0.jar, StringContains.class
package org.apache.spark.sql.sources
StringContains.attribute ( ) : String
[mangled: org/apache/spark/sql/sources/StringContains.attribute:()Ljava/lang/String;]
StringContains.value ( ) : String
[mangled: org/apache/spark/sql/sources/StringContains.value:()Ljava/lang/String;]
spark-sql_2.10-1.4.0.jar, StringEndsWith.class
package org.apache.spark.sql.sources
StringEndsWith.attribute ( ) : String
[mangled: org/apache/spark/sql/sources/StringEndsWith.attribute:()Ljava/lang/String;]
StringEndsWith.value ( ) : String
[mangled: org/apache/spark/sql/sources/StringEndsWith.value:()Ljava/lang/String;]
spark-sql_2.10-1.4.0.jar, StringStartsWith.class
package org.apache.spark.sql.sources
StringStartsWith.attribute ( ) : String
[mangled: org/apache/spark/sql/sources/StringStartsWith.attribute:()Ljava/lang/String;]
StringStartsWith.value ( ) : String
[mangled: org/apache/spark/sql/sources/StringStartsWith.value:()Ljava/lang/String;]
to the top
Problems with Data Types, High Severity (7)
spark-core_2.10-1.4.0.jar
package org.apache.spark
[+] TaskContext (1)
| Change | Effect |
---|
1 | Abstract method taskMemoryManager ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (15)
compute ( Partition, TaskContext )2nd parameter 'p2' of this abstract method has type 'TaskContext'.
computeOrReadCheckpoint ( Partition, TaskContext )2nd parameter 'context' of this method has type 'TaskContext'.
iterator ( Partition, TaskContext )2nd parameter 'context' of this method has type 'TaskContext'.
addTaskCompletionListener ( util.TaskCompletionListener )This abstract method is from 'TaskContext' abstract class.
addTaskCompletionListener ( scala.Function1<TaskContext,scala.runtime.BoxedUnit> )This abstract method is from 'TaskContext' abstract class.
attemptNumber ( )This abstract method is from 'TaskContext' abstract class.
get ( )This method is from 'TaskContext' abstract class.
isCompleted ( )This abstract method is from 'TaskContext' abstract class.
isInterrupted ( )This abstract method is from 'TaskContext' abstract class.
isRunningLocally ( )This abstract method is from 'TaskContext' abstract class.
partitionId ( )This abstract method is from 'TaskContext' abstract class.
stageId ( )This abstract method is from 'TaskContext' abstract class.
taskAttemptId ( )This abstract method is from 'TaskContext' abstract class.
TaskContext ( )This constructor is from 'TaskContext' abstract class.
taskMetrics ( )This abstract method is from 'TaskContext' abstract class.
spark-sql_2.10-1.4.0.jar
package org.apache.spark.sql
[+] DataFrameReader (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (17)
DataFrameReader ( SQLContext )This constructor is from 'DataFrameReader' class.
format ( java.lang.String )This method is from 'DataFrameReader' class.
jdbc ( java.lang.String, java.lang.String, java.lang.String, long, long, int, java.util.Properties )This method is from 'DataFrameReader' class.
jdbc ( java.lang.String, java.lang.String, java.lang.String[ ], java.util.Properties )This method is from 'DataFrameReader' class.
jdbc ( java.lang.String, java.lang.String, java.util.Properties )This method is from 'DataFrameReader' class.
json ( java.lang.String )This method is from 'DataFrameReader' class.
json ( org.apache.spark.api.java.JavaRDD<java.lang.String> )This method is from 'DataFrameReader' class.
json ( org.apache.spark.rdd.RDD<java.lang.String> )This method is from 'DataFrameReader' class.
load ( )This method is from 'DataFrameReader' class.
load ( java.lang.String )This method is from 'DataFrameReader' class.
option ( java.lang.String, java.lang.String )This method is from 'DataFrameReader' class.
options ( java.util.Map<java.lang.String,java.lang.String> )This method is from 'DataFrameReader' class.
options ( scala.collection.Map<java.lang.String,java.lang.String> )This method is from 'DataFrameReader' class.
parquet ( java.lang.String... )This method is from 'DataFrameReader' class.
parquet ( scala.collection.Seq<java.lang.String> )This method is from 'DataFrameReader' class.
schema ( types.StructType )This method is from 'DataFrameReader' class.
table ( java.lang.String )This method is from 'DataFrameReader' class.
[+] SQLConf (1)
| Change | Effect |
---|
1 | Removed super-interface catalyst.CatalystConf. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (1)
conf ( )Return value of this method has type 'SQLConf'.
package org.apache.spark.sql.sources
[+] DDLParser (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.sql.types.DataTypeParser. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (1)
ddlParser ( )Return value of this method has type 'DDLParser'.
[+] StringContains (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (2)
attribute ( )This method is from 'StringContains' class.
value ( )This method is from 'StringContains' class.
[+] StringEndsWith (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (2)
attribute ( )This method is from 'StringEndsWith' class.
value ( )This method is from 'StringEndsWith' class.
[+] StringStartsWith (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (2)
attribute ( )This method is from 'StringStartsWith' class.
value ( )This method is from 'StringStartsWith' class.
to the top
Problems with Data Types, Medium Severity (1)
spark-core_2.10-1.4.0.jar
package org.apache.spark.api.java
[+] JavaRDD<T> (1)
| Change | Effect |
---|
1 | Removed super-class AbstractJavaRDDLike<T,JavaRDD<T>>. | Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. |
[+] affected methods (1)
toJavaRDD ( )Return value of this method has type 'JavaRDD<T>'.
to the top
Java ARchives (4)
spark-core_2.10-1.4.0.jar
spark-sql_2.10-1.4.0.jar
spark-streaming-twitter_2.10-1.4.0.jar
spark-streaming_2.10-1.4.0.jar
to the top
Generated on Wed Jan 6 22:15:09 2016 for spark-ryft-connector-core_2.10-0.4.0 by Java API Compliance Checker 1.4.1
A tool for checking backward compatibility of a Java library API