Binary compatibility report for the elasticsearch-spark_2.10-2.1.0.Beta2 library between 1.1.0 and 1.2.0 versions (relating to the portability of client application elasticsearch-spark_2.10-2.1.0.Beta2.jar)
Test Info
Library Name | elasticsearch-spark_2.10-2.1.0.Beta2 |
Version #1 | 1.1.0 |
Version #2 | 1.2.0 |
Java Version | 1.7.0_75 |
Test Results
Total Java ARchives | 2 |
---|
Total Methods / Classes | 851 / 1440 |
---|
Verdict | Incompatible (6.8%) |
Problem Summary
| Severity | Count |
---|
Added Methods | - | 80 |
---|
Removed Methods | High | 9 |
---|
Problems with Data Types | High | 14 |
---|
Medium | 0 |
Low | 0 |
Problems with Methods | High | 9 |
---|
Medium | 0 |
Low | 0 |
Other Changes in Data Types | - | 3 |
Added Methods (80)
spark-core_2.10-1.2.0.jar, JavaSparkContext.class
package org.apache.spark.api.java
JavaSparkContext.binaryFiles ( String path ) : JavaPairRDD<String,org.apache.spark.input.PortableDataStream>
[mangled: org/apache/spark/api/java/JavaSparkContext.binaryFiles:(Ljava/lang/String;)Lorg/apache/spark/api/java/JavaPairRDD;]
JavaSparkContext.binaryFiles ( String path, int minPartitions ) : JavaPairRDD<String,org.apache.spark.input.PortableDataStream>
[mangled: org/apache/spark/api/java/JavaSparkContext.binaryFiles:(Ljava/lang/String;I)Lorg/apache/spark/api/java/JavaPairRDD;]
JavaSparkContext.binaryRecords ( String path, int recordLength ) : JavaRDD<byte[ ]>
[mangled: org/apache/spark/api/java/JavaSparkContext.binaryRecords:(Ljava/lang/String;I)Lorg/apache/spark/api/java/JavaRDD;]
JavaSparkContext.close ( ) : void
[mangled: org/apache/spark/api/java/JavaSparkContext.close:()V]
JavaSparkContext.statusTracker ( ) : JavaSparkStatusTracker
[mangled: org/apache/spark/api/java/JavaSparkContext.statusTracker:()Lorg/apache/spark/api/java/JavaSparkStatusTracker;]
spark-core_2.10-1.2.0.jar, RDD<T>.class
package org.apache.spark.rdd
RDD<T>.parent ( int j, scala.reflect.ClassTag<U> p2 ) : RDD<U>
[mangled: org/apache/spark/rdd/RDD<T>.parent:(ILscala/reflect/ClassTag;)Lorg/apache/spark/rdd/RDD;]
spark-core_2.10-1.2.0.jar, SparkConf.class
package org.apache.spark
SparkConf.getAppId ( ) : String
[mangled: org/apache/spark/SparkConf.getAppId:()Ljava/lang/String;]
SparkConf.registerKryoClasses ( Class<?>[ ] classes ) : SparkConf
[mangled: org/apache/spark/SparkConf.registerKryoClasses:([Ljava/lang/Class;)Lorg/apache/spark/SparkConf;]
spark-core_2.10-1.2.0.jar, SparkContext.class
package org.apache.spark
SparkContext.applicationId ( ) : String
[mangled: org/apache/spark/SparkContext.applicationId:()Ljava/lang/String;]
SparkContext.binaryFiles ( String path, int minPartitions ) : rdd.RDD<scala.Tuple2<String,input.PortableDataStream>>
[mangled: org/apache/spark/SparkContext.binaryFiles:(Ljava/lang/String;I)Lorg/apache/spark/rdd/RDD;]
SparkContext.binaryRecords ( String path, int recordLength, org.apache.hadoop.conf.Configuration conf ) : rdd.RDD<byte[ ]>
[mangled: org/apache/spark/SparkContext.binaryRecords:(Ljava/lang/String;ILorg/apache/hadoop/conf/Configuration;)Lorg/apache/spark/rdd/RDD;]
SparkContext.eventLogDir ( ) : scala.Option<String>
[mangled: org/apache/spark/SparkContext.eventLogDir:()Lscala/Option;]
SparkContext.executorAllocationManager ( ) : scala.Option<ExecutorAllocationManager>
[mangled: org/apache/spark/SparkContext.executorAllocationManager:()Lscala/Option;]
SparkContext.getExecutorThreadDump ( String executorId ) : scala.Option<util.ThreadStackTrace[ ]>
[mangled: org/apache/spark/SparkContext.getExecutorThreadDump:(Ljava/lang/String;)Lscala/Option;]
SparkContext.isEventLogEnabled ( ) : boolean
[mangled: org/apache/spark/SparkContext.isEventLogEnabled:()Z]
SparkContext.jobProgressListener ( ) : ui.jobs.JobProgressListener
[mangled: org/apache/spark/SparkContext.jobProgressListener:()Lorg/apache/spark/ui/jobs/JobProgressListener;]
SparkContext.killExecutor ( String executorId ) : boolean
[mangled: org/apache/spark/SparkContext.killExecutor:(Ljava/lang/String;)Z]
SparkContext.killExecutors ( scala.collection.Seq<String> executorIds ) : boolean
[mangled: org/apache/spark/SparkContext.killExecutors:(Lscala/collection/Seq;)Z]
SparkContext.metricsSystem ( ) : metrics.MetricsSystem
[mangled: org/apache/spark/SparkContext.metricsSystem:()Lorg/apache/spark/metrics/MetricsSystem;]
SparkContext.SparkContext..creationSite ( ) : util.CallSite
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext..creationSite:()Lorg/apache/spark/util/CallSite;]
SparkContext.progressBar ( ) : scala.Option<ui.ConsoleProgressBar>
[mangled: org/apache/spark/SparkContext.progressBar:()Lscala/Option;]
SparkContext.requestExecutors ( int numAdditionalExecutors ) : boolean
[mangled: org/apache/spark/SparkContext.requestExecutors:(I)Z]
SparkContext.schedulerBackend ( ) : scheduler.SchedulerBackend
[mangled: org/apache/spark/SparkContext.schedulerBackend:()Lorg/apache/spark/scheduler/SchedulerBackend;]
SparkContext.schedulerBackend_.eq ( scheduler.SchedulerBackend p1 ) : void
[mangled: org/apache/spark/SparkContext.schedulerBackend_.eq:(Lorg/apache/spark/scheduler/SchedulerBackend;)V]
SparkContext.setCallSite ( util.CallSite callSite ) : void
[mangled: org/apache/spark/SparkContext.setCallSite:(Lorg/apache/spark/util/CallSite;)V]
SparkContext.statusTracker ( ) : SparkStatusTracker
[mangled: org/apache/spark/SparkContext.statusTracker:()Lorg/apache/spark/SparkStatusTracker;]
SparkContext.ui ( ) : scala.Option<ui.SparkUI>
[mangled: org/apache/spark/SparkContext.ui:()Lscala/Option;]
spark-core_2.10-1.2.0.jar, TaskContext.class
package org.apache.spark
TaskContext.get ( ) [static] : TaskContext
[mangled: org/apache/spark/TaskContext.get:()Lorg/apache/spark/TaskContext;]
TaskContext.isRunningLocally ( ) [abstract] : boolean
[mangled: org/apache/spark/TaskContext.isRunningLocally:()Z]
TaskContext.TaskContext ( )
[mangled: org/apache/spark/TaskContext."<init>":()V]
spark-sql_2.10-1.2.0.jar, DataType.class
package org.apache.spark.sql.api.java
DataType.createStructField ( String name, DataType dataType, boolean nullable, Metadata metadata ) [static] : StructField
[mangled: org/apache/spark/sql/api/java/DataType.createStructField:(Ljava/lang/String;Lorg/apache/spark/sql/api/java/DataType;ZLorg/apache/spark/sql/api/java/Metadata;)Lorg/apache/spark/sql/api/java/StructField;]
spark-sql_2.10-1.2.0.jar, JavaSchemaRDD.class
package org.apache.spark.sql.api.java
JavaSchemaRDD.collectAsync ( ) : org.apache.spark.api.java.JavaFutureAction<java.util.List<Row>>
[mangled: org/apache/spark/sql/api/java/JavaSchemaRDD.collectAsync:()Lorg/apache/spark/api/java/JavaFutureAction;]
JavaSchemaRDD.countAsync ( ) : org.apache.spark.api.java.JavaFutureAction<Long>
[mangled: org/apache/spark/sql/api/java/JavaSchemaRDD.countAsync:()Lorg/apache/spark/api/java/JavaFutureAction;]
JavaSchemaRDD.foreachAsync ( org.apache.spark.api.java.function.VoidFunction<Row> f ) : org.apache.spark.api.java.JavaFutureAction<Void>
[mangled: org/apache/spark/sql/api/java/JavaSchemaRDD.foreachAsync:(Lorg/apache/spark/api/java/function/VoidFunction;)Lorg/apache/spark/api/java/JavaFutureAction;]
JavaSchemaRDD.foreachPartitionAsync ( org.apache.spark.api.java.function.VoidFunction<java.util.Iterator<Row>> f ) : org.apache.spark.api.java.JavaFutureAction<Void>
[mangled: org/apache/spark/sql/api/java/JavaSchemaRDD.foreachPartitionAsync:(Lorg/apache/spark/api/java/function/VoidFunction;)Lorg/apache/spark/api/java/JavaFutureAction;]
JavaSchemaRDD.schemaRDD ( ) : org.apache.spark.sql.SchemaRDD
[mangled: org/apache/spark/sql/api/java/JavaSchemaRDD.schemaRDD:()Lorg/apache/spark/sql/SchemaRDD;]
JavaSchemaRDD.takeAsync ( int num ) : org.apache.spark.api.java.JavaFutureAction<java.util.List<Row>>
[mangled: org/apache/spark/sql/api/java/JavaSchemaRDD.takeAsync:(I)Lorg/apache/spark/api/java/JavaFutureAction;]
JavaSchemaRDD.toJSON ( ) : org.apache.spark.api.java.JavaRDD<String>
[mangled: org/apache/spark/sql/api/java/JavaSchemaRDD.toJSON:()Lorg/apache/spark/api/java/JavaRDD;]
spark-sql_2.10-1.2.0.jar, JavaSQLContext.class
package org.apache.spark.sql.api.java
JavaSQLContext.baseRelationToSchemaRDD ( org.apache.spark.sql.sources.BaseRelation baseRelation ) : JavaSchemaRDD
[mangled: org/apache/spark/sql/api/java/JavaSQLContext.baseRelationToSchemaRDD:(Lorg/apache/spark/sql/sources/BaseRelation;)Lorg/apache/spark/sql/api/java/JavaSchemaRDD;]
spark-sql_2.10-1.2.0.jar, Row.class
package org.apache.spark.sql.api.java
Row.toString ( ) : String
[mangled: org/apache/spark/sql/api/java/Row.toString:()Ljava/lang/String;]
spark-sql_2.10-1.2.0.jar, SchemaRDD.class
package org.apache.spark.sql
SchemaRDD.cache ( ) : org.apache.spark.rdd.RDD
[mangled: org/apache/spark/sql/SchemaRDD.cache:()Lorg/apache/spark/rdd/RDD;]
SchemaRDD.cache ( ) : SchemaRDD
[mangled: org/apache/spark/sql/SchemaRDD.cache:()Lorg/apache/spark/sql/SchemaRDD;]
SchemaRDD.collectToPython ( ) : java.util.List<byte[ ]>
[mangled: org/apache/spark/sql/SchemaRDD.collectToPython:()Ljava/util/List;]
SchemaRDD.persist ( org.apache.spark.storage.StorageLevel newLevel ) : org.apache.spark.rdd.RDD
[mangled: org/apache/spark/sql/SchemaRDD.persist:(Lorg/apache/spark/storage/StorageLevel;)Lorg/apache/spark/rdd/RDD;]
SchemaRDD.persist ( org.apache.spark.storage.StorageLevel newLevel ) : SchemaRDD
[mangled: org/apache/spark/sql/SchemaRDD.persist:(Lorg/apache/spark/storage/StorageLevel;)Lorg/apache/spark/sql/SchemaRDD;]
SchemaRDD.toJSON ( ) : org.apache.spark.rdd.RDD<String>
[mangled: org/apache/spark/sql/SchemaRDD.toJSON:()Lorg/apache/spark/rdd/RDD;]
SchemaRDD.unpersist ( boolean blocking ) : org.apache.spark.rdd.RDD
[mangled: org/apache/spark/sql/SchemaRDD.unpersist:(Z)Lorg/apache/spark/rdd/RDD;]
SchemaRDD.unpersist ( boolean blocking ) : SchemaRDD
[mangled: org/apache/spark/sql/SchemaRDD.unpersist:(Z)Lorg/apache/spark/sql/SchemaRDD;]
spark-sql_2.10-1.2.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.abs ( catalyst.expressions.Expression e ) : catalyst.expressions.Abs
[mangled: org/apache/spark/sql/SQLContext.abs:(Lorg/apache/spark/sql/catalyst/expressions/Expression;)Lorg/apache/spark/sql/catalyst/expressions/Abs;]
SQLContext.baseRelationToSchemaRDD ( sources.BaseRelation baseRelation ) : SchemaRDD
[mangled: org/apache/spark/sql/SQLContext.baseRelationToSchemaRDD:(Lorg/apache/spark/sql/sources/BaseRelation;)Lorg/apache/spark/sql/SchemaRDD;]
SQLContext.bigDecimalToLiteral ( scala.math.BigDecimal d ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.bigDecimalToLiteral:(Lscala/math/BigDecimal;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.cacheQuery ( SchemaRDD query, scala.Option<String> tableName, org.apache.spark.storage.StorageLevel storageLevel ) : void
[mangled: org/apache/spark/sql/SQLContext.cacheQuery:(Lorg/apache/spark/sql/SchemaRDD;Lscala/Option;Lorg/apache/spark/storage/StorageLevel;)V]
SQLContext.clearCache ( ) : void
[mangled: org/apache/spark/sql/SQLContext.clearCache:()V]
SQLContext.columnNameOfCorruptRecord ( ) : String
[mangled: org/apache/spark/sql/SQLContext.columnNameOfCorruptRecord:()Ljava/lang/String;]
SQLContext.dateToLiteral ( java.sql.Date d ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.dateToLiteral:(Ljava/sql/Date;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.ddlParser ( ) : sources.DDLParser
[mangled: org/apache/spark/sql/SQLContext.ddlParser:()Lorg/apache/spark/sql/sources/DDLParser;]
SQLContext.decimalToLiteral ( catalyst.types.decimal.Decimal d ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.decimalToLiteral:(Lorg/apache/spark/sql/catalyst/types/decimal/Decimal;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.dropTempTable ( String tableName ) : void
[mangled: org/apache/spark/sql/SQLContext.dropTempTable:(Ljava/lang/String;)V]
SQLContext.externalSortEnabled ( ) : boolean
[mangled: org/apache/spark/sql/SQLContext.externalSortEnabled:()Z]
SQLContext.extraStrategies ( ) : scala.collection.Seq<catalyst.planning.GenericStrategy<execution.SparkPlan>>
[mangled: org/apache/spark/sql/SQLContext.extraStrategies:()Lscala/collection/Seq;]
SQLContext.extraStrategies_.eq ( scala.collection.Seq<catalyst.planning.GenericStrategy<execution.SparkPlan>> p1 ) : void
[mangled: org/apache/spark/sql/SQLContext.extraStrategies_.eq:(Lscala/collection/Seq;)V]
SQLContext.inMemoryPartitionPruning ( ) : boolean
[mangled: org/apache/spark/sql/SQLContext.inMemoryPartitionPruning:()Z]
SQLContext.invalidateCache ( catalyst.plans.logical.LogicalPlan plan ) : void
[mangled: org/apache/spark/sql/SQLContext.invalidateCache:(Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)V]
SQLContext.last ( catalyst.expressions.Expression e ) : catalyst.expressions.Last
[mangled: org/apache/spark/sql/SQLContext.last:(Lorg/apache/spark/sql/catalyst/expressions/Expression;)Lorg/apache/spark/sql/catalyst/expressions/Last;]
SQLContext.lookupCachedData ( catalyst.plans.logical.LogicalPlan plan ) : scala.Option<CachedData>
[mangled: org/apache/spark/sql/SQLContext.lookupCachedData:(Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)Lscala/Option;]
SQLContext.lookupCachedData ( SchemaRDD query ) : scala.Option<CachedData>
[mangled: org/apache/spark/sql/SQLContext.lookupCachedData:(Lorg/apache/spark/sql/SchemaRDD;)Lscala/Option;]
SQLContext.optimizer ( ) : catalyst.optimizer.Optimizer
[mangled: org/apache/spark/sql/SQLContext.optimizer:()Lorg/apache/spark/sql/catalyst/optimizer/Optimizer;]
SQLContext.CacheManager..cachedData ( ) : scala.collection.mutable.ArrayBuffer<CachedData>
[mangled: org/apache/spark/sql/SQLContext.org.apache.spark.sql.CacheManager..cachedData:()Lscala/collection/mutable/ArrayBuffer;]
SQLContext.CacheManager..cacheLock ( ) : java.util.concurrent.locks.ReentrantReadWriteLock
[mangled: org/apache/spark/sql/SQLContext.org.apache.spark.sql.CacheManager..cacheLock:()Ljava/util/concurrent/locks/ReentrantReadWriteLock;]
SQLContext.CacheManager._setter_.CacheManager..cachedData_.eq ( scala.collection.mutable.ArrayBuffer p1 ) : void
[mangled: org/apache/spark/sql/SQLContext.org.apache.spark.sql.CacheManager._setter_.org.apache.spark.sql.CacheManager..cachedData_.eq:(Lscala/collection/mutable/ArrayBuffer;)V]
SQLContext.CacheManager._setter_.CacheManager..cacheLock_.eq ( java.util.concurrent.locks.ReentrantReadWriteLock p1 ) : void
[mangled: org/apache/spark/sql/SQLContext.org.apache.spark.sql.CacheManager._setter_.org.apache.spark.sql.CacheManager..cacheLock_.eq:(Ljava/util/concurrent/locks/ReentrantReadWriteLock;)V]
SQLContext.parquetFilterPushDown ( ) : boolean
[mangled: org/apache/spark/sql/SQLContext.parquetFilterPushDown:()Z]
SQLContext.registerPython ( String name, byte[ ] command, java.util.Map<String,String> envVars, java.util.List<String> pythonIncludes, String pythonExec, java.util.List<org.apache.spark.broadcast.Broadcast<org.apache.spark.api.python.PythonBroadcast>> broadcastVars, org.apache.spark.Accumulator<java.util.List<byte[ ]>> accumulator, String stringDataType ) : void
[mangled: org/apache/spark/sql/SQLContext.registerPython:(Ljava/lang/String;[BLjava/util/Map;Ljava/util/List;Ljava/lang/String;Ljava/util/List;Lorg/apache/spark/Accumulator;Ljava/lang/String;)V]
SQLContext.sqlParser ( ) : catalyst.SparkSQLParser
[mangled: org/apache/spark/sql/SQLContext.sqlParser:()Lorg/apache/spark/sql/catalyst/SparkSQLParser;]
SQLContext.sqrt ( catalyst.expressions.Expression e ) : catalyst.expressions.Sqrt
[mangled: org/apache/spark/sql/SQLContext.sqrt:(Lorg/apache/spark/sql/catalyst/expressions/Expression;)Lorg/apache/spark/sql/catalyst/expressions/Sqrt;]
SQLContext.tryUncacheQuery ( SchemaRDD query, boolean blocking ) : boolean
[mangled: org/apache/spark/sql/SQLContext.tryUncacheQuery:(Lorg/apache/spark/sql/SchemaRDD;Z)Z]
SQLContext.uncacheQuery ( SchemaRDD query, boolean blocking ) : void
[mangled: org/apache/spark/sql/SQLContext.uncacheQuery:(Lorg/apache/spark/sql/SchemaRDD;Z)V]
SQLContext.useCachedData ( catalyst.plans.logical.LogicalPlan plan ) : catalyst.plans.logical.LogicalPlan
[mangled: org/apache/spark/sql/SQLContext.useCachedData:(Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;]
spark-sql_2.10-1.2.0.jar, StructField.class
package org.apache.spark.sql.api.java
StructField.getMetadata ( ) : Metadata
[mangled: org/apache/spark/sql/api/java/StructField.getMetadata:()Lorg/apache/spark/sql/api/java/Metadata;]
StructField.StructField ( String name, DataType dataType, boolean nullable, Metadata metadata )
[mangled: org/apache/spark/sql/api/java/StructField."<init>":(Ljava/lang/String;Lorg/apache/spark/sql/api/java/DataType;ZLorg/apache/spark/sql/api/java/Metadata;)V]
to the top
Removed Methods (9)
spark-core_2.10-1.1.0.jar, SparkContext.class
package org.apache.spark
SparkContext.ui ( ) : ui.SparkUI
[mangled: org/apache/spark/SparkContext.ui:()Lorg/apache/spark/ui/SparkUI;]
spark-core_2.10-1.1.0.jar, TaskContext.class
package org.apache.spark
TaskContext.markInterrupted ( ) : void
[mangled: org/apache/spark/TaskContext.markInterrupted:()V]
TaskContext.markTaskCompleted ( ) : void
[mangled: org/apache/spark/TaskContext.markTaskCompleted:()V]
TaskContext.TaskContext ( int stageId, int partitionId, long attemptId, boolean runningLocally, executor.TaskMetrics taskMetrics )
[mangled: org/apache/spark/TaskContext."<init>":(IIJZLorg/apache/spark/executor/TaskMetrics;)V]
spark-sql_2.10-1.1.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.decimalToLiteral ( scala.math.BigDecimal d ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.decimalToLiteral:(Lscala/math/BigDecimal;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.optimizer ( ) : catalyst.optimizer.Optimizer.
[mangled: org/apache/spark/sql/SQLContext.optimizer:()Lorg/apache/spark/sql/catalyst/optimizer/Optimizer$;]
SQLContext.parser ( ) : catalyst.SqlParser
[mangled: org/apache/spark/sql/SQLContext.parser:()Lorg/apache/spark/sql/catalyst/SqlParser;]
SQLContext.registerPython ( String name, byte[ ] command, java.util.Map<String,String> envVars, java.util.List<String> pythonIncludes, String pythonExec, org.apache.spark.Accumulator<java.util.List<byte[ ]>> accumulator, String stringDataType ) : void
[mangled: org/apache/spark/sql/SQLContext.registerPython:(Ljava/lang/String;[BLjava/util/Map;Ljava/util/List;Ljava/lang/String;Lorg/apache/spark/Accumulator;Ljava/lang/String;)V]
spark-sql_2.10-1.1.0.jar, StructField.class
package org.apache.spark.sql.api.java
StructField.StructField ( String name, DataType dataType, boolean nullable )
[mangled: org/apache/spark/sql/api/java/StructField."<init>":(Ljava/lang/String;Lorg/apache/spark/sql/api/java/DataType;Z)V]
to the top
Problems with Data Types, High Severity (14)
spark-core_2.10-1.1.0.jar
package org.apache.spark
[+] TaskContext (12)
| Change | Effect |
---|
1 | This class became abstract. | A client program may be interrupted by InstantiationError exception. |
2 | Method addTaskCompletionListener ( util.TaskCompletionListener ) became abstract. | A client program may be interrupted by InstantiationError exception. |
3 | Method addTaskCompletionListener ( scala.Function1<TaskContext,scala.runtime.BoxedUnit> ) became abstract. | A client program may be interrupted by InstantiationError exception. |
4 | Method attemptId ( ) became abstract. | A client program may be interrupted by InstantiationError exception. |
5 | Method isCompleted ( ) became abstract. | A client program may be interrupted by InstantiationError exception. |
6 | Method isInterrupted ( ) became abstract. | A client program may be interrupted by InstantiationError exception. |
7 | Method partitionId ( ) became abstract. | A client program may be interrupted by InstantiationError exception. |
8 | Method runningLocally ( ) became abstract. | A client program may be interrupted by InstantiationError exception. |
9 | Method stageId ( ) became abstract. | A client program may be interrupted by InstantiationError exception. |
10 | Method taskMetrics ( ) became abstract. | A client program may be interrupted by InstantiationError exception. |
11 | Removed super-interface scala.Serializable. | A client program may be interrupted by NoSuchMethodError exception. |
12 | Abstract method isRunningLocally ( ) has been added to this class. | This class became abstract and a client program may be interrupted by InstantiationError exception. |
[+] affected methods (14)
compute ( Partition, TaskContext )2nd parameter 'p2' of this abstract method has type 'TaskContext'.
computeOrReadCheckpoint ( Partition, TaskContext )2nd parameter 'context' of this method has type 'TaskContext'.
iterator ( Partition, TaskContext )2nd parameter 'context' of this method has type 'TaskContext'.
iterator ( Partition, TaskContext )2nd parameter 'taskContext' of this method has type 'TaskContext'.
compute ( Partition, TaskContext )2nd parameter 'context' of this method has type 'TaskContext'.
addTaskCompletionListener ( util.TaskCompletionListener )Return value of this method has type 'TaskContext'.
addTaskCompletionListener ( scala.Function1<TaskContext,scala.runtime.BoxedUnit> )Return value of this method has type 'TaskContext'.
attemptId ( )This method is from 'TaskContext' class.
isCompleted ( )This method is from 'TaskContext' class.
isInterrupted ( )This method is from 'TaskContext' class.
partitionId ( )This method is from 'TaskContext' class.
runningLocally ( )This method is from 'TaskContext' class.
stageId ( )This method is from 'TaskContext' class.
taskMetrics ( )This method is from 'TaskContext' class.
package org.apache.spark.rdd
[+] PairRDDFunctions<K,V> (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.hadoop.mapreduce.SparkHadoopMapReduceUtil. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (1)
rddToPairRDDFunctions ( RDD<scala.Tuple2<K,V>>, scala.reflect.ClassTag<K>, scala.reflect.ClassTag<V>, scala.math.Ordering<K> )Return value of this method has type 'PairRDDFunctions<K,V>'.
spark-sql_2.10-1.1.0.jar
package org.apache.spark.sql.api.java
[+] DataType (1)
| Change | Effect |
---|
1 | Field DecimalType of type DecimalType has been removed from this class. | A client program may be interrupted by NoSuchFieldError exception. |
[+] affected methods (34)
getElementType ( )Return value of this method has type 'DataType'.
createArrayType ( DataType )This method is from 'DataType' abstract class.
createArrayType ( DataType, boolean )This method is from 'DataType' abstract class.
createMapType ( DataType, DataType )This method is from 'DataType' abstract class.
createMapType ( DataType, DataType, boolean )This method is from 'DataType' abstract class.
createStructField ( java.lang.String, DataType, boolean )This method is from 'DataType' abstract class.
createStructType ( java.util.List<StructField> )This method is from 'DataType' abstract class.
createStructType ( StructField[ ] )This method is from 'DataType' abstract class.
DataType ( )This constructor is from 'DataType' abstract class.
registerFunction ( java.lang.String, UDF10<?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF11<?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF12<?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF13<?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF14<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF15<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF16<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF17<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF18<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF19<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF1<?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF20<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF21<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF22<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF2<?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF3<?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF4<?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF5<?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF6<?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF7<?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF8<?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF9<?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
getKeyType ( )Return value of this method has type 'DataType'.
getValueType ( )Return value of this method has type 'DataType'.
getDataType ( )Return value of this method has type 'DataType'.
to the top
Problems with Methods, High Severity (9)
spark-core_2.10-1.1.0.jar, TaskContext
package org.apache.spark
[+] TaskContext.addTaskCompletionListener ( util.TaskCompletionListener listener ) : TaskContext (1)
[mangled: org/apache/spark/TaskContext.addTaskCompletionListener:(Lorg/apache/spark/util/TaskCompletionListener;)Lorg/apache/spark/TaskContext;]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
[+] TaskContext.addTaskCompletionListener ( scala.Function1<TaskContext,scala.runtime.BoxedUnit> f ) : TaskContext (1)
[mangled: org/apache/spark/TaskContext.addTaskCompletionListener:(Lscala/Function1;)Lorg/apache/spark/TaskContext;]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
[+] TaskContext.attemptId ( ) : long (1)
[mangled: org/apache/spark/TaskContext.attemptId:()J]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
[+] TaskContext.isCompleted ( ) : boolean (1)
[mangled: org/apache/spark/TaskContext.isCompleted:()Z]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
[+] TaskContext.isInterrupted ( ) : boolean (1)
[mangled: org/apache/spark/TaskContext.isInterrupted:()Z]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
[+] TaskContext.partitionId ( ) : int (1)
[mangled: org/apache/spark/TaskContext.partitionId:()I]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
[+] TaskContext.runningLocally ( ) : boolean (1)
[mangled: org/apache/spark/TaskContext.runningLocally:()Z]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
[+] TaskContext.stageId ( ) : int (1)
[mangled: org/apache/spark/TaskContext.stageId:()I]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
[+] TaskContext.taskMetrics ( ) : executor.TaskMetrics (1)
[mangled: org/apache/spark/TaskContext.taskMetrics:()Lorg/apache/spark/executor/TaskMetrics;]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
to the top
Other Changes in Data Types (3)
spark-core_2.10-1.1.0.jar
package org.apache.spark.broadcast
[+] Broadcast<T> (1)
| Change | Effect |
---|
1 | Added super-interface org.apache.spark.Logging. | No effect. |
[+] affected methods (2)
broadcast ( T )Return value of this method has type 'Broadcast<T>'.
broadcast ( T, scala.reflect.ClassTag<T> )Return value of this method has type 'Broadcast<T>'.
spark-sql_2.10-1.1.0.jar
package org.apache.spark.sql.api.java
[+] DataType (2)
| Change | Effect |
---|
1 | Field DateType has been added to this class. | No effect. NOTE: A static field from a super-interface of a client class may hide an added field (with the same name) inherited from the super-class of a client class and cause IncompatibleClassChangeError exception. |
2 | Field NullType has been added to this class. | No effect. NOTE: A static field from a super-interface of a client class may hide an added field (with the same name) inherited from the super-class of a client class and cause IncompatibleClassChangeError exception. |
[+] affected methods (34)
getElementType ( )Return value of this method has type 'DataType'.
createArrayType ( DataType )1st parameter 'elementType' of this method has type 'DataType'.
createArrayType ( DataType, boolean )1st parameter 'elementType' of this method has type 'DataType'.
createMapType ( DataType, DataType )2nd parameter 'valueType' of this method has type 'DataType'.
createMapType ( DataType, DataType, boolean )2nd parameter 'valueType' of this method has type 'DataType'.
createStructField ( java.lang.String, DataType, boolean )2nd parameter 'dataType' of this method has type 'DataType'.
createStructType ( java.util.List<StructField> )This method is from 'DataType' abstract class.
createStructType ( StructField[ ] )This method is from 'DataType' abstract class.
DataType ( )This constructor is from 'DataType' abstract class.
registerFunction ( java.lang.String, UDF10<?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF11<?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF12<?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF13<?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF14<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF15<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF16<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF17<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF18<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF19<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF1<?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF20<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF21<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF22<?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF2<?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF3<?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF4<?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF5<?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF6<?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF7<?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF8<?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
registerFunction ( java.lang.String, UDF9<?,?,?,?,?,?,?,?,?,?>, DataType )3rd parameter 'dataType' of this method has type 'DataType'.
getKeyType ( )Return value of this method has type 'DataType'.
getValueType ( )Return value of this method has type 'DataType'.
getDataType ( )Return value of this method has type 'DataType'.
to the top
Java ARchives (2)
spark-core_2.10-1.1.0.jar
spark-sql_2.10-1.1.0.jar
to the top
Generated on Mon Jun 29 20:33:30 2015 for elasticsearch-spark_2.10-2.1.0.Beta2 by Java API Compliance Checker 1.4.1
A tool for checking backward compatibility of a Java library API