Binary compatibility report for the Spark-SQL-on-HBase-1.0.0 library between 1.4.0 and 1.0.0 versions (relating to the portability of client application Spark-SQL-on-HBase-1.0.0.jar)
Test Info
Library Name | Spark-SQL-on-HBase-1.0.0 |
Version #1 | 1.4.0 |
Version #2 | 1.0.0 |
Java Version | 1.7.0_75 |
Test Results
Total Java ARchives | 3 |
---|
Total Methods / Classes | 1150 / 3171 |
---|
Verdict | Incompatible (82.7%) |
Problem Summary
| Severity | Count |
---|
Added Methods | - | 140 |
---|
Removed Methods | High | 630 |
---|
Problems with Data Types | High | 64 |
---|
Medium | 14 |
Low | 6 |
Problems with Methods | High | 1 |
---|
Medium | 0 |
Low | 3 |
Other Changes in Data Types | - | 4 |
Added Methods (140)
spark-catalyst_2.10-1.0.0.jar, AttributeReference.class
package org.apache.spark.sql.catalyst.expressions
AttributeReference.AttributeReference ( String name, org.apache.spark.sql.catalyst.types.DataType dataType, boolean nullable, ExprId exprId, scala.collection.Seq<String> qualifiers )
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference."<init>":(Ljava/lang/String;Lorg/apache/spark/sql/catalyst/types/DataType;ZLorg/apache/spark/sql/catalyst/expressions/ExprId;Lscala/collection/Seq;)V]
AttributeReference.children ( ) : scala.collection.immutable.Nil.
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.children:()Lscala/collection/immutable/Nil$;]
AttributeReference.copy ( String name, org.apache.spark.sql.catalyst.types.DataType dataType, boolean nullable, ExprId exprId, scala.collection.Seq<String> qualifiers ) : AttributeReference
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.copy:(Ljava/lang/String;Lorg/apache/spark/sql/catalyst/types/DataType;ZLorg/apache/spark/sql/catalyst/expressions/ExprId;Lscala/collection/Seq;)Lorg/apache/spark/sql/catalyst/expressions/AttributeReference;]
AttributeReference.dataType ( ) : org.apache.spark.sql.catalyst.types.DataType
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.dataType:()Lorg/apache/spark/sql/catalyst/types/DataType;]
AttributeReference.eval ( Row input ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.eval:(Lorg/apache/spark/sql/catalyst/expressions/Row;)Ljava/lang/Object;]
spark-catalyst_2.10-1.0.0.jar, Catalog.class
package org.apache.spark.sql.catalyst.analysis
Catalog.lookupRelation ( scala.Option<String> p1, String p2, scala.Option<String> p3 ) [abstract] : org.apache.spark.sql.catalyst.plans.logical.LogicalPlan
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.lookupRelation:(Lscala/Option;Ljava/lang/String;Lscala/Option;)Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;]
Catalog.registerTable ( scala.Option<String> p1, String p2, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan p3 ) [abstract] : void
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.registerTable:(Lscala/Option;Ljava/lang/String;Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)V]
Catalog.unregisterTable ( scala.Option<String> p1, String p2 ) [abstract] : void
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.unregisterTable:(Lscala/Option;Ljava/lang/String;)V]
spark-catalyst_2.10-1.0.0.jar, Expression.class
package org.apache.spark.sql.catalyst.expressions
Expression.c2 ( Row i, Expression e1, Expression e2, scala.Function3<scala.math.Ordering<Object>,Object,Object,Object> f ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.c2:(Lorg/apache/spark/sql/catalyst/expressions/Row;Lorg/apache/spark/sql/catalyst/expressions/Expression;Lorg/apache/spark/sql/catalyst/expressions/Expression;Lscala/Function3;)Ljava/lang/Object;]
Expression.dataType ( ) [abstract] : org.apache.spark.sql.catalyst.types.DataType
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.dataType:()Lorg/apache/spark/sql/catalyst/types/DataType;]
Expression.eval ( Row p1 ) [abstract] : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.eval:(Lorg/apache/spark/sql/catalyst/expressions/Row;)Ljava/lang/Object;]
Expression.f2 ( Row i, Expression e1, Expression e2, scala.Function3<scala.math.Fractional<Object>,Object,Object,Object> f ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.f2:(Lorg/apache/spark/sql/catalyst/expressions/Row;Lorg/apache/spark/sql/catalyst/expressions/Expression;Lorg/apache/spark/sql/catalyst/expressions/Expression;Lscala/Function3;)Ljava/lang/Object;]
Expression.i2 ( Row i, Expression e1, Expression e2, scala.Function3<scala.math.Integral<Object>,Object,Object,Object> f ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.i2:(Lorg/apache/spark/sql/catalyst/expressions/Row;Lorg/apache/spark/sql/catalyst/expressions/Expression;Lorg/apache/spark/sql/catalyst/expressions/Expression;Lscala/Function3;)Ljava/lang/Object;]
Expression.n1 ( Expression e, Row i, scala.Function2<scala.math.Numeric<Object>,Object,Object> f ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.n1:(Lorg/apache/spark/sql/catalyst/expressions/Expression;Lorg/apache/spark/sql/catalyst/expressions/Row;Lscala/Function2;)Ljava/lang/Object;]
Expression.n2 ( Row i, Expression e1, Expression e2, scala.Function3<scala.math.Numeric<Object>,Object,Object,Object> f ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.n2:(Lorg/apache/spark/sql/catalyst/expressions/Row;Lorg/apache/spark/sql/catalyst/expressions/Expression;Lorg/apache/spark/sql/catalyst/expressions/Expression;Lscala/Function3;)Ljava/lang/Object;]
Expression.references ( ) [abstract] : scala.collection.immutable.Set<Attribute>
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.references:()Lscala/collection/immutable/Set;]
spark-catalyst_2.10-1.0.0.jar, GenericMutableRow.class
package org.apache.spark.sql.catalyst.expressions
GenericMutableRow.copy ( ) : GenericRow
[mangled: org/apache/spark/sql/catalyst/expressions/GenericMutableRow.copy:()Lorg/apache/spark/sql/catalyst/expressions/GenericRow;]
GenericMutableRow.copy ( ) : Row
[mangled: org/apache/spark/sql/catalyst/expressions/GenericMutableRow.copy:()Lorg/apache/spark/sql/catalyst/expressions/Row;]
GenericMutableRow.getStringBuilder ( int ordinal ) : scala.collection.mutable.StringBuilder
[mangled: org/apache/spark/sql/catalyst/expressions/GenericMutableRow.getStringBuilder:(I)Lscala/collection/mutable/StringBuilder;]
spark-catalyst_2.10-1.0.0.jar, Literal.class
package org.apache.spark.sql.catalyst.expressions
Literal.copy ( Object value, org.apache.spark.sql.catalyst.types.DataType dataType ) : Literal
[mangled: org/apache/spark/sql/catalyst/expressions/Literal.copy:(Ljava/lang/Object;Lorg/apache/spark/sql/catalyst/types/DataType;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
Literal.dataType ( ) : org.apache.spark.sql.catalyst.types.DataType
[mangled: org/apache/spark/sql/catalyst/expressions/Literal.dataType:()Lorg/apache/spark/sql/catalyst/types/DataType;]
Literal.eval ( Row input ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Literal.eval:(Lorg/apache/spark/sql/catalyst/expressions/Row;)Ljava/lang/Object;]
Literal.Literal ( Object value, org.apache.spark.sql.catalyst.types.DataType dataType )
[mangled: org/apache/spark/sql/catalyst/expressions/Literal."<init>":(Ljava/lang/Object;Lorg/apache/spark/sql/catalyst/types/DataType;)V]
Literal.references ( ) : scala.collection.immutable.Set<Attribute>
[mangled: org/apache/spark/sql/catalyst/expressions/Literal.references:()Lscala/collection/immutable/Set;]
spark-catalyst_2.10-1.0.0.jar, LogicalPlan.class
package org.apache.spark.sql.catalyst.plans.logical
LogicalPlan.inputSet ( ) : scala.collection.immutable.Set<org.apache.spark.sql.catalyst.expressions.Attribute>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.inputSet:()Lscala/collection/immutable/Set;]
LogicalPlan.references ( ) [abstract] : scala.collection.immutable.Set<org.apache.spark.sql.catalyst.expressions.Attribute>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.references:()Lscala/collection/immutable/Set;]
LogicalPlan.resolve ( String name ) : scala.Option<org.apache.spark.sql.catalyst.expressions.NamedExpression>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.resolve:(Ljava/lang/String;)Lscala/Option;]
spark-catalyst_2.10-1.0.0.jar, MutableRow.class
package org.apache.spark.sql.catalyst.expressions
MutableRow.getStringBuilder ( int p1 ) [abstract] : scala.collection.mutable.StringBuilder
[mangled: org/apache/spark/sql/catalyst/expressions/MutableRow.getStringBuilder:(I)Lscala/collection/mutable/StringBuilder;]
spark-catalyst_2.10-1.0.0.jar, SqlParser.class
package org.apache.spark.sql.catalyst
SqlParser.ALL ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ALL:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.AND ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.AND:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.apply ( String input ) : plans.logical.LogicalPlan
[mangled: org/apache/spark/sql/catalyst/SqlParser.apply:(Ljava/lang/String;)Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;]
SqlParser.APPROXIMATE ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.APPROXIMATE:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.AS ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.AS:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.ASC ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ASC:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.asParser ( SqlParser.Keyword k ) : scala.util.parsing.combinator.Parsers.Parser<String>
[mangled: org/apache/spark/sql/catalyst/SqlParser.asParser:(Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;)Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.AVG ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.AVG:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.baseExpression ( ) : scala.util.parsing.combinator.PackratParsers.PackratParser<expressions.Expression>
[mangled: org/apache/spark/sql/catalyst/SqlParser.baseExpression:()Lscala/util/parsing/combinator/PackratParsers$PackratParser;]
SqlParser.BY ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.BY:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.CAST ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.CAST:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.COUNT ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.COUNT:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.DESC ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.DESC:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.DISTINCT ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.DISTINCT:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.FALSE ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.FALSE:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.filter ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Expression>
[mangled: org/apache/spark/sql/catalyst/SqlParser.filter:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.FIRST ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.FIRST:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.FROM ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.FROM:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.from ( ) : scala.util.parsing.combinator.Parsers.Parser<plans.logical.LogicalPlan>
[mangled: org/apache/spark/sql/catalyst/SqlParser.from:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.FULL ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.FULL:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.GROUP ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.GROUP:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.grouping ( ) : scala.util.parsing.combinator.Parsers.Parser<scala.collection.Seq<expressions.Expression>>
[mangled: org/apache/spark/sql/catalyst/SqlParser.grouping:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.HAVING ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.HAVING:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.having ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Expression>
[mangled: org/apache/spark/sql/catalyst/SqlParser.having:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.IF ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.IF:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.IN ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.IN:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.INNER ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.INNER:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.INSERT ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.INSERT:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.INTO ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.INTO:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.inTo ( ) : scala.util.parsing.combinator.Parsers.Parser<plans.logical.LogicalPlan>
[mangled: org/apache/spark/sql/catalyst/SqlParser.inTo:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.IS ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.IS:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.JOIN ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.JOIN:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.Keyword ( ) : SqlParser.Keyword.
[mangled: org/apache/spark/sql/catalyst/SqlParser.Keyword:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword$;]
SqlParser.LEFT ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.LEFT:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.lexical ( ) : SqlParser.SqlLexical
[mangled: org/apache/spark/sql/catalyst/SqlParser.lexical:()Lorg/apache/spark/sql/catalyst/SqlParser$SqlLexical;]
SqlParser.lexical ( ) : scala.util.parsing.combinator.lexical.StdLexical
[mangled: org/apache/spark/sql/catalyst/SqlParser.lexical:()Lscala/util/parsing/combinator/lexical/StdLexical;]
SqlParser.lexical ( ) : scala.util.parsing.combinator.token.Tokens
[mangled: org/apache/spark/sql/catalyst/SqlParser.lexical:()Lscala/util/parsing/combinator/token/Tokens;]
SqlParser.LIKE ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.LIKE:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.LIMIT ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.LIMIT:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.limit ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Expression>
[mangled: org/apache/spark/sql/catalyst/SqlParser.limit:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.MAX ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.MAX:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.memo ( scala.util.parsing.combinator.Parsers.Parser<T> p ) : scala.util.parsing.combinator.PackratParsers.PackratParser<T>
[mangled: org/apache/spark/sql/catalyst/SqlParser.memo:(Lscala/util/parsing/combinator/Parsers$Parser;)Lscala/util/parsing/combinator/PackratParsers$PackratParser;]
SqlParser.MIN ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.MIN:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.NOT ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.NOT:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.NULL ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.NULL:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.ON ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ON:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.OR ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.OR:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.ORDER ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ORDER:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.orderBy ( ) : scala.util.parsing.combinator.Parsers.Parser<scala.collection.Seq<expressions.SortOrder>>
[mangled: org/apache/spark/sql/catalyst/SqlParser.orderBy:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.SqlParser..allCaseVersions ( String s, String prefix ) : scala.collection.immutable.Stream<String>
[mangled: org/apache/spark/sql/catalyst/SqlParser.org.apache.spark.sql.catalyst.SqlParser..allCaseVersions:(Ljava/lang/String;Ljava/lang/String;)Lscala/collection/immutable/Stream;]
SqlParser.OUTER ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.OUTER:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.OVERWRITE ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.OVERWRITE:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.parser2packrat ( scala.Function0<scala.util.parsing.combinator.Parsers.Parser<T>> p ) : scala.util.parsing.combinator.PackratParsers.PackratParser<T>
[mangled: org/apache/spark/sql/catalyst/SqlParser.parser2packrat:(Lscala/Function0;)Lscala/util/parsing/combinator/PackratParsers$PackratParser;]
SqlParser.phrase ( scala.util.parsing.combinator.Parsers.Parser p ) : scala.util.parsing.combinator.Parsers.Parser
[mangled: org/apache/spark/sql/catalyst/SqlParser.phrase:(Lscala/util/parsing/combinator/Parsers$Parser;)Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.phrase ( scala.util.parsing.combinator.Parsers.Parser<T> p ) : scala.util.parsing.combinator.PackratParsers.PackratParser<T>
[mangled: org/apache/spark/sql/catalyst/SqlParser.phrase:(Lscala/util/parsing/combinator/Parsers$Parser;)Lscala/util/parsing/combinator/PackratParsers$PackratParser;]
SqlParser.projections ( ) : scala.util.parsing.combinator.Parsers.Parser<scala.collection.Seq<expressions.Expression>>
[mangled: org/apache/spark/sql/catalyst/SqlParser.projections:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.query ( ) : scala.util.parsing.combinator.Parsers.Parser<plans.logical.LogicalPlan>
[mangled: org/apache/spark/sql/catalyst/SqlParser.query:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.REGEXP ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.REGEXP:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.reservedWords ( ) : SqlParser.Keyword[ ]
[mangled: org/apache/spark/sql/catalyst/SqlParser.reservedWords:()[Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.RIGHT ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.RIGHT:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.RLIKE ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.RLIKE:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.scala.util.parsing.combinator.PackratParsers..Head ( ) : scala.util.parsing.combinator.PackratParsers.Head.
[mangled: org/apache/spark/sql/catalyst/SqlParser.scala.util.parsing.combinator.PackratParsers..Head:()Lscala/util/parsing/combinator/PackratParsers$Head$;]
SqlParser.scala.util.parsing.combinator.PackratParsers..LR ( ) : scala.util.parsing.combinator.PackratParsers.LR.
[mangled: org/apache/spark/sql/catalyst/SqlParser.scala.util.parsing.combinator.PackratParsers..LR:()Lscala/util/parsing/combinator/PackratParsers$LR$;]
SqlParser.scala.util.parsing.combinator.PackratParsers..MemoEntry ( ) : scala.util.parsing.combinator.PackratParsers.MemoEntry.
[mangled: org/apache/spark/sql/catalyst/SqlParser.scala.util.parsing.combinator.PackratParsers..MemoEntry:()Lscala/util/parsing/combinator/PackratParsers$MemoEntry$;]
SqlParser.scala.util.parsing.combinator.PackratParsers..super.Parser ( scala.Function1 f ) : scala.util.parsing.combinator.Parsers.Parser
[mangled: org/apache/spark/sql/catalyst/SqlParser.scala.util.parsing.combinator.PackratParsers..super.Parser:(Lscala/Function1;)Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.scala.util.parsing.combinator.PackratParsers..super.phrase ( scala.util.parsing.combinator.Parsers.Parser p ) : scala.util.parsing.combinator.Parsers.Parser
[mangled: org/apache/spark/sql/catalyst/SqlParser.scala.util.parsing.combinator.PackratParsers..super.phrase:(Lscala/util/parsing/combinator/Parsers$Parser;)Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.SELECT ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SELECT:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.singleOrder ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.SortOrder>
[mangled: org/apache/spark/sql/catalyst/SqlParser.singleOrder:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.STRING ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.STRING:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.SUM ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SUM:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.TRUE ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.TRUE:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.UNION ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.UNION:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
SqlParser.WHERE ( ) : SqlParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.WHERE:()Lorg/apache/spark/sql/catalyst/SqlParser$Keyword;]
spark-core_2.10-1.0.0.jar, SparkContext.class
package org.apache.spark
SparkContext.clean ( F f ) : F
[mangled: org/apache/spark/SparkContext.clean:(Ljava/lang/Object;)Ljava/lang/Object;]
SparkContext.getCallSite ( ) : String
[mangled: org/apache/spark/SparkContext.getCallSite:()Ljava/lang/String;]
SparkContext.ui ( ) : ui.SparkUI
[mangled: org/apache/spark/SparkContext.ui:()Lorg/apache/spark/ui/SparkUI;]
spark-core_2.10-1.0.0.jar, TaskContext.class
package org.apache.spark
TaskContext.completed ( ) : boolean
[mangled: org/apache/spark/TaskContext.completed:()Z]
TaskContext.completed_.eq ( boolean p1 ) : void
[mangled: org/apache/spark/TaskContext.completed_.eq:(Z)V]
TaskContext.executeOnCompleteCallbacks ( ) : void
[mangled: org/apache/spark/TaskContext.executeOnCompleteCallbacks:()V]
TaskContext.interrupted ( ) : boolean
[mangled: org/apache/spark/TaskContext.interrupted:()Z]
TaskContext.interrupted_.eq ( boolean p1 ) : void
[mangled: org/apache/spark/TaskContext.interrupted_.eq:(Z)V]
TaskContext.TaskContext ( int stageId, int partitionId, long attemptId, boolean runningLocally, executor.TaskMetrics taskMetrics )
[mangled: org/apache/spark/TaskContext."<init>":(IIJZLorg/apache/spark/executor/TaskMetrics;)V]
spark-sql_2.10-1.0.0.jar, SparkPlan.class
package org.apache.spark.sql.execution
SparkPlan.buildRow ( scala.collection.Seq<Object> values ) : org.apache.spark.sql.catalyst.expressions.Row
[mangled: org/apache/spark/sql/execution/SparkPlan.buildRow:(Lscala/collection/Seq;)Lorg/apache/spark/sql/catalyst/expressions/Row;]
SparkPlan.executeCollect ( ) : org.apache.spark.sql.catalyst.expressions.Row[ ]
[mangled: org/apache/spark/sql/execution/SparkPlan.executeCollect:()[Lorg/apache/spark/sql/catalyst/expressions/Row;]
SparkPlan.logger ( ) : com.typesafe.scalalogging.slf4j.Logger
[mangled: org/apache/spark/sql/execution/SparkPlan.logger:()Lcom/typesafe/scalalogging/slf4j/Logger;]
spark-sql_2.10-1.0.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.binaryToLiteral ( byte[ ] a ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.binaryToLiteral:([B)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.booleanToLiteral ( boolean b ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.booleanToLiteral:(Z)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.byteToLiteral ( byte b ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.byteToLiteral:(B)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.createParquetFile ( String path, boolean allowExisting, org.apache.hadoop.conf.Configuration conf, scala.reflect.api.TypeTags.TypeTag<A> p4 ) : SchemaRDD
[mangled: org/apache/spark/sql/SQLContext.createParquetFile:(Ljava/lang/String;ZLorg/apache/hadoop/conf/Configuration;Lscala/reflect/api/TypeTags$TypeTag;)Lorg/apache/spark/sql/SchemaRDD;]
SQLContext.createSchemaRDD ( org.apache.spark.rdd.RDD<A> rdd, scala.reflect.api.TypeTags.TypeTag<A> p2 ) : SchemaRDD
[mangled: org/apache/spark/sql/SQLContext.createSchemaRDD:(Lorg/apache/spark/rdd/RDD;Lscala/reflect/api/TypeTags$TypeTag;)Lorg/apache/spark/sql/SchemaRDD;]
SQLContext.decimalToLiteral ( scala.math.BigDecimal d ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.decimalToLiteral:(Lscala/math/BigDecimal;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.doubleToLiteral ( double d ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.doubleToLiteral:(D)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.DslAttribute ( catalyst.expressions.AttributeReference a ) : catalyst.dsl.package.ExpressionConversions.DslAttribute
[mangled: org/apache/spark/sql/SQLContext.DslAttribute:(Lorg/apache/spark/sql/catalyst/expressions/AttributeReference;)Lorg/apache/spark/sql/catalyst/dsl/package$ExpressionConversions$DslAttribute;]
SQLContext.DslExpression ( catalyst.expressions.Expression e ) : catalyst.dsl.package.ExpressionConversions.DslExpression
[mangled: org/apache/spark/sql/SQLContext.DslExpression:(Lorg/apache/spark/sql/catalyst/expressions/Expression;)Lorg/apache/spark/sql/catalyst/dsl/package$ExpressionConversions$DslExpression;]
SQLContext.DslString ( String s ) : catalyst.dsl.package.ExpressionConversions.DslString
[mangled: org/apache/spark/sql/SQLContext.DslString:(Ljava/lang/String;)Lorg/apache/spark/sql/catalyst/dsl/package$ExpressionConversions$DslString;]
SQLContext.DslSymbol ( scala.Symbol sym ) : catalyst.dsl.package.ExpressionConversions.DslSymbol
[mangled: org/apache/spark/sql/SQLContext.DslSymbol:(Lscala/Symbol;)Lorg/apache/spark/sql/catalyst/dsl/package$ExpressionConversions$DslSymbol;]
SQLContext.floatToLiteral ( float f ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.floatToLiteral:(F)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.inferSchema ( org.apache.spark.rdd.RDD<scala.collection.immutable.Map<String,Object>> rdd ) : SchemaRDD
[mangled: org/apache/spark/sql/SQLContext.inferSchema:(Lorg/apache/spark/rdd/RDD;)Lorg/apache/spark/sql/SchemaRDD;]
SQLContext.intToLiteral ( int i ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.intToLiteral:(I)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.logger ( ) : com.typesafe.scalalogging.slf4j.Logger
[mangled: org/apache/spark/sql/SQLContext.logger:()Lcom/typesafe/scalalogging/slf4j/Logger;]
SQLContext.logicalPlanToSparkQuery ( catalyst.plans.logical.LogicalPlan plan ) : SchemaRDD
[mangled: org/apache/spark/sql/SQLContext.logicalPlanToSparkQuery:(Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)Lorg/apache/spark/sql/SchemaRDD;]
SQLContext.longToLiteral ( long l ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.longToLiteral:(J)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.optimizer ( ) : catalyst.optimizer.Optimizer.
[mangled: org/apache/spark/sql/SQLContext.optimizer:()Lorg/apache/spark/sql/catalyst/optimizer/Optimizer$;]
SQLContext.parquetFile ( String path ) : SchemaRDD
[mangled: org/apache/spark/sql/SQLContext.parquetFile:(Ljava/lang/String;)Lorg/apache/spark/sql/SchemaRDD;]
SQLContext.parser ( ) : catalyst.SqlParser
[mangled: org/apache/spark/sql/SQLContext.parser:()Lorg/apache/spark/sql/catalyst/SqlParser;]
SQLContext.registerRDDAsTable ( SchemaRDD rdd, String tableName ) : void
[mangled: org/apache/spark/sql/SQLContext.registerRDDAsTable:(Lorg/apache/spark/sql/SchemaRDD;Ljava/lang/String;)V]
SQLContext.shortToLiteral ( short s ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.shortToLiteral:(S)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.sql ( String sqlText ) : SchemaRDD
[mangled: org/apache/spark/sql/SQLContext.sql:(Ljava/lang/String;)Lorg/apache/spark/sql/SchemaRDD;]
SQLContext.stringToLiteral ( String s ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.stringToLiteral:(Ljava/lang/String;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
SQLContext.symbolToUnresolvedAttribute ( scala.Symbol s ) : catalyst.analysis.UnresolvedAttribute
[mangled: org/apache/spark/sql/SQLContext.symbolToUnresolvedAttribute:(Lscala/Symbol;)Lorg/apache/spark/sql/catalyst/analysis/UnresolvedAttribute;]
SQLContext.table ( String tableName ) : SchemaRDD
[mangled: org/apache/spark/sql/SQLContext.table:(Ljava/lang/String;)Lorg/apache/spark/sql/SchemaRDD;]
SQLContext.timestampToLiteral ( java.sql.Timestamp t ) : catalyst.expressions.Literal
[mangled: org/apache/spark/sql/SQLContext.timestampToLiteral:(Ljava/sql/Timestamp;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
to the top
Removed Methods (630)
spark-catalyst_2.10-1.4.0.jar, AtomicType.class
package org.apache.spark.sql.types
AtomicType.AtomicType ( )
[mangled: org/apache/spark/sql/types/AtomicType."<init>":()V]
AtomicType.classTag ( ) : scala.reflect.ClassTag<Object>
[mangled: org/apache/spark/sql/types/AtomicType.classTag:()Lscala/reflect/ClassTag;]
AtomicType.ordering ( ) [abstract] : scala.math.Ordering<Object>
[mangled: org/apache/spark/sql/types/AtomicType.ordering:()Lscala/math/Ordering;]
AtomicType.tag ( ) [abstract] : scala.reflect.api.TypeTags.TypeTag<Object>
[mangled: org/apache/spark/sql/types/AtomicType.tag:()Lscala/reflect/api/TypeTags$TypeTag;]
spark-catalyst_2.10-1.4.0.jar, AttributeReference.class
package org.apache.spark.sql.catalyst.expressions
AttributeReference.AttributeReference ( String name, org.apache.spark.sql.types.DataType dataType, boolean nullable, org.apache.spark.sql.types.Metadata metadata, ExprId exprId, scala.collection.Seq<String> qualifiers )
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference."<init>":(Ljava/lang/String;Lorg/apache/spark/sql/types/DataType;ZLorg/apache/spark/sql/types/Metadata;Lorg/apache/spark/sql/catalyst/expressions/ExprId;Lscala/collection/Seq;)V]
AttributeReference.copy ( String name, org.apache.spark.sql.types.DataType dataType, boolean nullable, org.apache.spark.sql.types.Metadata metadata, ExprId exprId, scala.collection.Seq<String> qualifiers ) : AttributeReference
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.copy:(Ljava/lang/String;Lorg/apache/spark/sql/types/DataType;ZLorg/apache/spark/sql/types/Metadata;Lorg/apache/spark/sql/catalyst/expressions/ExprId;Lscala/collection/Seq;)Lorg/apache/spark/sql/catalyst/expressions/AttributeReference;]
AttributeReference.eval ( org.apache.spark.sql.Row input ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.eval:(Lorg/apache/spark/sql/Row;)Ljava/lang/Object;]
AttributeReference.metadata ( ) : org.apache.spark.sql.types.Metadata
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.metadata:()Lorg/apache/spark/sql/types/Metadata;]
AttributeReference.sameRef ( AttributeReference other ) : boolean
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.sameRef:(Lorg/apache/spark/sql/catalyst/expressions/AttributeReference;)Z]
AttributeReference.semanticEquals ( Expression other ) : boolean
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.semanticEquals:(Lorg/apache/spark/sql/catalyst/expressions/Expression;)Z]
AttributeReference.withName ( String newName ) : Attribute
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.withName:(Ljava/lang/String;)Lorg/apache/spark/sql/catalyst/expressions/Attribute;]
AttributeReference.withName ( String newName ) : AttributeReference
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.withName:(Ljava/lang/String;)Lorg/apache/spark/sql/catalyst/expressions/AttributeReference;]
AttributeReference.withNullability ( boolean newNullability ) : Attribute
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeReference.withNullability:(Z)Lorg/apache/spark/sql/catalyst/expressions/Attribute;]
spark-catalyst_2.10-1.4.0.jar, AttributeSet.class
package org.apache.spark.sql.catalyst.expressions
AttributeSet.exists ( scala.Function1<Attribute,Object> p ) : boolean
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeSet.exists:(Lscala/Function1;)Z]
AttributeSet.map ( scala.Function1<Attribute,B> f, scala.collection.generic.CanBuildFrom<scala.collection.Traversable<Attribute>,B,That> bf ) : That
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeSet.map:(Lscala/Function1;Lscala/collection/generic/CanBuildFrom;)Ljava/lang/Object;]
AttributeSet.size ( ) : int
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeSet.size:()I]
AttributeSet.subsetOf ( AttributeSet other ) : boolean
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeSet.subsetOf:(Lorg/apache/spark/sql/catalyst/expressions/AttributeSet;)Z]
AttributeSet.toSeq ( ) : scala.collection.Seq<Attribute>
[mangled: org/apache/spark/sql/catalyst/expressions/AttributeSet.toSeq:()Lscala/collection/Seq;]
spark-catalyst_2.10-1.4.0.jar, BoundReference.class
package org.apache.spark.sql.catalyst.expressions
BoundReference.eval ( org.apache.spark.sql.Row input ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/BoundReference.eval:(Lorg/apache/spark/sql/Row;)Ljava/lang/Object;]
spark-catalyst_2.10-1.4.0.jar, Catalog.class
package org.apache.spark.sql.catalyst.analysis
Catalog.conf ( ) [abstract] : org.apache.spark.sql.catalyst.CatalystConf
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.conf:()Lorg/apache/spark/sql/catalyst/CatalystConf;]
Catalog.getDBTable ( scala.collection.Seq<String> p1 ) [abstract] : scala.Tuple2<scala.Option<String>,String>
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.getDBTable:(Lscala/collection/Seq;)Lscala/Tuple2;]
Catalog.getDbTableName ( scala.collection.Seq<String> p1 ) [abstract] : String
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.getDbTableName:(Lscala/collection/Seq;)Ljava/lang/String;]
Catalog.getTables ( scala.Option<String> p1 ) [abstract] : scala.collection.Seq<scala.Tuple2<String,Object>>
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.getTables:(Lscala/Option;)Lscala/collection/Seq;]
Catalog.lookupRelation ( scala.collection.Seq<String> p1, scala.Option<String> p2 ) [abstract] : org.apache.spark.sql.catalyst.plans.logical.LogicalPlan
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.lookupRelation:(Lscala/collection/Seq;Lscala/Option;)Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;]
Catalog.processTableIdentifier ( scala.collection.Seq<String> p1 ) [abstract] : scala.collection.Seq<String>
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.processTableIdentifier:(Lscala/collection/Seq;)Lscala/collection/Seq;]
Catalog.refreshTable ( String p1, String p2 ) [abstract] : void
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.refreshTable:(Ljava/lang/String;Ljava/lang/String;)V]
Catalog.registerTable ( scala.collection.Seq<String> p1, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan p2 ) [abstract] : void
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.registerTable:(Lscala/collection/Seq;Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)V]
Catalog.tableExists ( scala.collection.Seq<String> p1 ) [abstract] : boolean
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.tableExists:(Lscala/collection/Seq;)Z]
Catalog.unregisterTable ( scala.collection.Seq<String> p1 ) [abstract] : void
[mangled: org/apache/spark/sql/catalyst/analysis/Catalog.unregisterTable:(Lscala/collection/Seq;)V]
spark-catalyst_2.10-1.4.0.jar, DataType.class
package org.apache.spark.sql.types
DataType.asNullable ( ) [abstract] : DataType
[mangled: org/apache/spark/sql/types/DataType.asNullable:()Lorg/apache/spark/sql/types/DataType;]
DataType.DataType ( )
[mangled: org/apache/spark/sql/types/DataType."<init>":()V]
DataType.defaultSize ( ) [abstract] : int
[mangled: org/apache/spark/sql/types/DataType.defaultSize:()I]
DataType.fromCaseClassString ( String p1 ) [static] : DataType
[mangled: org/apache/spark/sql/types/DataType.fromCaseClassString:(Ljava/lang/String;)Lorg/apache/spark/sql/types/DataType;]
DataType.fromJson ( String p1 ) [static] : DataType
[mangled: org/apache/spark/sql/types/DataType.fromJson:(Ljava/lang/String;)Lorg/apache/spark/sql/types/DataType;]
DataType.json ( ) : String
[mangled: org/apache/spark/sql/types/DataType.json:()Ljava/lang/String;]
DataType.jsonValue ( ) : org.json4s.JsonAST.JValue
[mangled: org/apache/spark/sql/types/DataType.jsonValue:()Lorg/json4s/JsonAST$JValue;]
DataType.prettyJson ( ) : String
[mangled: org/apache/spark/sql/types/DataType.prettyJson:()Ljava/lang/String;]
DataType.sameType ( DataType other ) : boolean
[mangled: org/apache/spark/sql/types/DataType.sameType:(Lorg/apache/spark/sql/types/DataType;)Z]
DataType.simpleString ( ) : String
[mangled: org/apache/spark/sql/types/DataType.simpleString:()Ljava/lang/String;]
DataType.typeName ( ) : String
[mangled: org/apache/spark/sql/types/DataType.typeName:()Ljava/lang/String;]
DataType.unapply ( org.apache.spark.sql.catalyst.expressions.Expression e ) : boolean
[mangled: org/apache/spark/sql/types/DataType.unapply:(Lorg/apache/spark/sql/catalyst/expressions/Expression;)Z]
spark-catalyst_2.10-1.4.0.jar, EqualTo.class
package org.apache.spark.sql.catalyst.expressions
EqualTo.EqualTo ( Expression left, Expression right )
[mangled: org/apache/spark/sql/catalyst/expressions/EqualTo."<init>":(Lorg/apache/spark/sql/catalyst/expressions/Expression;Lorg/apache/spark/sql/catalyst/expressions/Expression;)V]
EqualTo.left ( ) : Expression
[mangled: org/apache/spark/sql/catalyst/expressions/EqualTo.left:()Lorg/apache/spark/sql/catalyst/expressions/Expression;]
EqualTo.right ( ) : Expression
[mangled: org/apache/spark/sql/catalyst/expressions/EqualTo.right:()Lorg/apache/spark/sql/catalyst/expressions/Expression;]
spark-catalyst_2.10-1.4.0.jar, Expression.class
package org.apache.spark.sql.catalyst.expressions
Expression.dataType ( ) [abstract] : org.apache.spark.sql.types.DataType
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.dataType:()Lorg/apache/spark/sql/types/DataType;]
Expression.deterministic ( ) : boolean
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.deterministic:()Z]
Expression.eval ( org.apache.spark.sql.Row p1 ) [abstract] : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.eval:(Lorg/apache/spark/sql/Row;)Ljava/lang/Object;]
Expression.prettyString ( ) : String
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.prettyString:()Ljava/lang/String;]
Expression.references ( ) : AttributeSet
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.references:()Lorg/apache/spark/sql/catalyst/expressions/AttributeSet;]
Expression.semanticEquals ( Expression other ) : boolean
[mangled: org/apache/spark/sql/catalyst/expressions/Expression.semanticEquals:(Lorg/apache/spark/sql/catalyst/expressions/Expression;)Z]
spark-catalyst_2.10-1.4.0.jar, GenericMutableRow.class
package org.apache.spark.sql.catalyst.expressions
GenericMutableRow.GenericMutableRow ( Object[ ] v )
[mangled: org/apache/spark/sql/catalyst/expressions/GenericMutableRow."<init>":([Ljava/lang/Object;)V]
spark-catalyst_2.10-1.4.0.jar, InSet.class
package org.apache.spark.sql.catalyst.expressions
InSet.hset ( ) : scala.collection.immutable.Set<Object>
[mangled: org/apache/spark/sql/catalyst/expressions/InSet.hset:()Lscala/collection/immutable/Set;]
InSet.InSet ( Expression value, scala.collection.immutable.Set<Object> hset )
[mangled: org/apache/spark/sql/catalyst/expressions/InSet."<init>":(Lorg/apache/spark/sql/catalyst/expressions/Expression;Lscala/collection/immutable/Set;)V]
InSet.value ( ) : Expression
[mangled: org/apache/spark/sql/catalyst/expressions/InSet.value:()Lorg/apache/spark/sql/catalyst/expressions/Expression;]
spark-catalyst_2.10-1.4.0.jar, Literal.class
package org.apache.spark.sql.catalyst.expressions
Literal.copy ( Object value, org.apache.spark.sql.types.DataType dataType ) : Literal
[mangled: org/apache/spark/sql/catalyst/expressions/Literal.copy:(Ljava/lang/Object;Lorg/apache/spark/sql/types/DataType;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
Literal.create ( Object p1, org.apache.spark.sql.types.DataType p2 ) [static] : Literal
[mangled: org/apache/spark/sql/catalyst/expressions/Literal.create:(Ljava/lang/Object;Lorg/apache/spark/sql/types/DataType;)Lorg/apache/spark/sql/catalyst/expressions/Literal;]
Literal.eval ( org.apache.spark.sql.Row input ) : Object
[mangled: org/apache/spark/sql/catalyst/expressions/Literal.eval:(Lorg/apache/spark/sql/Row;)Ljava/lang/Object;]
Literal.Literal ( Object value, org.apache.spark.sql.types.DataType dataType )
[mangled: org/apache/spark/sql/catalyst/expressions/Literal."<init>":(Ljava/lang/Object;Lorg/apache/spark/sql/types/DataType;)V]
spark-catalyst_2.10-1.4.0.jar, LogicalPlan.class
package org.apache.spark.sql.catalyst.plans.logical
LogicalPlan.cleanArgs ( ) : scala.collection.Seq<Object>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.cleanArgs:()Lscala/collection/Seq;]
LogicalPlan.isTraceEnabled ( ) : boolean
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.isTraceEnabled:()Z]
LogicalPlan.log ( ) : org.slf4j.Logger
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.log:()Lorg/slf4j/Logger;]
LogicalPlan.logDebug ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logDebug:(Lscala/Function0;)V]
LogicalPlan.logDebug ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logDebug:(Lscala/Function0;Ljava/lang/Throwable;)V]
LogicalPlan.logError ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logError:(Lscala/Function0;)V]
LogicalPlan.logError ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logError:(Lscala/Function0;Ljava/lang/Throwable;)V]
LogicalPlan.logInfo ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logInfo:(Lscala/Function0;)V]
LogicalPlan.logInfo ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logInfo:(Lscala/Function0;Ljava/lang/Throwable;)V]
LogicalPlan.logName ( ) : String
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logName:()Ljava/lang/String;]
LogicalPlan.logTrace ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logTrace:(Lscala/Function0;)V]
LogicalPlan.logTrace ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logTrace:(Lscala/Function0;Ljava/lang/Throwable;)V]
LogicalPlan.logWarning ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logWarning:(Lscala/Function0;)V]
LogicalPlan.logWarning ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.logWarning:(Lscala/Function0;Ljava/lang/Throwable;)V]
LogicalPlan.org.apache.spark.Logging..log_ ( ) : org.slf4j.Logger
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.org.apache.spark.Logging..log_:()Lorg/slf4j/Logger;]
LogicalPlan.org.apache.spark.Logging..log__.eq ( org.slf4j.Logger p1 ) : void
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.org.apache.spark.Logging..log__.eq:(Lorg/slf4j/Logger;)V]
LogicalPlan.LogicalPlan..resolveAsColumn ( scala.collection.Seq<String> nameParts, scala.Function2<String,String,Object> resolver, org.apache.spark.sql.catalyst.expressions.Attribute attribute ) : scala.Option<scala.Tuple2<org.apache.spark.sql.catalyst.expressions.Attribute,scala.collection.immutable.List<String>>>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.org.apache.spark.sql.catalyst.plans.logical.LogicalPlan..resolveAsColumn:(Lscala/collection/Seq;Lscala/Function2;Lorg/apache/spark/sql/catalyst/expressions/Attribute;)Lscala/Option;]
LogicalPlan.LogicalPlan..resolveAsTableColumn ( scala.collection.Seq<String> nameParts, scala.Function2<String,String,Object> resolver, org.apache.spark.sql.catalyst.expressions.Attribute attribute ) : scala.Option<scala.Tuple2<org.apache.spark.sql.catalyst.expressions.Attribute,scala.collection.immutable.List<String>>>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.org.apache.spark.sql.catalyst.plans.logical.LogicalPlan..resolveAsTableColumn:(Lscala/collection/Seq;Lscala/Function2;Lorg/apache/spark/sql/catalyst/expressions/Attribute;)Lscala/Option;]
LogicalPlan.resolve ( scala.collection.Seq<String> nameParts, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> input, scala.Function2<String,String,Object> resolver, boolean throwErrors ) : scala.Option<org.apache.spark.sql.catalyst.expressions.NamedExpression>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.resolve:(Lscala/collection/Seq;Lscala/collection/Seq;Lscala/Function2;Z)Lscala/Option;]
LogicalPlan.resolve ( scala.collection.Seq<String> nameParts, scala.Function2<String,String,Object> resolver, boolean throwErrors ) : scala.Option<org.apache.spark.sql.catalyst.expressions.NamedExpression>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.resolve:(Lscala/collection/Seq;Lscala/Function2;Z)Lscala/Option;]
LogicalPlan.resolveChildren ( scala.collection.Seq<String> nameParts, scala.Function2<String,String,Object> resolver, boolean throwErrors ) : scala.Option<org.apache.spark.sql.catalyst.expressions.NamedExpression>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.resolveChildren:(Lscala/collection/Seq;Lscala/Function2;Z)Lscala/Option;]
LogicalPlan.resolveQuoted ( String name, scala.Function2<String,String,Object> resolver ) : scala.Option<org.apache.spark.sql.catalyst.expressions.NamedExpression>
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.resolveQuoted:(Ljava/lang/String;Lscala/Function2;)Lscala/Option;]
LogicalPlan.sameResult ( LogicalPlan plan ) : boolean
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.sameResult:(Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)Z]
LogicalPlan.statePrefix ( ) : String
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.statePrefix:()Ljava/lang/String;]
LogicalPlan.statistics ( ) : Statistics
[mangled: org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.statistics:()Lorg/apache/spark/sql/catalyst/plans/logical/Statistics;]
spark-catalyst_2.10-1.4.0.jar, ParserDialect.class
package org.apache.spark.sql.catalyst
ParserDialect.parse ( String p1 ) [abstract] : plans.logical.LogicalPlan
[mangled: org/apache/spark/sql/catalyst/ParserDialect.parse:(Ljava/lang/String;)Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;]
ParserDialect.ParserDialect ( )
[mangled: org/apache/spark/sql/catalyst/ParserDialect."<init>":()V]
spark-catalyst_2.10-1.4.0.jar, Row.class
package org.apache.spark.sql
Row.anyNull ( ) [abstract] : boolean
[mangled: org/apache/spark/sql/Row.anyNull:()Z]
Row.apply ( int p1 ) [abstract] : Object
[mangled: org/apache/spark/sql/Row.apply:(I)Ljava/lang/Object;]
Row.copy ( ) [abstract] : Row
[mangled: org/apache/spark/sql/Row.copy:()Lorg/apache/spark/sql/Row;]
Row.equals ( Object p1 ) [abstract] : boolean
[mangled: org/apache/spark/sql/Row.equals:(Ljava/lang/Object;)Z]
Row.fieldIndex ( String p1 ) [abstract] : int
[mangled: org/apache/spark/sql/Row.fieldIndex:(Ljava/lang/String;)I]
Row.get ( int p1 ) [abstract] : Object
[mangled: org/apache/spark/sql/Row.get:(I)Ljava/lang/Object;]
Row.getAs ( int p1 ) [abstract] : T
[mangled: org/apache/spark/sql/Row.getAs:(I)Ljava/lang/Object;]
Row.getAs ( String p1 ) [abstract] : T
[mangled: org/apache/spark/sql/Row.getAs:(Ljava/lang/String;)Ljava/lang/Object;]
Row.getBoolean ( int p1 ) [abstract] : boolean
[mangled: org/apache/spark/sql/Row.getBoolean:(I)Z]
Row.getByte ( int p1 ) [abstract] : byte
[mangled: org/apache/spark/sql/Row.getByte:(I)B]
Row.getDate ( int p1 ) [abstract] : java.sql.Date
[mangled: org/apache/spark/sql/Row.getDate:(I)Ljava/sql/Date;]
Row.getDecimal ( int p1 ) [abstract] : java.math.BigDecimal
[mangled: org/apache/spark/sql/Row.getDecimal:(I)Ljava/math/BigDecimal;]
Row.getDouble ( int p1 ) [abstract] : double
[mangled: org/apache/spark/sql/Row.getDouble:(I)D]
Row.getFloat ( int p1 ) [abstract] : float
[mangled: org/apache/spark/sql/Row.getFloat:(I)F]
Row.getInt ( int p1 ) [abstract] : int
[mangled: org/apache/spark/sql/Row.getInt:(I)I]
Row.getJavaMap ( int p1 ) [abstract] : java.util.Map<K,V>
[mangled: org/apache/spark/sql/Row.getJavaMap:(I)Ljava/util/Map;]
Row.getList ( int p1 ) [abstract] : java.util.List<T>
[mangled: org/apache/spark/sql/Row.getList:(I)Ljava/util/List;]
Row.getLong ( int p1 ) [abstract] : long
[mangled: org/apache/spark/sql/Row.getLong:(I)J]
Row.getMap ( int p1 ) [abstract] : scala.collection.Map<K,V>
[mangled: org/apache/spark/sql/Row.getMap:(I)Lscala/collection/Map;]
Row.getSeq ( int p1 ) [abstract] : scala.collection.Seq<T>
[mangled: org/apache/spark/sql/Row.getSeq:(I)Lscala/collection/Seq;]
Row.getShort ( int p1 ) [abstract] : short
[mangled: org/apache/spark/sql/Row.getShort:(I)S]
Row.getString ( int p1 ) [abstract] : String
[mangled: org/apache/spark/sql/Row.getString:(I)Ljava/lang/String;]
Row.getStruct ( int p1 ) [abstract] : Row
[mangled: org/apache/spark/sql/Row.getStruct:(I)Lorg/apache/spark/sql/Row;]
Row.getValuesMap ( scala.collection.Seq<String> p1 ) [abstract] : scala.collection.immutable.Map<String,T>
[mangled: org/apache/spark/sql/Row.getValuesMap:(Lscala/collection/Seq;)Lscala/collection/immutable/Map;]
Row.hashCode ( ) [abstract] : int
[mangled: org/apache/spark/sql/Row.hashCode:()I]
Row.isNullAt ( int p1 ) [abstract] : boolean
[mangled: org/apache/spark/sql/Row.isNullAt:(I)Z]
Row.length ( ) [abstract] : int
[mangled: org/apache/spark/sql/Row.length:()I]
Row.mkString ( ) [abstract] : String
[mangled: org/apache/spark/sql/Row.mkString:()Ljava/lang/String;]
Row.mkString ( String p1 ) [abstract] : String
[mangled: org/apache/spark/sql/Row.mkString:(Ljava/lang/String;)Ljava/lang/String;]
Row.mkString ( String p1, String p2, String p3 ) [abstract] : String
[mangled: org/apache/spark/sql/Row.mkString:(Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Ljava/lang/String;]
Row.schema ( ) [abstract] : types.StructType
[mangled: org/apache/spark/sql/Row.schema:()Lorg/apache/spark/sql/types/StructType;]
Row.size ( ) [abstract] : int
[mangled: org/apache/spark/sql/Row.size:()I]
Row.toSeq ( ) [abstract] : scala.collection.Seq<Object>
[mangled: org/apache/spark/sql/Row.toSeq:()Lscala/collection/Seq;]
Row.toString ( ) [abstract] : String
[mangled: org/apache/spark/sql/Row.toString:()Ljava/lang/String;]
spark-catalyst_2.10-1.4.0.jar, SimpleCatalystConf.class
package org.apache.spark.sql.catalyst
SimpleCatalystConf.SimpleCatalystConf ( boolean caseSensitiveAnalysis )
[mangled: org/apache/spark/sql/catalyst/SimpleCatalystConf."<init>":(Z)V]
spark-catalyst_2.10-1.4.0.jar, SqlParser.class
package org.apache.spark.sql.catalyst
SqlParser.ABS ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ABS:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.ALL ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ALL:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.AND ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.AND:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.APPROXIMATE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.APPROXIMATE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.arrayType ( ) : scala.util.parsing.combinator.Parsers.Parser<org.apache.spark.sql.types.DataType>
[mangled: org/apache/spark/sql/catalyst/SqlParser.arrayType:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.AS ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.AS:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.ASC ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ASC:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.AVG ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.AVG:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.baseExpression ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Expression>
[mangled: org/apache/spark/sql/catalyst/SqlParser.baseExpression:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.BETWEEN ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.BETWEEN:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.booleanLiteral ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Literal>
[mangled: org/apache/spark/sql/catalyst/SqlParser.booleanLiteral:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.BY ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.BY:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.CASE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.CASE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.CAST ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.CAST:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.COALESCE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.COALESCE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.COUNT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.COUNT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.cte ( ) : scala.util.parsing.combinator.Parsers.Parser<plans.logical.LogicalPlan>
[mangled: org/apache/spark/sql/catalyst/SqlParser.cte:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.DESC ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.DESC:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.DISTINCT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.DISTINCT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.dotExpressionHeader ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Expression>
[mangled: org/apache/spark/sql/catalyst/SqlParser.dotExpressionHeader:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.ELSE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ELSE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.END ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.END:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.EXCEPT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.EXCEPT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.FALSE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.FALSE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.FIRST ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.FIRST:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.fixedDecimalType ( ) : scala.util.parsing.combinator.Parsers.Parser<org.apache.spark.sql.types.DataType>
[mangled: org/apache/spark/sql/catalyst/SqlParser.fixedDecimalType:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.FROM ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.FROM:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.FULL ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.FULL:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.GROUP ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.GROUP:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.HAVING ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.HAVING:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.IF ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.IF:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.IN ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.IN:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.INNER ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.INNER:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.INSERT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.INSERT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.INTERSECT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.INTERSECT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.INTO ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.INTO:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.IS ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.IS:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.JOIN ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.JOIN:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.LAST ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.LAST:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.LEFT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.LEFT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.LIKE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.LIKE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.LIMIT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.LIMIT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.LOWER ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.LOWER:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.mapType ( ) : scala.util.parsing.combinator.Parsers.Parser<org.apache.spark.sql.types.DataType>
[mangled: org/apache/spark/sql/catalyst/SqlParser.mapType:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.MAX ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.MAX:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.MIN ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.MIN:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.NOT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.NOT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.NULL ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.NULL:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.numericLiteral ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Literal>
[mangled: org/apache/spark/sql/catalyst/SqlParser.numericLiteral:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.ON ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ON:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.OR ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.OR:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.ORDER ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.ORDER:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.SqlParser..toNarrowestIntegerType ( String value ) : Object
[mangled: org/apache/spark/sql/catalyst/SqlParser.org.apache.spark.sql.catalyst.SqlParser..toNarrowestIntegerType:(Ljava/lang/String;)Ljava/lang/Object;]
SqlParser.OUTER ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.OUTER:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.OVERWRITE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.OVERWRITE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.parseExpression ( String input ) : expressions.Expression
[mangled: org/apache/spark/sql/catalyst/SqlParser.parseExpression:(Ljava/lang/String;)Lorg/apache/spark/sql/catalyst/expressions/Expression;]
SqlParser.primary ( ) : scala.util.parsing.combinator.PackratParsers.PackratParser<expressions.Expression>
[mangled: org/apache/spark/sql/catalyst/SqlParser.primary:()Lscala/util/parsing/combinator/PackratParsers$PackratParser;]
SqlParser.primitiveType ( ) : scala.util.parsing.combinator.Parsers.Parser<org.apache.spark.sql.types.DataType>
[mangled: org/apache/spark/sql/catalyst/SqlParser.primitiveType:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.REGEXP ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.REGEXP:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.regexToParser ( scala.util.matching.Regex regex ) : scala.util.parsing.combinator.Parsers.Parser<String>
[mangled: org/apache/spark/sql/catalyst/SqlParser.regexToParser:(Lscala/util/matching/Regex;)Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.RIGHT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.RIGHT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.RLIKE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.RLIKE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.SELECT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SELECT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.SEMI ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SEMI:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.sign ( ) : scala.util.parsing.combinator.Parsers.Parser<String>
[mangled: org/apache/spark/sql/catalyst/SqlParser.sign:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.signedNumericLiteral ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Literal>
[mangled: org/apache/spark/sql/catalyst/SqlParser.signedNumericLiteral:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.signedPrimary ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Expression>
[mangled: org/apache/spark/sql/catalyst/SqlParser.signedPrimary:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.SORT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SORT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.sortType ( ) : scala.util.parsing.combinator.Parsers.Parser<scala.Function1<plans.logical.LogicalPlan,plans.logical.LogicalPlan>>
[mangled: org/apache/spark/sql/catalyst/SqlParser.sortType:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.SQRT ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SQRT:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.start ( ) : scala.util.parsing.combinator.Parsers.Parser<plans.logical.LogicalPlan>
[mangled: org/apache/spark/sql/catalyst/SqlParser.start:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.start1 ( ) : scala.util.parsing.combinator.Parsers.Parser<plans.logical.LogicalPlan>
[mangled: org/apache/spark/sql/catalyst/SqlParser.start1:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.structField ( ) : scala.util.parsing.combinator.Parsers.Parser<org.apache.spark.sql.types.StructField>
[mangled: org/apache/spark/sql/catalyst/SqlParser.structField:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.structType ( ) : scala.util.parsing.combinator.Parsers.Parser<org.apache.spark.sql.types.DataType>
[mangled: org/apache/spark/sql/catalyst/SqlParser.structType:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.SUBSTR ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SUBSTR:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.SUBSTRING ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SUBSTRING:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.SUM ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.SUM:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.TABLE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.TABLE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.THEN ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.THEN:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.toDataType ( String dataTypeString ) : org.apache.spark.sql.types.DataType
[mangled: org/apache/spark/sql/catalyst/SqlParser.toDataType:(Ljava/lang/String;)Lorg/apache/spark/sql/types/DataType;]
SqlParser.TRUE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.TRUE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.UNION ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.UNION:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.unsignedNumericLiteral ( ) : scala.util.parsing.combinator.Parsers.Parser<expressions.Literal>
[mangled: org/apache/spark/sql/catalyst/SqlParser.unsignedNumericLiteral:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.UPPER ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.UPPER:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.varchar ( ) : scala.util.parsing.combinator.Parsers.Parser<org.apache.spark.sql.types.DataType>
[mangled: org/apache/spark/sql/catalyst/SqlParser.varchar:()Lscala/util/parsing/combinator/Parsers$Parser;]
SqlParser.WHEN ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.WHEN:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.WHERE ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.WHERE:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
SqlParser.WITH ( ) : AbstractSparkSQLParser.Keyword
[mangled: org/apache/spark/sql/catalyst/SqlParser.WITH:()Lorg/apache/spark/sql/catalyst/AbstractSparkSQLParser$Keyword;]
spark-catalyst_2.10-1.4.0.jar, StructField.class
package org.apache.spark.sql.types
StructField.StructField ( String name, DataType dataType, boolean nullable, Metadata metadata )
[mangled: org/apache/spark/sql/types/StructField."<init>":(Ljava/lang/String;Lorg/apache/spark/sql/types/DataType;ZLorg/apache/spark/sql/types/Metadata;)V]
spark-catalyst_2.10-1.4.0.jar, StructType.class
package org.apache.spark.sql.types
StructType.toAttributes ( ) : scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference>
[mangled: org/apache/spark/sql/types/StructType.toAttributes:()Lscala/collection/Seq;]
spark-catalyst_2.10-1.4.0.jar, UTF8String.class
package org.apache.spark.sql.types
UTF8String.apply ( byte[ ] p1 ) [static] : UTF8String
[mangled: org/apache/spark/sql/types/UTF8String.apply:([B)Lorg/apache/spark/sql/types/UTF8String;]
UTF8String.apply ( String that ) [static] : UTF8String
[mangled: org/apache/spark/sql/types/UTF8String.apply:(Ljava/lang/String;)Lorg/apache/spark/sql/types/UTF8String;]
UTF8String.clone ( ) : Object
[mangled: org/apache/spark/sql/types/UTF8String.clone:()Ljava/lang/Object;]
UTF8String.clone ( ) : UTF8String
[mangled: org/apache/spark/sql/types/UTF8String.clone:()Lorg/apache/spark/sql/types/UTF8String;]
UTF8String.compare ( Object that ) : int
[mangled: org/apache/spark/sql/types/UTF8String.compare:(Ljava/lang/Object;)I]
UTF8String.compare ( UTF8String other ) : int
[mangled: org/apache/spark/sql/types/UTF8String.compare:(Lorg/apache/spark/sql/types/UTF8String;)I]
UTF8String.compareTo ( Object that ) : int
[mangled: org/apache/spark/sql/types/UTF8String.compareTo:(Ljava/lang/Object;)I]
UTF8String.compareTo ( UTF8String other ) : int
[mangled: org/apache/spark/sql/types/UTF8String.compareTo:(Lorg/apache/spark/sql/types/UTF8String;)I]
UTF8String.contains ( UTF8String sub ) : boolean
[mangled: org/apache/spark/sql/types/UTF8String.contains:(Lorg/apache/spark/sql/types/UTF8String;)Z]
UTF8String.endsWith ( UTF8String suffix ) : boolean
[mangled: org/apache/spark/sql/types/UTF8String.endsWith:(Lorg/apache/spark/sql/types/UTF8String;)Z]
UTF8String.equals ( Object other ) : boolean
[mangled: org/apache/spark/sql/types/UTF8String.equals:(Ljava/lang/Object;)Z]
UTF8String.getBytes ( ) : byte[ ]
[mangled: org/apache/spark/sql/types/UTF8String.getBytes:()[B]
UTF8String.hashCode ( ) : int
[mangled: org/apache/spark/sql/types/UTF8String.hashCode:()I]
UTF8String.length ( ) : int
[mangled: org/apache/spark/sql/types/UTF8String.length:()I]
UTF8String.set ( byte[ ] bytes ) : UTF8String
[mangled: org/apache/spark/sql/types/UTF8String.set:([B)Lorg/apache/spark/sql/types/UTF8String;]
UTF8String.set ( String str ) : UTF8String
[mangled: org/apache/spark/sql/types/UTF8String.set:(Ljava/lang/String;)Lorg/apache/spark/sql/types/UTF8String;]
UTF8String.slice ( int start, int until ) : UTF8String
[mangled: org/apache/spark/sql/types/UTF8String.slice:(II)Lorg/apache/spark/sql/types/UTF8String;]
UTF8String.startsWith ( UTF8String prefix ) : boolean
[mangled: org/apache/spark/sql/types/UTF8String.startsWith:(Lorg/apache/spark/sql/types/UTF8String;)Z]
UTF8String.toLowerCase ( ) : UTF8String
[mangled: org/apache/spark/sql/types/UTF8String.toLowerCase:()Lorg/apache/spark/sql/types/UTF8String;]
UTF8String.toString ( ) : String
[mangled: org/apache/spark/sql/types/UTF8String.toString:()Ljava/lang/String;]
UTF8String.toUpperCase ( ) : UTF8String
[mangled: org/apache/spark/sql/types/UTF8String.toUpperCase:()Lorg/apache/spark/sql/types/UTF8String;]
UTF8String.UTF8String ( )
[mangled: org/apache/spark/sql/types/UTF8String."<init>":()V]
spark-core_2.10-1.4.0.jar, JavaSparkContext.class
package org.apache.spark.api.java
JavaSparkContext.accumulable ( T initialValue, String name, org.apache.spark.AccumulableParam<T,R> param ) : org.apache.spark.Accumulable<T,R>
[mangled: org/apache/spark/api/java/JavaSparkContext.accumulable:(Ljava/lang/Object;Ljava/lang/String;Lorg/apache/spark/AccumulableParam;)Lorg/apache/spark/Accumulable;]
JavaSparkContext.accumulator ( double initialValue, String name ) : org.apache.spark.Accumulator<Double>
[mangled: org/apache/spark/api/java/JavaSparkContext.accumulator:(DLjava/lang/String;)Lorg/apache/spark/Accumulator;]
JavaSparkContext.accumulator ( int initialValue, String name ) : org.apache.spark.Accumulator<Integer>
[mangled: org/apache/spark/api/java/JavaSparkContext.accumulator:(ILjava/lang/String;)Lorg/apache/spark/Accumulator;]
JavaSparkContext.accumulator ( T initialValue, String name, org.apache.spark.AccumulatorParam<T> accumulatorParam ) : org.apache.spark.Accumulator<T>
[mangled: org/apache/spark/api/java/JavaSparkContext.accumulator:(Ljava/lang/Object;Ljava/lang/String;Lorg/apache/spark/AccumulatorParam;)Lorg/apache/spark/Accumulator;]
JavaSparkContext.binaryFiles ( String path ) : JavaPairRDD<String,org.apache.spark.input.PortableDataStream>
[mangled: org/apache/spark/api/java/JavaSparkContext.binaryFiles:(Ljava/lang/String;)Lorg/apache/spark/api/java/JavaPairRDD;]
JavaSparkContext.binaryFiles ( String path, int minPartitions ) : JavaPairRDD<String,org.apache.spark.input.PortableDataStream>
[mangled: org/apache/spark/api/java/JavaSparkContext.binaryFiles:(Ljava/lang/String;I)Lorg/apache/spark/api/java/JavaPairRDD;]
JavaSparkContext.binaryRecords ( String path, int recordLength ) : JavaRDD<byte[ ]>
[mangled: org/apache/spark/api/java/JavaSparkContext.binaryRecords:(Ljava/lang/String;I)Lorg/apache/spark/api/java/JavaRDD;]
JavaSparkContext.close ( ) : void
[mangled: org/apache/spark/api/java/JavaSparkContext.close:()V]
JavaSparkContext.doubleAccumulator ( double initialValue, String name ) : org.apache.spark.Accumulator<Double>
[mangled: org/apache/spark/api/java/JavaSparkContext.doubleAccumulator:(DLjava/lang/String;)Lorg/apache/spark/Accumulator;]
JavaSparkContext.emptyRDD ( ) : JavaRDD<T>
[mangled: org/apache/spark/api/java/JavaSparkContext.emptyRDD:()Lorg/apache/spark/api/java/JavaRDD;]
JavaSparkContext.intAccumulator ( int initialValue, String name ) : org.apache.spark.Accumulator<Integer>
[mangled: org/apache/spark/api/java/JavaSparkContext.intAccumulator:(ILjava/lang/String;)Lorg/apache/spark/Accumulator;]
JavaSparkContext.setLogLevel ( String logLevel ) : void
[mangled: org/apache/spark/api/java/JavaSparkContext.setLogLevel:(Ljava/lang/String;)V]
JavaSparkContext.statusTracker ( ) : JavaSparkStatusTracker
[mangled: org/apache/spark/api/java/JavaSparkContext.statusTracker:()Lorg/apache/spark/api/java/JavaSparkStatusTracker;]
JavaSparkContext.version ( ) : String
[mangled: org/apache/spark/api/java/JavaSparkContext.version:()Ljava/lang/String;]
spark-core_2.10-1.4.0.jar, Logging.class
package org.apache.spark
Logging.logName ( ) [abstract] : String
[mangled: org/apache/spark/Logging.logName:()Ljava/lang/String;]
spark-core_2.10-1.4.0.jar, SparkContext.class
package org.apache.spark
SparkContext.accumulable ( R initialValue, String name, AccumulableParam<R,T> param ) : Accumulable<R,T>
[mangled: org/apache/spark/SparkContext.accumulable:(Ljava/lang/Object;Ljava/lang/String;Lorg/apache/spark/AccumulableParam;)Lorg/apache/spark/Accumulable;]
SparkContext.accumulator ( T initialValue, String name, AccumulatorParam<T> param ) : Accumulator<T>
[mangled: org/apache/spark/SparkContext.accumulator:(Ljava/lang/Object;Ljava/lang/String;Lorg/apache/spark/AccumulatorParam;)Lorg/apache/spark/Accumulator;]
SparkContext.addFile ( String path, boolean recursive ) : void
[mangled: org/apache/spark/SparkContext.addFile:(Ljava/lang/String;Z)V]
SparkContext.applicationAttemptId ( ) : scala.Option<String>
[mangled: org/apache/spark/SparkContext.applicationAttemptId:()Lscala/Option;]
SparkContext.applicationId ( ) : String
[mangled: org/apache/spark/SparkContext.applicationId:()Ljava/lang/String;]
SparkContext.binaryFiles ( String path, int minPartitions ) : rdd.RDD<scala.Tuple2<String,input.PortableDataStream>>
[mangled: org/apache/spark/SparkContext.binaryFiles:(Ljava/lang/String;I)Lorg/apache/spark/rdd/RDD;]
SparkContext.binaryRecords ( String path, int recordLength, org.apache.hadoop.conf.Configuration conf ) : rdd.RDD<byte[ ]>
[mangled: org/apache/spark/SparkContext.binaryRecords:(Ljava/lang/String;ILorg/apache/hadoop/conf/Configuration;)Lorg/apache/spark/rdd/RDD;]
SparkContext.clean ( F f, boolean checkSerializable ) : F
[mangled: org/apache/spark/SparkContext.clean:(Ljava/lang/Object;Z)Ljava/lang/Object;]
SparkContext.createSparkEnv ( SparkConf conf, boolean isLocal, scheduler.LiveListenerBus listenerBus ) : SparkEnv
[mangled: org/apache/spark/SparkContext.createSparkEnv:(Lorg/apache/spark/SparkConf;ZLorg/apache/spark/scheduler/LiveListenerBus;)Lorg/apache/spark/SparkEnv;]
SparkContext.eventLogCodec ( ) : scala.Option<String>
[mangled: org/apache/spark/SparkContext.eventLogCodec:()Lscala/Option;]
SparkContext.eventLogDir ( ) : scala.Option<java.net.URI>
[mangled: org/apache/spark/SparkContext.eventLogDir:()Lscala/Option;]
SparkContext.executorAllocationManager ( ) : scala.Option<ExecutorAllocationManager>
[mangled: org/apache/spark/SparkContext.executorAllocationManager:()Lscala/Option;]
SparkContext.externalBlockStoreFolderName ( ) : String
[mangled: org/apache/spark/SparkContext.externalBlockStoreFolderName:()Ljava/lang/String;]
SparkContext.getCallSite ( ) : util.CallSite
[mangled: org/apache/spark/SparkContext.getCallSite:()Lorg/apache/spark/util/CallSite;]
SparkContext.getExecutorThreadDump ( String executorId ) : scala.Option<util.ThreadStackTrace[ ]>
[mangled: org/apache/spark/SparkContext.getExecutorThreadDump:(Ljava/lang/String;)Lscala/Option;]
SparkContext.getOrCreate ( ) [static] : SparkContext
[mangled: org/apache/spark/SparkContext.getOrCreate:()Lorg/apache/spark/SparkContext;]
SparkContext.getOrCreate ( SparkConf p1 ) [static] : SparkContext
[mangled: org/apache/spark/SparkContext.getOrCreate:(Lorg/apache/spark/SparkConf;)Lorg/apache/spark/SparkContext;]
SparkContext.isEventLogEnabled ( ) : boolean
[mangled: org/apache/spark/SparkContext.isEventLogEnabled:()Z]
SparkContext.jobProgressListener ( ) : ui.jobs.JobProgressListener
[mangled: org/apache/spark/SparkContext.jobProgressListener:()Lorg/apache/spark/ui/jobs/JobProgressListener;]
SparkContext.killExecutor ( String executorId ) : boolean
[mangled: org/apache/spark/SparkContext.killExecutor:(Ljava/lang/String;)Z]
SparkContext.killExecutors ( scala.collection.Seq<String> executorIds ) : boolean
[mangled: org/apache/spark/SparkContext.killExecutors:(Lscala/collection/Seq;)Z]
SparkContext.logName ( ) : String
[mangled: org/apache/spark/SparkContext.logName:()Ljava/lang/String;]
SparkContext.metricsSystem ( ) : metrics.MetricsSystem
[mangled: org/apache/spark/SparkContext.metricsSystem:()Lorg/apache/spark/metrics/MetricsSystem;]
SparkContext.SparkContext.._conf ( ) : SparkConf
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext.._conf:()Lorg/apache/spark/SparkConf;]
SparkContext.SparkContext.._env ( ) : SparkEnv
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext.._env:()Lorg/apache/spark/SparkEnv;]
SparkContext.SparkContext..assertNotStopped ( ) : void
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext..assertNotStopped:()V]
SparkContext.SparkContext..creationSite ( ) : util.CallSite
[mangled: org/apache/spark/SparkContext.org.apache.spark.SparkContext..creationSite:()Lorg/apache/spark/util/CallSite;]
SparkContext.progressBar ( ) : scala.Option<ui.ConsoleProgressBar>
[mangled: org/apache/spark/SparkContext.progressBar:()Lscala/Option;]
SparkContext.range ( long start, long end, long step, int numSlices ) : rdd.RDD<Object>
[mangled: org/apache/spark/SparkContext.range:(JJJI)Lorg/apache/spark/rdd/RDD;]
SparkContext.requestExecutors ( int numAdditionalExecutors ) : boolean
[mangled: org/apache/spark/SparkContext.requestExecutors:(I)Z]
SparkContext.requestTotalExecutors ( int numExecutors ) : boolean
[mangled: org/apache/spark/SparkContext.requestTotalExecutors:(I)Z]
SparkContext.schedulerBackend ( ) : scheduler.SchedulerBackend
[mangled: org/apache/spark/SparkContext.schedulerBackend:()Lorg/apache/spark/scheduler/SchedulerBackend;]
SparkContext.schedulerBackend_.eq ( scheduler.SchedulerBackend sb ) : void
[mangled: org/apache/spark/SparkContext.schedulerBackend_.eq:(Lorg/apache/spark/scheduler/SchedulerBackend;)V]
SparkContext.setCallSite ( util.CallSite callSite ) : void
[mangled: org/apache/spark/SparkContext.setCallSite:(Lorg/apache/spark/util/CallSite;)V]
SparkContext.setLogLevel ( String logLevel ) : void
[mangled: org/apache/spark/SparkContext.setLogLevel:(Ljava/lang/String;)V]
SparkContext.statusTracker ( ) : SparkStatusTracker
[mangled: org/apache/spark/SparkContext.statusTracker:()Lorg/apache/spark/SparkStatusTracker;]
SparkContext.supportDynamicAllocation ( ) : boolean
[mangled: org/apache/spark/SparkContext.supportDynamicAllocation:()Z]
SparkContext.ui ( ) : scala.Option<ui.SparkUI>
[mangled: org/apache/spark/SparkContext.ui:()Lscala/Option;]
SparkContext.withScope ( scala.Function0<U> body ) : U
[mangled: org/apache/spark/SparkContext.withScope:(Lscala/Function0;)Ljava/lang/Object;]
spark-core_2.10-1.4.0.jar, SparkHadoopMapReduceUtil.class
package org.apache.spark.mapreduce
SparkHadoopMapReduceUtil.newJobContext ( org.apache.hadoop.conf.Configuration p1, org.apache.hadoop.mapreduce.JobID p2 ) [abstract] : org.apache.hadoop.mapreduce.JobContext
[mangled: org/apache/spark/mapreduce/SparkHadoopMapReduceUtil.newJobContext:(Lorg/apache/hadoop/conf/Configuration;Lorg/apache/hadoop/mapreduce/JobID;)Lorg/apache/hadoop/mapreduce/JobContext;]
SparkHadoopMapReduceUtil.newTaskAttemptContext ( org.apache.hadoop.conf.Configuration p1, org.apache.hadoop.mapreduce.TaskAttemptID p2 ) [abstract] : org.apache.hadoop.mapreduce.TaskAttemptContext
[mangled: org/apache/spark/mapreduce/SparkHadoopMapReduceUtil.newTaskAttemptContext:(Lorg/apache/hadoop/conf/Configuration;Lorg/apache/hadoop/mapreduce/TaskAttemptID;)Lorg/apache/hadoop/mapreduce/TaskAttemptContext;]
SparkHadoopMapReduceUtil.newTaskAttemptID ( String p1, int p2, boolean p3, int p4, int p5 ) [abstract] : org.apache.hadoop.mapreduce.TaskAttemptID
[mangled: org/apache/spark/mapreduce/SparkHadoopMapReduceUtil.newTaskAttemptID:(Ljava/lang/String;IZII)Lorg/apache/hadoop/mapreduce/TaskAttemptID;]
spark-core_2.10-1.4.0.jar, TaskContext.class
package org.apache.spark
TaskContext.addTaskCompletionListener ( util.TaskCompletionListener p1 ) [abstract] : TaskContext
[mangled: org/apache/spark/TaskContext.addTaskCompletionListener:(Lorg/apache/spark/util/TaskCompletionListener;)Lorg/apache/spark/TaskContext;]
TaskContext.addTaskCompletionListener ( scala.Function1<TaskContext,scala.runtime.BoxedUnit> p1 ) [abstract] : TaskContext
[mangled: org/apache/spark/TaskContext.addTaskCompletionListener:(Lscala/Function1;)Lorg/apache/spark/TaskContext;]
TaskContext.attemptNumber ( ) [abstract] : int
[mangled: org/apache/spark/TaskContext.attemptNumber:()I]
TaskContext.get ( ) [static] : TaskContext
[mangled: org/apache/spark/TaskContext.get:()Lorg/apache/spark/TaskContext;]
TaskContext.isCompleted ( ) [abstract] : boolean
[mangled: org/apache/spark/TaskContext.isCompleted:()Z]
TaskContext.isInterrupted ( ) [abstract] : boolean
[mangled: org/apache/spark/TaskContext.isInterrupted:()Z]
TaskContext.isRunningLocally ( ) [abstract] : boolean
[mangled: org/apache/spark/TaskContext.isRunningLocally:()Z]
TaskContext.taskAttemptId ( ) [abstract] : long
[mangled: org/apache/spark/TaskContext.taskAttemptId:()J]
TaskContext.TaskContext ( )
[mangled: org/apache/spark/TaskContext."<init>":()V]
TaskContext.taskMemoryManager ( ) [abstract] : unsafe.memory.TaskMemoryManager
[mangled: org/apache/spark/TaskContext.taskMemoryManager:()Lorg/apache/spark/unsafe/memory/TaskMemoryManager;]
spark-core_2.10-1.4.0.jar, TaskContextImpl.class
package org.apache.spark
TaskContextImpl.attemptNumber ( ) : int
[mangled: org/apache/spark/TaskContextImpl.attemptNumber:()I]
TaskContextImpl.partitionId ( ) : int
[mangled: org/apache/spark/TaskContextImpl.partitionId:()I]
TaskContextImpl.stageId ( ) : int
[mangled: org/apache/spark/TaskContextImpl.stageId:()I]
TaskContextImpl.taskAttemptId ( ) : long
[mangled: org/apache/spark/TaskContextImpl.taskAttemptId:()J]
TaskContextImpl.TaskContextImpl ( int stageId, int partitionId, long taskAttemptId, int attemptNumber, unsafe.memory.TaskMemoryManager taskMemoryManager, boolean runningLocally, executor.TaskMetrics taskMetrics )
[mangled: org/apache/spark/TaskContextImpl."<init>":(IIJILorg/apache/spark/unsafe/memory/TaskMemoryManager;ZLorg/apache/spark/executor/TaskMetrics;)V]
spark-sql_2.10-1.4.0.jar, Aggregate.class
package org.apache.spark.sql.execution
Aggregate.Aggregate ( boolean partial, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupingExpressions, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> aggregateExpressions, SparkPlan child )
[mangled: org/apache/spark/sql/execution/Aggregate."<init>":(ZLscala/collection/Seq;Lscala/collection/Seq;Lorg/apache/spark/sql/execution/SparkPlan;)V]
spark-sql_2.10-1.4.0.jar, BaseRelation.class
package org.apache.spark.sql.sources
BaseRelation.BaseRelation ( )
[mangled: org/apache/spark/sql/sources/BaseRelation."<init>":()V]
BaseRelation.needConversion ( ) : boolean
[mangled: org/apache/spark/sql/sources/BaseRelation.needConversion:()Z]
BaseRelation.schema ( ) [abstract] : org.apache.spark.sql.types.StructType
[mangled: org/apache/spark/sql/sources/BaseRelation.schema:()Lorg/apache/spark/sql/types/StructType;]
BaseRelation.sizeInBytes ( ) : long
[mangled: org/apache/spark/sql/sources/BaseRelation.sizeInBytes:()J]
BaseRelation.sqlContext ( ) [abstract] : org.apache.spark.sql.SQLContext
[mangled: org/apache/spark/sql/sources/BaseRelation.sqlContext:()Lorg/apache/spark/sql/SQLContext;]
spark-sql_2.10-1.4.0.jar, CatalystScan.class
package org.apache.spark.sql.sources
CatalystScan.buildScan ( scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> p1, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> p2 ) [abstract] : org.apache.spark.rdd.RDD<org.apache.spark.sql.Row>
[mangled: org/apache/spark/sql/sources/CatalystScan.buildScan:(Lscala/collection/Seq;Lscala/collection/Seq;)Lorg/apache/spark/rdd/RDD;]
spark-sql_2.10-1.4.0.jar, DataFrame.class
package org.apache.spark.sql
DataFrame.agg ( java.util.Map<String,String> exprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.agg:(Ljava/util/Map;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.agg ( Column expr, Column... exprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.agg:(Lorg/apache/spark/sql/Column;[Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.agg ( Column expr, scala.collection.Seq<Column> exprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.agg:(Lorg/apache/spark/sql/Column;Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.agg ( scala.collection.immutable.Map<String,String> exprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.agg:(Lscala/collection/immutable/Map;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.agg ( scala.Tuple2<String,String> aggExpr, scala.collection.Seq<scala.Tuple2<String,String>> aggExprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.agg:(Lscala/Tuple2;Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.apply ( String colName ) : Column
[mangled: org/apache/spark/sql/DataFrame.apply:(Ljava/lang/String;)Lorg/apache/spark/sql/Column;]
DataFrame.as ( scala.Symbol alias ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.as:(Lscala/Symbol;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.as ( String alias ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.as:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.cache ( ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.cache:()Lorg/apache/spark/sql/DataFrame;]
DataFrame.cache ( ) : RDDApi
[mangled: org/apache/spark/sql/DataFrame.cache:()Lorg/apache/spark/sql/RDDApi;]
DataFrame.coalesce ( int numPartitions ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.coalesce:(I)Lorg/apache/spark/sql/DataFrame;]
DataFrame.col ( String colName ) : Column
[mangled: org/apache/spark/sql/DataFrame.col:(Ljava/lang/String;)Lorg/apache/spark/sql/Column;]
DataFrame.collect ( ) : Object
[mangled: org/apache/spark/sql/DataFrame.collect:()Ljava/lang/Object;]
DataFrame.collect ( ) : Row[ ]
[mangled: org/apache/spark/sql/DataFrame.collect:()[Lorg/apache/spark/sql/Row;]
DataFrame.collectAsList ( ) : java.util.List<Row>
[mangled: org/apache/spark/sql/DataFrame.collectAsList:()Ljava/util/List;]
DataFrame.columns ( ) : String[ ]
[mangled: org/apache/spark/sql/DataFrame.columns:()[Ljava/lang/String;]
DataFrame.count ( ) : long
[mangled: org/apache/spark/sql/DataFrame.count:()J]
DataFrame.cube ( Column... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.cube:([Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.cube ( scala.collection.Seq<Column> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.cube:(Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.cube ( String col1, scala.collection.Seq<String> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.cube:(Ljava/lang/String;Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.cube ( String col1, String... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.cube:(Ljava/lang/String;[Ljava/lang/String;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.DataFrame ( SQLContext sqlContext, catalyst.plans.logical.LogicalPlan logicalPlan )
[mangled: org/apache/spark/sql/DataFrame."<init>":(Lorg/apache/spark/sql/SQLContext;Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)V]
DataFrame.DataFrame ( SQLContext sqlContext, SQLContext.QueryExecution queryExecution )
[mangled: org/apache/spark/sql/DataFrame."<init>":(Lorg/apache/spark/sql/SQLContext;Lorg/apache/spark/sql/SQLContext$QueryExecution;)V]
DataFrame.describe ( scala.collection.Seq<String> cols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.describe:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.describe ( String... cols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.describe:([Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.distinct ( ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.distinct:()Lorg/apache/spark/sql/DataFrame;]
DataFrame.drop ( String colName ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.drop:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.dropDuplicates ( ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.dropDuplicates:()Lorg/apache/spark/sql/DataFrame;]
DataFrame.dropDuplicates ( scala.collection.Seq<String> colNames ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.dropDuplicates:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.dropDuplicates ( String[ ] colNames ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.dropDuplicates:([Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.dtypes ( ) : scala.Tuple2<String,String>[ ]
[mangled: org/apache/spark/sql/DataFrame.dtypes:()[Lscala/Tuple2;]
DataFrame.except ( DataFrame other ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.except:(Lorg/apache/spark/sql/DataFrame;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.explain ( ) : void
[mangled: org/apache/spark/sql/DataFrame.explain:()V]
DataFrame.explain ( boolean extended ) : void
[mangled: org/apache/spark/sql/DataFrame.explain:(Z)V]
DataFrame.explode ( scala.collection.Seq<Column> input, scala.Function1<Row,scala.collection.TraversableOnce<A>> f, scala.reflect.api.TypeTags.TypeTag<A> p3 ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.explode:(Lscala/collection/Seq;Lscala/Function1;Lscala/reflect/api/TypeTags$TypeTag;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.explode ( String inputColumn, String outputColumn, scala.Function1<A,scala.collection.TraversableOnce<B>> f, scala.reflect.api.TypeTags.TypeTag<B> p4 ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.explode:(Ljava/lang/String;Ljava/lang/String;Lscala/Function1;Lscala/reflect/api/TypeTags$TypeTag;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.filter ( Column condition ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.filter:(Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.filter ( String conditionExpr ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.filter:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.first ( ) : Object
[mangled: org/apache/spark/sql/DataFrame.first:()Ljava/lang/Object;]
DataFrame.first ( ) : Row
[mangled: org/apache/spark/sql/DataFrame.first:()Lorg/apache/spark/sql/Row;]
DataFrame.flatMap ( scala.Function1<Row,scala.collection.TraversableOnce<R>> f, scala.reflect.ClassTag<R> p2 ) : org.apache.spark.rdd.RDD<R>
[mangled: org/apache/spark/sql/DataFrame.flatMap:(Lscala/Function1;Lscala/reflect/ClassTag;)Lorg/apache/spark/rdd/RDD;]
DataFrame.foreach ( scala.Function1<Row,scala.runtime.BoxedUnit> f ) : void
[mangled: org/apache/spark/sql/DataFrame.foreach:(Lscala/Function1;)V]
DataFrame.foreachPartition ( scala.Function1<scala.collection.Iterator<Row>,scala.runtime.BoxedUnit> f ) : void
[mangled: org/apache/spark/sql/DataFrame.foreachPartition:(Lscala/Function1;)V]
DataFrame.groupBy ( Column... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.groupBy:([Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.groupBy ( scala.collection.Seq<Column> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.groupBy:(Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.groupBy ( String col1, scala.collection.Seq<String> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.groupBy:(Ljava/lang/String;Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.groupBy ( String col1, String... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.groupBy:(Ljava/lang/String;[Ljava/lang/String;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.head ( ) : Row
[mangled: org/apache/spark/sql/DataFrame.head:()Lorg/apache/spark/sql/Row;]
DataFrame.head ( int n ) : Row[ ]
[mangled: org/apache/spark/sql/DataFrame.head:(I)[Lorg/apache/spark/sql/Row;]
DataFrame.intersect ( DataFrame other ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.intersect:(Lorg/apache/spark/sql/DataFrame;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.isLocal ( ) : boolean
[mangled: org/apache/spark/sql/DataFrame.isLocal:()Z]
DataFrame.javaRDD ( ) : org.apache.spark.api.java.JavaRDD<Row>
[mangled: org/apache/spark/sql/DataFrame.javaRDD:()Lorg/apache/spark/api/java/JavaRDD;]
DataFrame.javaToPython ( ) : org.apache.spark.api.java.JavaRDD<byte[ ]>
[mangled: org/apache/spark/sql/DataFrame.javaToPython:()Lorg/apache/spark/api/java/JavaRDD;]
DataFrame.join ( DataFrame right ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.join:(Lorg/apache/spark/sql/DataFrame;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.join ( DataFrame right, Column joinExprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.join:(Lorg/apache/spark/sql/DataFrame;Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.join ( DataFrame right, Column joinExprs, String joinType ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.join:(Lorg/apache/spark/sql/DataFrame;Lorg/apache/spark/sql/Column;Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.join ( DataFrame right, String usingColumn ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.join:(Lorg/apache/spark/sql/DataFrame;Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.limit ( int n ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.limit:(I)Lorg/apache/spark/sql/DataFrame;]
DataFrame.logicalPlan ( ) : catalyst.plans.logical.LogicalPlan
[mangled: org/apache/spark/sql/DataFrame.logicalPlan:()Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;]
DataFrame.map ( scala.Function1<Row,R> f, scala.reflect.ClassTag<R> p2 ) : org.apache.spark.rdd.RDD<R>
[mangled: org/apache/spark/sql/DataFrame.map:(Lscala/Function1;Lscala/reflect/ClassTag;)Lorg/apache/spark/rdd/RDD;]
DataFrame.mapPartitions ( scala.Function1<scala.collection.Iterator<Row>,scala.collection.Iterator<R>> f, scala.reflect.ClassTag<R> p2 ) : org.apache.spark.rdd.RDD<R>
[mangled: org/apache/spark/sql/DataFrame.mapPartitions:(Lscala/Function1;Lscala/reflect/ClassTag;)Lorg/apache/spark/rdd/RDD;]
DataFrame.na ( ) : DataFrameNaFunctions
[mangled: org/apache/spark/sql/DataFrame.na:()Lorg/apache/spark/sql/DataFrameNaFunctions;]
DataFrame.numericColumns ( ) : scala.collection.Seq<catalyst.expressions.Expression>
[mangled: org/apache/spark/sql/DataFrame.numericColumns:()Lscala/collection/Seq;]
DataFrame.orderBy ( Column... sortExprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.orderBy:([Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.orderBy ( scala.collection.Seq<Column> sortExprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.orderBy:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.orderBy ( String sortCol, scala.collection.Seq<String> sortCols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.orderBy:(Ljava/lang/String;Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.orderBy ( String sortCol, String... sortCols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.orderBy:(Ljava/lang/String;[Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.DataFrame..logicalPlanToDataFrame ( catalyst.plans.logical.LogicalPlan logicalPlan ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.org.apache.spark.sql.DataFrame..logicalPlanToDataFrame:(Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.persist ( ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.persist:()Lorg/apache/spark/sql/DataFrame;]
DataFrame.persist ( ) : RDDApi
[mangled: org/apache/spark/sql/DataFrame.persist:()Lorg/apache/spark/sql/RDDApi;]
DataFrame.persist ( org.apache.spark.storage.StorageLevel newLevel ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.persist:(Lorg/apache/spark/storage/StorageLevel;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.persist ( org.apache.spark.storage.StorageLevel newLevel ) : RDDApi
[mangled: org/apache/spark/sql/DataFrame.persist:(Lorg/apache/spark/storage/StorageLevel;)Lorg/apache/spark/sql/RDDApi;]
DataFrame.printSchema ( ) : void
[mangled: org/apache/spark/sql/DataFrame.printSchema:()V]
DataFrame.queryExecution ( ) : SQLContext.QueryExecution
[mangled: org/apache/spark/sql/DataFrame.queryExecution:()Lorg/apache/spark/sql/SQLContext$QueryExecution;]
DataFrame.randomSplit ( double[ ] weights ) : DataFrame[ ]
[mangled: org/apache/spark/sql/DataFrame.randomSplit:([D)[Lorg/apache/spark/sql/DataFrame;]
DataFrame.randomSplit ( double[ ] weights, long seed ) : DataFrame[ ]
[mangled: org/apache/spark/sql/DataFrame.randomSplit:([DJ)[Lorg/apache/spark/sql/DataFrame;]
DataFrame.randomSplit ( scala.collection.immutable.List<Object> weights, long seed ) : DataFrame[ ]
[mangled: org/apache/spark/sql/DataFrame.randomSplit:(Lscala/collection/immutable/List;J)[Lorg/apache/spark/sql/DataFrame;]
DataFrame.rdd ( ) : org.apache.spark.rdd.RDD<Row>
[mangled: org/apache/spark/sql/DataFrame.rdd:()Lorg/apache/spark/rdd/RDD;]
DataFrame.registerTempTable ( String tableName ) : void
[mangled: org/apache/spark/sql/DataFrame.registerTempTable:(Ljava/lang/String;)V]
DataFrame.repartition ( int numPartitions ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.repartition:(I)Lorg/apache/spark/sql/DataFrame;]
DataFrame.resolve ( String colName ) : catalyst.expressions.NamedExpression
[mangled: org/apache/spark/sql/DataFrame.resolve:(Ljava/lang/String;)Lorg/apache/spark/sql/catalyst/expressions/NamedExpression;]
DataFrame.rollup ( Column... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.rollup:([Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.rollup ( scala.collection.Seq<Column> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.rollup:(Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.rollup ( String col1, scala.collection.Seq<String> cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.rollup:(Ljava/lang/String;Lscala/collection/Seq;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.rollup ( String col1, String... cols ) : GroupedData
[mangled: org/apache/spark/sql/DataFrame.rollup:(Ljava/lang/String;[Ljava/lang/String;)Lorg/apache/spark/sql/GroupedData;]
DataFrame.sample ( boolean withReplacement, double fraction ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.sample:(ZD)Lorg/apache/spark/sql/DataFrame;]
DataFrame.sample ( boolean withReplacement, double fraction, long seed ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.sample:(ZDJ)Lorg/apache/spark/sql/DataFrame;]
DataFrame.schema ( ) : types.StructType
[mangled: org/apache/spark/sql/DataFrame.schema:()Lorg/apache/spark/sql/types/StructType;]
DataFrame.select ( Column... cols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.select:([Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.select ( scala.collection.Seq<Column> cols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.select:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.select ( String col, scala.collection.Seq<String> cols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.select:(Ljava/lang/String;Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.select ( String col, String... cols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.select:(Ljava/lang/String;[Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.selectExpr ( scala.collection.Seq<String> exprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.selectExpr:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.selectExpr ( String... exprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.selectExpr:([Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.show ( ) : void
[mangled: org/apache/spark/sql/DataFrame.show:()V]
DataFrame.show ( int numRows ) : void
[mangled: org/apache/spark/sql/DataFrame.show:(I)V]
DataFrame.showString ( int numRows ) : String
[mangled: org/apache/spark/sql/DataFrame.showString:(I)Ljava/lang/String;]
DataFrame.sort ( Column... sortExprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.sort:([Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.sort ( scala.collection.Seq<Column> sortExprs ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.sort:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.sort ( String sortCol, scala.collection.Seq<String> sortCols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.sort:(Ljava/lang/String;Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.sort ( String sortCol, String... sortCols ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.sort:(Ljava/lang/String;[Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.sqlContext ( ) : SQLContext
[mangled: org/apache/spark/sql/DataFrame.sqlContext:()Lorg/apache/spark/sql/SQLContext;]
DataFrame.stat ( ) : DataFrameStatFunctions
[mangled: org/apache/spark/sql/DataFrame.stat:()Lorg/apache/spark/sql/DataFrameStatFunctions;]
DataFrame.take ( int n ) : Object
[mangled: org/apache/spark/sql/DataFrame.take:(I)Ljava/lang/Object;]
DataFrame.take ( int n ) : Row[ ]
[mangled: org/apache/spark/sql/DataFrame.take:(I)[Lorg/apache/spark/sql/Row;]
DataFrame.toDF ( ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.toDF:()Lorg/apache/spark/sql/DataFrame;]
DataFrame.toDF ( scala.collection.Seq<String> colNames ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.toDF:(Lscala/collection/Seq;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.toDF ( String... colNames ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.toDF:([Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.toJavaRDD ( ) : org.apache.spark.api.java.JavaRDD<Row>
[mangled: org/apache/spark/sql/DataFrame.toJavaRDD:()Lorg/apache/spark/api/java/JavaRDD;]
DataFrame.toJSON ( ) : org.apache.spark.rdd.RDD<String>
[mangled: org/apache/spark/sql/DataFrame.toJSON:()Lorg/apache/spark/rdd/RDD;]
DataFrame.toString ( ) : String
[mangled: org/apache/spark/sql/DataFrame.toString:()Ljava/lang/String;]
DataFrame.unionAll ( DataFrame other ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.unionAll:(Lorg/apache/spark/sql/DataFrame;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.unpersist ( ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.unpersist:()Lorg/apache/spark/sql/DataFrame;]
DataFrame.unpersist ( ) : RDDApi
[mangled: org/apache/spark/sql/DataFrame.unpersist:()Lorg/apache/spark/sql/RDDApi;]
DataFrame.unpersist ( boolean blocking ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.unpersist:(Z)Lorg/apache/spark/sql/DataFrame;]
DataFrame.unpersist ( boolean blocking ) : RDDApi
[mangled: org/apache/spark/sql/DataFrame.unpersist:(Z)Lorg/apache/spark/sql/RDDApi;]
DataFrame.where ( Column condition ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.where:(Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.withColumn ( String colName, Column col ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.withColumn:(Ljava/lang/String;Lorg/apache/spark/sql/Column;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.withColumnRenamed ( String existingName, String newName ) : DataFrame
[mangled: org/apache/spark/sql/DataFrame.withColumnRenamed:(Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
DataFrame.write ( ) : DataFrameWriter
[mangled: org/apache/spark/sql/DataFrame.write:()Lorg/apache/spark/sql/DataFrameWriter;]
spark-sql_2.10-1.4.0.jar, GeneratedAggregate.class
package org.apache.spark.sql.execution
GeneratedAggregate.GeneratedAggregate ( boolean partial, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupingExpressions, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> aggregateExpressions, boolean unsafeEnabled, SparkPlan child )
[mangled: org/apache/spark/sql/execution/GeneratedAggregate."<init>":(ZLscala/collection/Seq;Lscala/collection/Seq;ZLorg/apache/spark/sql/execution/SparkPlan;)V]
spark-sql_2.10-1.4.0.jar, InsertableRelation.class
package org.apache.spark.sql.sources
InsertableRelation.insert ( org.apache.spark.sql.DataFrame p1, boolean p2 ) [abstract] : void
[mangled: org/apache/spark/sql/sources/InsertableRelation.insert:(Lorg/apache/spark/sql/DataFrame;Z)V]
spark-sql_2.10-1.4.0.jar, LogicalRelation.class
package org.apache.spark.sql.sources
LogicalRelation.andThen ( scala.Function1<LogicalRelation,A> p1 ) [static] : scala.Function1<BaseRelation,A>
[mangled: org/apache/spark/sql/sources/LogicalRelation.andThen:(Lscala/Function1;)Lscala/Function1;]
LogicalRelation.attributeMap ( ) : org.apache.spark.sql.catalyst.expressions.AttributeMap<org.apache.spark.sql.catalyst.expressions.AttributeReference>
[mangled: org/apache/spark/sql/sources/LogicalRelation.attributeMap:()Lorg/apache/spark/sql/catalyst/expressions/AttributeMap;]
LogicalRelation.canEqual ( Object p1 ) : boolean
[mangled: org/apache/spark/sql/sources/LogicalRelation.canEqual:(Ljava/lang/Object;)Z]
LogicalRelation.compose ( scala.Function1<A,BaseRelation> p1 ) [static] : scala.Function1<A,LogicalRelation>
[mangled: org/apache/spark/sql/sources/LogicalRelation.compose:(Lscala/Function1;)Lscala/Function1;]
LogicalRelation.copy ( BaseRelation relation ) : LogicalRelation
[mangled: org/apache/spark/sql/sources/LogicalRelation.copy:(Lorg/apache/spark/sql/sources/BaseRelation;)Lorg/apache/spark/sql/sources/LogicalRelation;]
LogicalRelation.equals ( Object other ) : boolean
[mangled: org/apache/spark/sql/sources/LogicalRelation.equals:(Ljava/lang/Object;)Z]
LogicalRelation.hashCode ( ) : int
[mangled: org/apache/spark/sql/sources/LogicalRelation.hashCode:()I]
LogicalRelation.LogicalRelation ( BaseRelation relation )
[mangled: org/apache/spark/sql/sources/LogicalRelation."<init>":(Lorg/apache/spark/sql/sources/BaseRelation;)V]
LogicalRelation.newInstance ( ) : org.apache.spark.sql.catalyst.plans.logical.LogicalPlan
[mangled: org/apache/spark/sql/sources/LogicalRelation.newInstance:()Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;]
LogicalRelation.newInstance ( ) : LogicalRelation
[mangled: org/apache/spark/sql/sources/LogicalRelation.newInstance:()Lorg/apache/spark/sql/sources/LogicalRelation;]
LogicalRelation.output ( ) : scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference>
[mangled: org/apache/spark/sql/sources/LogicalRelation.output:()Lscala/collection/Seq;]
LogicalRelation.productArity ( ) : int
[mangled: org/apache/spark/sql/sources/LogicalRelation.productArity:()I]
LogicalRelation.productElement ( int p1 ) : Object
[mangled: org/apache/spark/sql/sources/LogicalRelation.productElement:(I)Ljava/lang/Object;]
LogicalRelation.productIterator ( ) : scala.collection.Iterator<Object>
[mangled: org/apache/spark/sql/sources/LogicalRelation.productIterator:()Lscala/collection/Iterator;]
LogicalRelation.productPrefix ( ) : String
[mangled: org/apache/spark/sql/sources/LogicalRelation.productPrefix:()Ljava/lang/String;]
LogicalRelation.relation ( ) : BaseRelation
[mangled: org/apache/spark/sql/sources/LogicalRelation.relation:()Lorg/apache/spark/sql/sources/BaseRelation;]
LogicalRelation.sameResult ( org.apache.spark.sql.catalyst.plans.logical.LogicalPlan otherPlan ) : boolean
[mangled: org/apache/spark/sql/sources/LogicalRelation.sameResult:(Lorg/apache/spark/sql/catalyst/plans/logical/LogicalPlan;)Z]
LogicalRelation.simpleString ( ) : String
[mangled: org/apache/spark/sql/sources/LogicalRelation.simpleString:()Ljava/lang/String;]
LogicalRelation.statistics ( ) : org.apache.spark.sql.catalyst.plans.logical.Statistics
[mangled: org/apache/spark/sql/sources/LogicalRelation.statistics:()Lorg/apache/spark/sql/catalyst/plans/logical/Statistics;]
spark-sql_2.10-1.4.0.jar, RelationProvider.class
package org.apache.spark.sql.sources
RelationProvider.createRelation ( org.apache.spark.sql.SQLContext p1, scala.collection.immutable.Map<String,String> p2 ) [abstract] : BaseRelation
[mangled: org/apache/spark/sql/sources/RelationProvider.createRelation:(Lorg/apache/spark/sql/SQLContext;Lscala/collection/immutable/Map;)Lorg/apache/spark/sql/sources/BaseRelation;]
spark-sql_2.10-1.4.0.jar, RunnableCommand.class
package org.apache.spark.sql.execution
RunnableCommand.children ( ) [abstract] : scala.collection.Seq<org.apache.spark.sql.catalyst.plans.logical.LogicalPlan>
[mangled: org/apache/spark/sql/execution/RunnableCommand.children:()Lscala/collection/Seq;]
RunnableCommand.output ( ) [abstract] : scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute>
[mangled: org/apache/spark/sql/execution/RunnableCommand.output:()Lscala/collection/Seq;]
RunnableCommand.run ( org.apache.spark.sql.SQLContext p1 ) [abstract] : scala.collection.Seq<org.apache.spark.sql.Row>
[mangled: org/apache/spark/sql/execution/RunnableCommand.run:(Lorg/apache/spark/sql/SQLContext;)Lscala/collection/Seq;]
spark-sql_2.10-1.4.0.jar, SparkPlan.class
package org.apache.spark.sql.execution
SparkPlan.codegenEnabled ( ) : boolean
[mangled: org/apache/spark/sql/execution/SparkPlan.codegenEnabled:()Z]
SparkPlan.doExecute ( ) [abstract] : org.apache.spark.rdd.RDD<org.apache.spark.sql.Row>
[mangled: org/apache/spark/sql/execution/SparkPlan.doExecute:()Lorg/apache/spark/rdd/RDD;]
SparkPlan.executeCollect ( ) : org.apache.spark.sql.Row[ ]
[mangled: org/apache/spark/sql/execution/SparkPlan.executeCollect:()[Lorg/apache/spark/sql/Row;]
SparkPlan.executeTake ( int n ) : org.apache.spark.sql.Row[ ]
[mangled: org/apache/spark/sql/execution/SparkPlan.executeTake:(I)[Lorg/apache/spark/sql/Row;]
SparkPlan.isTraceEnabled ( ) : boolean
[mangled: org/apache/spark/sql/execution/SparkPlan.isTraceEnabled:()Z]
SparkPlan.log ( ) : org.slf4j.Logger
[mangled: org/apache/spark/sql/execution/SparkPlan.log:()Lorg/slf4j/Logger;]
SparkPlan.logDebug ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logDebug:(Lscala/Function0;)V]
SparkPlan.logDebug ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logDebug:(Lscala/Function0;Ljava/lang/Throwable;)V]
SparkPlan.logError ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logError:(Lscala/Function0;)V]
SparkPlan.logError ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logError:(Lscala/Function0;Ljava/lang/Throwable;)V]
SparkPlan.logInfo ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logInfo:(Lscala/Function0;)V]
SparkPlan.logInfo ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logInfo:(Lscala/Function0;Ljava/lang/Throwable;)V]
SparkPlan.logName ( ) : String
[mangled: org/apache/spark/sql/execution/SparkPlan.logName:()Ljava/lang/String;]
SparkPlan.logTrace ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logTrace:(Lscala/Function0;)V]
SparkPlan.logTrace ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logTrace:(Lscala/Function0;Ljava/lang/Throwable;)V]
SparkPlan.logWarning ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logWarning:(Lscala/Function0;)V]
SparkPlan.logWarning ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.logWarning:(Lscala/Function0;Ljava/lang/Throwable;)V]
SparkPlan.makeCopy ( Object[ ] newArgs ) : org.apache.spark.sql.catalyst.trees.TreeNode
[mangled: org/apache/spark/sql/execution/SparkPlan.makeCopy:([Ljava/lang/Object;)Lorg/apache/spark/sql/catalyst/trees/TreeNode;]
SparkPlan.makeCopy ( Object[ ] newArgs ) : SparkPlan
[mangled: org/apache/spark/sql/execution/SparkPlan.makeCopy:([Ljava/lang/Object;)Lorg/apache/spark/sql/execution/SparkPlan;]
SparkPlan.newMutableProjection ( scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> inputSchema ) : scala.Function0<org.apache.spark.sql.catalyst.expressions.package.MutableProjection>
[mangled: org/apache/spark/sql/execution/SparkPlan.newMutableProjection:(Lscala/collection/Seq;Lscala/collection/Seq;)Lscala/Function0;]
SparkPlan.newOrdering ( scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.SortOrder> order, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> inputSchema ) : scala.math.Ordering<org.apache.spark.sql.Row>
[mangled: org/apache/spark/sql/execution/SparkPlan.newOrdering:(Lscala/collection/Seq;Lscala/collection/Seq;)Lscala/math/Ordering;]
SparkPlan.newPredicate ( org.apache.spark.sql.catalyst.expressions.Expression expression, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> inputSchema ) : scala.Function1<org.apache.spark.sql.Row,Object>
[mangled: org/apache/spark/sql/execution/SparkPlan.newPredicate:(Lorg/apache/spark/sql/catalyst/expressions/Expression;Lscala/collection/Seq;)Lscala/Function1;]
SparkPlan.newProjection ( scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> inputSchema ) : org.apache.spark.sql.catalyst.expressions.package.Projection
[mangled: org/apache/spark/sql/execution/SparkPlan.newProjection:(Lscala/collection/Seq;Lscala/collection/Seq;)Lorg/apache/spark/sql/catalyst/expressions/package$Projection;]
SparkPlan.org.apache.spark.Logging..log_ ( ) : org.slf4j.Logger
[mangled: org/apache/spark/sql/execution/SparkPlan.org.apache.spark.Logging..log_:()Lorg/slf4j/Logger;]
SparkPlan.org.apache.spark.Logging..log__.eq ( org.slf4j.Logger p1 ) : void
[mangled: org/apache/spark/sql/execution/SparkPlan.org.apache.spark.Logging..log__.eq:(Lorg/slf4j/Logger;)V]
SparkPlan.outputOrdering ( ) : scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.SortOrder>
[mangled: org/apache/spark/sql/execution/SparkPlan.outputOrdering:()Lscala/collection/Seq;]
SparkPlan.requiredChildOrdering ( ) : scala.collection.Seq<scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.SortOrder>>
[mangled: org/apache/spark/sql/execution/SparkPlan.requiredChildOrdering:()Lscala/collection/Seq;]
SparkPlan.sparkContext ( ) : org.apache.spark.SparkContext
[mangled: org/apache/spark/sql/execution/SparkPlan.sparkContext:()Lorg/apache/spark/SparkContext;]
SparkPlan.sqlContext ( ) : org.apache.spark.sql.SQLContext
[mangled: org/apache/spark/sql/execution/SparkPlan.sqlContext:()Lorg/apache/spark/sql/SQLContext;]
spark-sql_2.10-1.4.0.jar, SQLConf.class
package org.apache.spark.sql
SQLConf.AUTO_BROADCASTJOIN_THRESHOLD ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.AUTO_BROADCASTJOIN_THRESHOLD:()Ljava/lang/String;]
SQLConf.autoBroadcastJoinThreshold ( ) : int
[mangled: org/apache/spark/sql/SQLConf.autoBroadcastJoinThreshold:()I]
SQLConf.BROADCAST_TIMEOUT ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.BROADCAST_TIMEOUT:()Ljava/lang/String;]
SQLConf.broadcastTimeout ( ) : int
[mangled: org/apache/spark/sql/SQLConf.broadcastTimeout:()I]
SQLConf.CASE_SENSITIVE ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.CASE_SENSITIVE:()Ljava/lang/String;]
SQLConf.caseSensitiveAnalysis ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.caseSensitiveAnalysis:()Z]
SQLConf.clear ( ) : void
[mangled: org/apache/spark/sql/SQLConf.clear:()V]
SQLConf.CODEGEN_ENABLED ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.CODEGEN_ENABLED:()Ljava/lang/String;]
SQLConf.codegenEnabled ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.codegenEnabled:()Z]
SQLConf.COLUMN_BATCH_SIZE ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.COLUMN_BATCH_SIZE:()Ljava/lang/String;]
SQLConf.COLUMN_NAME_OF_CORRUPT_RECORD ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.COLUMN_NAME_OF_CORRUPT_RECORD:()Ljava/lang/String;]
SQLConf.columnBatchSize ( ) : int
[mangled: org/apache/spark/sql/SQLConf.columnBatchSize:()I]
SQLConf.columnNameOfCorruptRecord ( ) : String
[mangled: org/apache/spark/sql/SQLConf.columnNameOfCorruptRecord:()Ljava/lang/String;]
SQLConf.COMPRESS_CACHED ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.COMPRESS_CACHED:()Ljava/lang/String;]
SQLConf.DATAFRAME_EAGER_ANALYSIS ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.DATAFRAME_EAGER_ANALYSIS:()Ljava/lang/String;]
SQLConf.DATAFRAME_RETAIN_GROUP_COLUMNS ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.DATAFRAME_RETAIN_GROUP_COLUMNS:()Ljava/lang/String;]
SQLConf.DATAFRAME_SELF_JOIN_AUTO_RESOLVE_AMBIGUITY ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.DATAFRAME_SELF_JOIN_AUTO_RESOLVE_AMBIGUITY:()Ljava/lang/String;]
SQLConf.dataFrameEagerAnalysis ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.dataFrameEagerAnalysis:()Z]
SQLConf.dataFrameRetainGroupColumns ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.dataFrameRetainGroupColumns:()Z]
SQLConf.dataFrameSelfJoinAutoResolveAmbiguity ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.dataFrameSelfJoinAutoResolveAmbiguity:()Z]
SQLConf.DEFAULT_DATA_SOURCE_NAME ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.DEFAULT_DATA_SOURCE_NAME:()Ljava/lang/String;]
SQLConf.DEFAULT_SIZE_IN_BYTES ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.DEFAULT_SIZE_IN_BYTES:()Ljava/lang/String;]
SQLConf.defaultDataSourceName ( ) : String
[mangled: org/apache/spark/sql/SQLConf.defaultDataSourceName:()Ljava/lang/String;]
SQLConf.defaultSizeInBytes ( ) : long
[mangled: org/apache/spark/sql/SQLConf.defaultSizeInBytes:()J]
SQLConf.dialect ( ) : String
[mangled: org/apache/spark/sql/SQLConf.dialect:()Ljava/lang/String;]
SQLConf.DIALECT ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.DIALECT:()Ljava/lang/String;]
SQLConf.EXTERNAL_SORT ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.EXTERNAL_SORT:()Ljava/lang/String;]
SQLConf.externalSortEnabled ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.externalSortEnabled:()Z]
SQLConf.getAllConfs ( ) : scala.collection.immutable.Map<String,String>
[mangled: org/apache/spark/sql/SQLConf.getAllConfs:()Lscala/collection/immutable/Map;]
SQLConf.getConf ( String key ) : String
[mangled: org/apache/spark/sql/SQLConf.getConf:(Ljava/lang/String;)Ljava/lang/String;]
SQLConf.getConf ( String key, String defaultValue ) : String
[mangled: org/apache/spark/sql/SQLConf.getConf:(Ljava/lang/String;Ljava/lang/String;)Ljava/lang/String;]
SQLConf.HIVE_VERIFY_PARTITIONPATH ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.HIVE_VERIFY_PARTITIONPATH:()Ljava/lang/String;]
SQLConf.IN_MEMORY_PARTITION_PRUNING ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.IN_MEMORY_PARTITION_PRUNING:()Ljava/lang/String;]
SQLConf.inMemoryPartitionPruning ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.inMemoryPartitionPruning:()Z]
SQLConf.isParquetBinaryAsString ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.isParquetBinaryAsString:()Z]
SQLConf.isParquetINT96AsTimestamp ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.isParquetINT96AsTimestamp:()Z]
SQLConf.numShufflePartitions ( ) : int
[mangled: org/apache/spark/sql/SQLConf.numShufflePartitions:()I]
SQLConf.ORC_FILTER_PUSHDOWN_ENABLED ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.ORC_FILTER_PUSHDOWN_ENABLED:()Ljava/lang/String;]
SQLConf.orcFilterPushDown ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.orcFilterPushDown:()Z]
SQLConf.OUTPUT_COMMITTER_CLASS ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.OUTPUT_COMMITTER_CLASS:()Ljava/lang/String;]
SQLConf.PARQUET_BINARY_AS_STRING ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.PARQUET_BINARY_AS_STRING:()Ljava/lang/String;]
SQLConf.PARQUET_CACHE_METADATA ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.PARQUET_CACHE_METADATA:()Ljava/lang/String;]
SQLConf.PARQUET_COMPRESSION ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.PARQUET_COMPRESSION:()Ljava/lang/String;]
SQLConf.PARQUET_FILTER_PUSHDOWN_ENABLED ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.PARQUET_FILTER_PUSHDOWN_ENABLED:()Ljava/lang/String;]
SQLConf.PARQUET_INT96_AS_TIMESTAMP ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.PARQUET_INT96_AS_TIMESTAMP:()Ljava/lang/String;]
SQLConf.PARQUET_USE_DATA_SOURCE_API ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.PARQUET_USE_DATA_SOURCE_API:()Ljava/lang/String;]
SQLConf.parquetCompressionCodec ( ) : String
[mangled: org/apache/spark/sql/SQLConf.parquetCompressionCodec:()Ljava/lang/String;]
SQLConf.parquetFilterPushDown ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.parquetFilterPushDown:()Z]
SQLConf.parquetUseDataSourceApi ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.parquetUseDataSourceApi:()Z]
SQLConf.PARTITION_DISCOVERY_ENABLED ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.PARTITION_DISCOVERY_ENABLED:()Ljava/lang/String;]
SQLConf.partitionDiscoveryEnabled ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.partitionDiscoveryEnabled:()Z]
SQLConf.SCHEMA_STRING_LENGTH_THRESHOLD ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.SCHEMA_STRING_LENGTH_THRESHOLD:()Ljava/lang/String;]
SQLConf.schemaStringLengthThreshold ( ) : int
[mangled: org/apache/spark/sql/SQLConf.schemaStringLengthThreshold:()I]
SQLConf.setConf ( java.util.Properties props ) : void
[mangled: org/apache/spark/sql/SQLConf.setConf:(Ljava/util/Properties;)V]
SQLConf.setConf ( String key, String value ) : void
[mangled: org/apache/spark/sql/SQLConf.setConf:(Ljava/lang/String;Ljava/lang/String;)V]
SQLConf.settings ( ) : java.util.Map<String,String>
[mangled: org/apache/spark/sql/SQLConf.settings:()Ljava/util/Map;]
SQLConf.SHUFFLE_PARTITIONS ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.SHUFFLE_PARTITIONS:()Ljava/lang/String;]
SQLConf.SORTMERGE_JOIN ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.SORTMERGE_JOIN:()Ljava/lang/String;]
SQLConf.sortMergeJoinEnabled ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.sortMergeJoinEnabled:()Z]
SQLConf.SQLConf ( )
[mangled: org/apache/spark/sql/SQLConf."<init>":()V]
SQLConf.THRIFTSERVER_POOL ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.THRIFTSERVER_POOL:()Ljava/lang/String;]
SQLConf.THRIFTSERVER_UI_SESSION_LIMIT ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.THRIFTSERVER_UI_SESSION_LIMIT:()Ljava/lang/String;]
SQLConf.THRIFTSERVER_UI_STATEMENT_LIMIT ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.THRIFTSERVER_UI_STATEMENT_LIMIT:()Ljava/lang/String;]
SQLConf.UNSAFE_ENABLED ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.UNSAFE_ENABLED:()Ljava/lang/String;]
SQLConf.unsafeEnabled ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.unsafeEnabled:()Z]
SQLConf.unsetConf ( String key ) : void
[mangled: org/apache/spark/sql/SQLConf.unsetConf:(Ljava/lang/String;)V]
SQLConf.USE_JACKSON_STREAMING_API ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.USE_JACKSON_STREAMING_API:()Ljava/lang/String;]
SQLConf.USE_SQL_SERIALIZER2 ( ) [static] : String
[mangled: org/apache/spark/sql/SQLConf.USE_SQL_SERIALIZER2:()Ljava/lang/String;]
SQLConf.useCompression ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.useCompression:()Z]
SQLConf.useJacksonStreamingAPI ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.useJacksonStreamingAPI:()Z]
SQLConf.useSqlSerializer2 ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.useSqlSerializer2:()Z]
SQLConf.verifyPartitionPath ( ) : boolean
[mangled: org/apache/spark/sql/SQLConf.verifyPartitionPath:()Z]
spark-sql_2.10-1.4.0.jar, SQLContext.class
package org.apache.spark.sql
SQLContext.applySchemaToPythonRDD ( org.apache.spark.rdd.RDD<Object[ ]> rdd, types.StructType schema ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.applySchemaToPythonRDD:(Lorg/apache/spark/rdd/RDD;Lorg/apache/spark/sql/types/StructType;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.applySchemaToPythonRDD ( org.apache.spark.rdd.RDD<Object[ ]> rdd, String schemaString ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.applySchemaToPythonRDD:(Lorg/apache/spark/rdd/RDD;Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.baseRelationToDataFrame ( sources.BaseRelation baseRelation ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.baseRelationToDataFrame:(Lorg/apache/spark/sql/sources/BaseRelation;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.cacheManager ( ) : execution.CacheManager
[mangled: org/apache/spark/sql/SQLContext.cacheManager:()Lorg/apache/spark/sql/execution/CacheManager;]
SQLContext.clearCache ( ) : void
[mangled: org/apache/spark/sql/SQLContext.clearCache:()V]
SQLContext.conf ( ) : SQLConf
[mangled: org/apache/spark/sql/SQLContext.conf:()Lorg/apache/spark/sql/SQLConf;]
SQLContext.createDataFrame ( org.apache.spark.api.java.JavaRDD<?> rdd, Class<?> beanClass ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/api/java/JavaRDD;Ljava/lang/Class;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createDataFrame ( org.apache.spark.api.java.JavaRDD<Row> rowRDD, types.StructType schema ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/api/java/JavaRDD;Lorg/apache/spark/sql/types/StructType;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createDataFrame ( org.apache.spark.rdd.RDD<?> rdd, Class<?> beanClass ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/rdd/RDD;Ljava/lang/Class;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createDataFrame ( org.apache.spark.rdd.RDD<A> rdd, scala.reflect.api.TypeTags.TypeTag<A> p2 ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/rdd/RDD;Lscala/reflect/api/TypeTags$TypeTag;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createDataFrame ( org.apache.spark.rdd.RDD<Row> rowRDD, types.StructType schema ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/rdd/RDD;Lorg/apache/spark/sql/types/StructType;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createDataFrame ( org.apache.spark.rdd.RDD<Row> rowRDD, types.StructType schema, boolean needsConversion ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lorg/apache/spark/rdd/RDD;Lorg/apache/spark/sql/types/StructType;Z)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createDataFrame ( scala.collection.Seq<A> data, scala.reflect.api.TypeTags.TypeTag<A> p2 ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createDataFrame:(Lscala/collection/Seq;Lscala/reflect/api/TypeTags$TypeTag;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createExternalTable ( String tableName, String path ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createExternalTable:(Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createExternalTable ( String tableName, String path, String source ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createExternalTable:(Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createExternalTable ( String tableName, String source, java.util.Map<String,String> options ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createExternalTable:(Ljava/lang/String;Ljava/lang/String;Ljava/util/Map;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createExternalTable ( String tableName, String source, types.StructType schema, java.util.Map<String,String> options ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createExternalTable:(Ljava/lang/String;Ljava/lang/String;Lorg/apache/spark/sql/types/StructType;Ljava/util/Map;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createExternalTable ( String tableName, String source, types.StructType schema, scala.collection.immutable.Map<String,String> options ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createExternalTable:(Ljava/lang/String;Ljava/lang/String;Lorg/apache/spark/sql/types/StructType;Lscala/collection/immutable/Map;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createExternalTable ( String tableName, String source, scala.collection.immutable.Map<String,String> options ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.createExternalTable:(Ljava/lang/String;Ljava/lang/String;Lscala/collection/immutable/Map;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.createSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.createSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.currentSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.currentSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.ddlParser ( ) : sources.DDLParser
[mangled: org/apache/spark/sql/SQLContext.ddlParser:()Lorg/apache/spark/sql/sources/DDLParser;]
SQLContext.defaultSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.defaultSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.detachSession ( ) : void
[mangled: org/apache/spark/sql/SQLContext.detachSession:()V]
SQLContext.dialectClassName ( ) : String
[mangled: org/apache/spark/sql/SQLContext.dialectClassName:()Ljava/lang/String;]
SQLContext.dropTempTable ( String tableName ) : void
[mangled: org/apache/spark/sql/SQLContext.dropTempTable:(Ljava/lang/String;)V]
SQLContext.emptyDataFrame ( ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.emptyDataFrame:()Lorg/apache/spark/sql/DataFrame;]
SQLContext.emptyResult ( ) : org.apache.spark.rdd.RDD<Row>
[mangled: org/apache/spark/sql/SQLContext.emptyResult:()Lorg/apache/spark/rdd/RDD;]
SQLContext.experimental ( ) : ExperimentalMethods
[mangled: org/apache/spark/sql/SQLContext.experimental:()Lorg/apache/spark/sql/ExperimentalMethods;]
SQLContext.functionRegistry ( ) : catalyst.analysis.FunctionRegistry
[mangled: org/apache/spark/sql/SQLContext.functionRegistry:()Lorg/apache/spark/sql/catalyst/analysis/FunctionRegistry;]
SQLContext.getAllConfs ( ) : scala.collection.immutable.Map<String,String>
[mangled: org/apache/spark/sql/SQLContext.getAllConfs:()Lscala/collection/immutable/Map;]
SQLContext.getConf ( String key ) : String
[mangled: org/apache/spark/sql/SQLContext.getConf:(Ljava/lang/String;)Ljava/lang/String;]
SQLContext.getConf ( String key, String defaultValue ) : String
[mangled: org/apache/spark/sql/SQLContext.getConf:(Ljava/lang/String;Ljava/lang/String;)Ljava/lang/String;]
SQLContext.getOrCreate ( org.apache.spark.SparkContext p1 ) [static] : SQLContext
[mangled: org/apache/spark/sql/SQLContext.getOrCreate:(Lorg/apache/spark/SparkContext;)Lorg/apache/spark/sql/SQLContext;]
SQLContext.getSchema ( Class<?> beanClass ) : scala.collection.Seq<catalyst.expressions.AttributeReference>
[mangled: org/apache/spark/sql/SQLContext.getSchema:(Ljava/lang/Class;)Lscala/collection/Seq;]
SQLContext.getSQLDialect ( ) : catalyst.ParserDialect
[mangled: org/apache/spark/sql/SQLContext.getSQLDialect:()Lorg/apache/spark/sql/catalyst/ParserDialect;]
SQLContext.implicits ( ) : SQLContext.implicits.
[mangled: org/apache/spark/sql/SQLContext.implicits:()Lorg/apache/spark/sql/SQLContext$implicits$;]
SQLContext.isCached ( String tableName ) : boolean
[mangled: org/apache/spark/sql/SQLContext.isCached:(Ljava/lang/String;)Z]
SQLContext.isTraceEnabled ( ) : boolean
[mangled: org/apache/spark/sql/SQLContext.isTraceEnabled:()Z]
SQLContext.log ( ) : org.slf4j.Logger
[mangled: org/apache/spark/sql/SQLContext.log:()Lorg/slf4j/Logger;]
SQLContext.logDebug ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/SQLContext.logDebug:(Lscala/Function0;)V]
SQLContext.logDebug ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/SQLContext.logDebug:(Lscala/Function0;Ljava/lang/Throwable;)V]
SQLContext.logError ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/SQLContext.logError:(Lscala/Function0;)V]
SQLContext.logError ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/SQLContext.logError:(Lscala/Function0;Ljava/lang/Throwable;)V]
SQLContext.logInfo ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/SQLContext.logInfo:(Lscala/Function0;)V]
SQLContext.logInfo ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/SQLContext.logInfo:(Lscala/Function0;Ljava/lang/Throwable;)V]
SQLContext.logName ( ) : String
[mangled: org/apache/spark/sql/SQLContext.logName:()Ljava/lang/String;]
SQLContext.logTrace ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/SQLContext.logTrace:(Lscala/Function0;)V]
SQLContext.logTrace ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/SQLContext.logTrace:(Lscala/Function0;Ljava/lang/Throwable;)V]
SQLContext.logWarning ( scala.Function0<String> msg ) : void
[mangled: org/apache/spark/sql/SQLContext.logWarning:(Lscala/Function0;)V]
SQLContext.logWarning ( scala.Function0<String> msg, Throwable throwable ) : void
[mangled: org/apache/spark/sql/SQLContext.logWarning:(Lscala/Function0;Ljava/lang/Throwable;)V]
SQLContext.openSession ( ) : SQLContext.SQLSession
[mangled: org/apache/spark/sql/SQLContext.openSession:()Lorg/apache/spark/sql/SQLContext$SQLSession;]
SQLContext.optimizer ( ) : catalyst.optimizer.Optimizer
[mangled: org/apache/spark/sql/SQLContext.optimizer:()Lorg/apache/spark/sql/catalyst/optimizer/Optimizer;]
SQLContext.org.apache.spark.Logging..log_ ( ) : org.slf4j.Logger
[mangled: org/apache/spark/sql/SQLContext.org.apache.spark.Logging..log_:()Lorg/slf4j/Logger;]
SQLContext.org.apache.spark.Logging..log__.eq ( org.slf4j.Logger p1 ) : void
[mangled: org/apache/spark/sql/SQLContext.org.apache.spark.Logging..log__.eq:(Lorg/slf4j/Logger;)V]
SQLContext.parquetFile ( String... paths ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.parquetFile:([Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.parseDataType ( String dataTypeString ) : types.DataType
[mangled: org/apache/spark/sql/SQLContext.parseDataType:(Ljava/lang/String;)Lorg/apache/spark/sql/types/DataType;]
SQLContext.range ( long start, long end ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.range:(JJ)Lorg/apache/spark/sql/DataFrame;]
SQLContext.range ( long start, long end, long step, int numPartitions ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.range:(JJJI)Lorg/apache/spark/sql/DataFrame;]
SQLContext.read ( ) : DataFrameReader
[mangled: org/apache/spark/sql/SQLContext.read:()Lorg/apache/spark/sql/DataFrameReader;]
SQLContext.registerDataFrameAsTable ( DataFrame df, String tableName ) : void
[mangled: org/apache/spark/sql/SQLContext.registerDataFrameAsTable:(Lorg/apache/spark/sql/DataFrame;Ljava/lang/String;)V]
SQLContext.setConf ( java.util.Properties props ) : void
[mangled: org/apache/spark/sql/SQLContext.setConf:(Ljava/util/Properties;)V]
SQLContext.setConf ( String key, String value ) : void
[mangled: org/apache/spark/sql/SQLContext.setConf:(Ljava/lang/String;Ljava/lang/String;)V]
SQLContext.sql ( String sqlText ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.sql:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.SQLContext ( org.apache.spark.api.java.JavaSparkContext sparkContext )
[mangled: org/apache/spark/sql/SQLContext."<init>":(Lorg/apache/spark/api/java/JavaSparkContext;)V]
SQLContext.sqlParser ( ) : SparkSQLParser
[mangled: org/apache/spark/sql/SQLContext.sqlParser:()Lorg/apache/spark/sql/SparkSQLParser;]
SQLContext.table ( String tableName ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.table:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.tableNames ( ) : String[ ]
[mangled: org/apache/spark/sql/SQLContext.tableNames:()[Ljava/lang/String;]
SQLContext.tableNames ( String databaseName ) : String[ ]
[mangled: org/apache/spark/sql/SQLContext.tableNames:(Ljava/lang/String;)[Ljava/lang/String;]
SQLContext.tables ( ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.tables:()Lorg/apache/spark/sql/DataFrame;]
SQLContext.tables ( String databaseName ) : DataFrame
[mangled: org/apache/spark/sql/SQLContext.tables:(Ljava/lang/String;)Lorg/apache/spark/sql/DataFrame;]
SQLContext.tlSession ( ) : ThreadLocal<SQLContext.SQLSession>
[mangled: org/apache/spark/sql/SQLContext.tlSession:()Ljava/lang/ThreadLocal;]
SQLContext.udf ( ) : UDFRegistration
[mangled: org/apache/spark/sql/SQLContext.udf:()Lorg/apache/spark/sql/UDFRegistration;]
to the top
Problems with Data Types, High Severity (64)
spark-catalyst_2.10-1.4.0.jar
package org.apache.spark.sql
[+] Row (1)
| Change | Effect |
---|
1 | This interface has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (34)
anyNull ( )This abstract method is from 'Row' interface.
apply ( int )This abstract method is from 'Row' interface.
copy ( )This abstract method is from 'Row' interface.
equals ( java.lang.Object )This abstract method is from 'Row' interface.
fieldIndex ( java.lang.String )This abstract method is from 'Row' interface.
get ( int )This abstract method is from 'Row' interface.
getAs ( int )This abstract method is from 'Row' interface.
getAs ( java.lang.String )This abstract method is from 'Row' interface.
getBoolean ( int )This abstract method is from 'Row' interface.
getByte ( int )This abstract method is from 'Row' interface.
getDate ( int )This abstract method is from 'Row' interface.
getDecimal ( int )This abstract method is from 'Row' interface.
getDouble ( int )This abstract method is from 'Row' interface.
getFloat ( int )This abstract method is from 'Row' interface.
getInt ( int )This abstract method is from 'Row' interface.
getJavaMap ( int )This abstract method is from 'Row' interface.
getList ( int )This abstract method is from 'Row' interface.
getLong ( int )This abstract method is from 'Row' interface.
getMap ( int )This abstract method is from 'Row' interface.
getSeq ( int )This abstract method is from 'Row' interface.
getShort ( int )This abstract method is from 'Row' interface.
getString ( int )This abstract method is from 'Row' interface.
getStruct ( int )This abstract method is from 'Row' interface.
getValuesMap ( scala.collection.Seq<java.lang.String> )This abstract method is from 'Row' interface.
hashCode ( )This abstract method is from 'Row' interface.
isNullAt ( int )This abstract method is from 'Row' interface.
length ( )This abstract method is from 'Row' interface.
mkString ( )This abstract method is from 'Row' interface.
mkString ( java.lang.String )This abstract method is from 'Row' interface.
mkString ( java.lang.String, java.lang.String, java.lang.String )This abstract method is from 'Row' interface.
schema ( )This abstract method is from 'Row' interface.
size ( )This abstract method is from 'Row' interface.
toSeq ( )This abstract method is from 'Row' interface.
toString ( )This abstract method is from 'Row' interface.
package org.apache.spark.sql.catalyst
[+] ParserDialect (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (2)
parse ( java.lang.String )This abstract method is from 'ParserDialect' abstract class.
ParserDialect ( )This constructor is from 'ParserDialect' abstract class.
[+] SimpleCatalystConf (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (1)
SimpleCatalystConf ( boolean )This constructor is from 'SimpleCatalystConf' class.
[+] SqlParser (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.sql.types.DataTypeParser. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (24)
andExpression ( )This method is from 'SqlParser' class.
assignAliases ( scala.collection.Seq<expressions.Expression> )This method is from 'SqlParser' class.
cast ( )This method is from 'SqlParser' class.
comparisonExpression ( )This method is from 'SqlParser' class.
dataType ( )This method is from 'SqlParser' class.
direction ( )This method is from 'SqlParser' class.
expression ( )This method is from 'SqlParser' class.
floatLit ( )This method is from 'SqlParser' class.
function ( )This method is from 'SqlParser' class.
insert ( )This method is from 'SqlParser' class.
joinConditions ( )This method is from 'SqlParser' class.
joinedRelation ( )This method is from 'SqlParser' class.
joinType ( )This method is from 'SqlParser' class.
literal ( )This method is from 'SqlParser' class.
ordering ( )This method is from 'SqlParser' class.
orExpression ( )This method is from 'SqlParser' class.
productExpression ( )This method is from 'SqlParser' class.
projection ( )This method is from 'SqlParser' class.
relation ( )This method is from 'SqlParser' class.
relationFactor ( )This method is from 'SqlParser' class.
relations ( )This method is from 'SqlParser' class.
select ( )This method is from 'SqlParser' class.
SqlParser ( )This constructor is from 'SqlParser' class.
termExpression ( )This method is from 'SqlParser' class.
package org.apache.spark.sql.catalyst.analysis
[+] Analyzer (1)
| Change | Effect |
---|
1 | Removed super-interface CheckAnalysis. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (1)
analyzer ( )Return value of this method has type 'Analyzer'.
[+] Catalog (10)
| Change | Effect |
---|
1 | Abstract method conf ( ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
2 | Abstract method getDBTable ( scala.collection.Seq<java.lang.String> ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
3 | Abstract method getDbTableName ( scala.collection.Seq<java.lang.String> ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
4 | Abstract method getTables ( scala.Option<java.lang.String> ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
5 | Abstract method lookupRelation ( scala.collection.Seq<java.lang.String>, scala.Option<java.lang.String> ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
6 | Abstract method processTableIdentifier ( scala.collection.Seq<java.lang.String> ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
7 | Abstract method refreshTable ( java.lang.String, java.lang.String ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
8 | Abstract method registerTable ( scala.collection.Seq<java.lang.String>, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
9 | Abstract method tableExists ( scala.collection.Seq<java.lang.String> ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
10 | Abstract method unregisterTable ( scala.collection.Seq<java.lang.String> ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (2)
unregisterAllTables ( )This abstract method is from 'Catalog' interface.
catalog ( )Return value of this method has type 'Catalog'.
package org.apache.spark.sql.catalyst.expressions
[+] And (2)
| Change | Effect |
---|
1 | Removed super-interface ExpectsInputTypes. | A client program may be interrupted by NoSuchMethodError exception. |
2 | Removed super-interface Predicate. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (3)
And ( Expression, Expression )This constructor is from 'And' class.
left ( )This method is from 'And' class.
right ( )This method is from 'And' class.
[+] AttributeSet (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (5)
exists ( scala.Function1<Attribute,java.lang.Object> )This method is from 'AttributeSet' class.
map ( scala.Function1<Attribute,B>, scala.collection.generic.CanBuildFrom<scala.collection.Traversable<Attribute>,B,That> )This method is from 'AttributeSet' class.
size ( )This method is from 'AttributeSet' class.
subsetOf ( AttributeSet )This method is from 'AttributeSet' class.
toSeq ( )This method is from 'AttributeSet' class.
[+] EqualTo (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (3)
EqualTo ( Expression, Expression )This constructor is from 'EqualTo' class.
left ( )This method is from 'EqualTo' class.
right ( )This method is from 'EqualTo' class.
[+] Expression (2)
| Change | Effect |
---|
1 | Abstract method dataType ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
2 | Abstract method eval ( org.apache.spark.sql.Row ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (32)
And ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
childrenResolved ( )This method is from 'Expression' abstract class.
Expression ( )This constructor is from 'Expression' abstract class.
foldable ( )This method is from 'Expression' abstract class.
nullable ( )This abstract method is from 'Expression' abstract class.
resolved ( )This method is from 'Expression' abstract class.
GreaterThan ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
GreaterThanOrEqual ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
falseValue ( )Return value of this method has type 'Expression'.
predicate ( )Return value of this method has type 'Expression'.
trueValue ( )Return value of this method has type 'Expression'.
In ( Expression, scala.collection.Seq<Expression> )1st parameter 'value' of this method has type 'Expression'.
value ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
LessThan ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
LessThanOrEqual ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
Not ( Expression )1st parameter 'child' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
Or ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
[+] GenericRow (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.sql.Row. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (1)
GenericRow ( int )This constructor is from 'GenericRow' class.
[+] InSet (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (3)
hset ( )This method is from 'InSet' class.
InSet ( Expression, scala.collection.immutable.Set<java.lang.Object> )This constructor is from 'InSet' class.
value ( )This method is from 'InSet' class.
[+] MutableRow (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.sql.Row. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (10)
setBoolean ( int, boolean )This abstract method is from 'MutableRow' interface.
setByte ( int, byte )This abstract method is from 'MutableRow' interface.
setDouble ( int, double )This abstract method is from 'MutableRow' interface.
setFloat ( int, float )This abstract method is from 'MutableRow' interface.
setInt ( int, int )This abstract method is from 'MutableRow' interface.
setLong ( int, long )This abstract method is from 'MutableRow' interface.
setNullAt ( int )This abstract method is from 'MutableRow' interface.
setShort ( int, short )This abstract method is from 'MutableRow' interface.
setString ( int, java.lang.String )This abstract method is from 'MutableRow' interface.
update ( int, java.lang.Object )This abstract method is from 'MutableRow' interface.
[+] Not (1)
| Change | Effect |
---|
1 | Removed super-interface ExpectsInputTypes. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (2)
child ( )This method is from 'Not' class.
Not ( Expression )This constructor is from 'Not' class.
[+] Or (2)
| Change | Effect |
---|
1 | Removed super-interface ExpectsInputTypes. | A client program may be interrupted by NoSuchMethodError exception. |
2 | Removed super-interface Predicate. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (3)
left ( )This method is from 'Or' class.
Or ( Expression, Expression )This constructor is from 'Or' class.
right ( )This method is from 'Or' class.
package org.apache.spark.sql.catalyst.plans.logical
[+] LogicalPlan (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.Logging. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (8)
child ( )Return value of this method has type 'LogicalPlan'.
childrenResolved ( )This method is from 'LogicalPlan' abstract class.
LogicalPlan ( )This constructor is from 'LogicalPlan' abstract class.
resolved ( )This method is from 'LogicalPlan' abstract class.
child ( )Return value of this method has type 'LogicalPlan'.
Subquery ( java.lang.String, LogicalPlan )2nd parameter 'child' of this method has type 'LogicalPlan'.
executePlan ( LogicalPlan )1st parameter 'plan' of this method has type 'LogicalPlan'.
parseSql ( java.lang.String )Return value of this method has type 'LogicalPlan'.
package org.apache.spark.sql.types
[+] AtomicType (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (4)
AtomicType ( )This constructor is from 'AtomicType' abstract class.
classTag ( )This method is from 'AtomicType' abstract class.
ordering ( )This abstract method is from 'AtomicType' abstract class.
tag ( )This abstract method is from 'AtomicType' abstract class.
[+] DataType (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (12)
asNullable ( )This abstract method is from 'DataType' abstract class.
DataType ( )This constructor is from 'DataType' abstract class.
defaultSize ( )This abstract method is from 'DataType' abstract class.
fromCaseClassString ( java.lang.String )This method is from 'DataType' abstract class.
fromJson ( java.lang.String )This method is from 'DataType' abstract class.
json ( )This method is from 'DataType' abstract class.
jsonValue ( )This method is from 'DataType' abstract class.
prettyJson ( )This method is from 'DataType' abstract class.
sameType ( DataType )This method is from 'DataType' abstract class.
simpleString ( )This method is from 'DataType' abstract class.
typeName ( )This method is from 'DataType' abstract class.
unapply ( org.apache.spark.sql.catalyst.expressions.Expression )This method is from 'DataType' abstract class.
[+] StructField (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (1)
StructField ( java.lang.String, DataType, boolean, Metadata )This constructor is from 'StructField' class.
[+] StructType (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (1)
toAttributes ( )This method is from 'StructType' class.
[+] UTF8String (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (22)
apply ( byte[ ] )This method is from 'UTF8String' class.
apply ( java.lang.String )This method is from 'UTF8String' class.
clone ( )This method is from 'UTF8String' class.
clone ( )This method is from 'UTF8String' class.
compare ( java.lang.Object )This method is from 'UTF8String' class.
compare ( UTF8String )This method is from 'UTF8String' class.
compareTo ( java.lang.Object )This method is from 'UTF8String' class.
compareTo ( UTF8String )This method is from 'UTF8String' class.
contains ( UTF8String )This method is from 'UTF8String' class.
endsWith ( UTF8String )This method is from 'UTF8String' class.
equals ( java.lang.Object )This method is from 'UTF8String' class.
getBytes ( )This method is from 'UTF8String' class.
hashCode ( )This method is from 'UTF8String' class.
length ( )This method is from 'UTF8String' class.
set ( byte[ ] )This method is from 'UTF8String' class.
set ( java.lang.String )This method is from 'UTF8String' class.
slice ( int, int )This method is from 'UTF8String' class.
startsWith ( UTF8String )This method is from 'UTF8String' class.
toLowerCase ( )This method is from 'UTF8String' class.
toString ( )This method is from 'UTF8String' class.
toUpperCase ( )This method is from 'UTF8String' class.
UTF8String ( )This constructor is from 'UTF8String' class.
spark-core_2.10-1.4.0.jar
package org.apache.spark
[+] Logging (1)
| Change | Effect |
---|
1 | Abstract method logName ( ) has been removed from this interface. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (14)
isTraceEnabled ( )This abstract method is from 'Logging' interface.
log ( )This abstract method is from 'Logging' interface.
logDebug ( scala.Function0<java.lang.String> )This abstract method is from 'Logging' interface.
logDebug ( scala.Function0<java.lang.String>, java.lang.Throwable )This abstract method is from 'Logging' interface.
logError ( scala.Function0<java.lang.String> )This abstract method is from 'Logging' interface.
logError ( scala.Function0<java.lang.String>, java.lang.Throwable )This abstract method is from 'Logging' interface.
logInfo ( scala.Function0<java.lang.String> )This abstract method is from 'Logging' interface.
logInfo ( scala.Function0<java.lang.String>, java.lang.Throwable )This abstract method is from 'Logging' interface.
logTrace ( scala.Function0<java.lang.String> )This abstract method is from 'Logging' interface.
logTrace ( scala.Function0<java.lang.String>, java.lang.Throwable )This abstract method is from 'Logging' interface.
logWarning ( scala.Function0<java.lang.String> )This abstract method is from 'Logging' interface.
logWarning ( scala.Function0<java.lang.String>, java.lang.Throwable )This abstract method is from 'Logging' interface.
Logging..log_ ( )This abstract method is from 'Logging' interface.
Logging..log__.eq ( org.slf4j.Logger )This abstract method is from 'Logging' interface.
[+] SparkContext (1)
| Change | Effect |
---|
1 | Removed super-interface ExecutorAllocationClient. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (149)
fromSparkContext ( SparkContext )1st parameter 'p1' of this method has type 'SparkContext'.
JavaSparkContext ( SparkContext )1st parameter 'sc' of this method has type 'SparkContext'.
sc ( )Return value of this method has type 'SparkContext'.
toSparkContext ( api.java.JavaSparkContext )Return value of this method has type 'SparkContext'.
accumulable ( R, AccumulableParam<R,T> )This method is from 'SparkContext' class.
accumulableCollection ( R, scala.Function1<R,scala.collection.generic.Growable<T>>, scala.reflect.ClassTag<R> )This method is from 'SparkContext' class.
accumulator ( T, AccumulatorParam<T> )This method is from 'SparkContext' class.
addedFiles ( )This method is from 'SparkContext' class.
addedJars ( )This method is from 'SparkContext' class.
addFile ( java.lang.String )This method is from 'SparkContext' class.
addJar ( java.lang.String )This method is from 'SparkContext' class.
addSparkListener ( scheduler.SparkListener )This method is from 'SparkContext' class.
appName ( )This method is from 'SparkContext' class.
booleanWritableConverter ( )This method is from 'SparkContext' class.
boolToBoolWritable ( boolean )This method is from 'SparkContext' class.
broadcast ( T, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
bytesToBytesWritable ( byte[ ] )This method is from 'SparkContext' class.
bytesWritableConverter ( )This method is from 'SparkContext' class.
cancelAllJobs ( )This method is from 'SparkContext' class.
cancelJob ( int )This method is from 'SparkContext' class.
cancelJobGroup ( java.lang.String )This method is from 'SparkContext' class.
cancelStage ( int )This method is from 'SparkContext' class.
checkpointDir ( )This method is from 'SparkContext' class.
checkpointDir_.eq ( scala.Option<java.lang.String> )This method is from 'SparkContext' class.
checkpointFile ( java.lang.String, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
cleaner ( )This method is from 'SparkContext' class.
cleanup ( long )This method is from 'SparkContext' class.
clearCallSite ( )This method is from 'SparkContext' class.
clearJobGroup ( )This method is from 'SparkContext' class.
conf ( )This method is from 'SparkContext' class.
dagScheduler ( )This method is from 'SparkContext' class.
dagScheduler_.eq ( scheduler.DAGScheduler )This method is from 'SparkContext' class.
defaultMinPartitions ( )This method is from 'SparkContext' class.
defaultParallelism ( )This method is from 'SparkContext' class.
doubleRDDToDoubleRDDFunctions ( rdd.RDD<java.lang.Object> )This method is from 'SparkContext' class.
doubleToDoubleWritable ( double )This method is from 'SparkContext' class.
doubleWritableConverter ( )This method is from 'SparkContext' class.
emptyRDD ( scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
env ( )This method is from 'SparkContext' class.
eventLogger ( )This method is from 'SparkContext' class.
executorEnvs ( )This method is from 'SparkContext' class.
executorMemory ( )This method is from 'SparkContext' class.
files ( )This method is from 'SparkContext' class.
floatToFloatWritable ( float )This method is from 'SparkContext' class.
floatWritableConverter ( )This method is from 'SparkContext' class.
getAllPools ( )This method is from 'SparkContext' class.
getCheckpointDir ( )This method is from 'SparkContext' class.
getConf ( )This method is from 'SparkContext' class.
getExecutorMemoryStatus ( )This method is from 'SparkContext' class.
getExecutorStorageStatus ( )This method is from 'SparkContext' class.
getLocalProperties ( )This method is from 'SparkContext' class.
getLocalProperty ( java.lang.String )This method is from 'SparkContext' class.
getPersistentRDDs ( )This method is from 'SparkContext' class.
getPoolForName ( java.lang.String )This method is from 'SparkContext' class.
getPreferredLocs ( rdd.RDD<?>, int )This method is from 'SparkContext' class.
getRDDStorageInfo ( )This method is from 'SparkContext' class.
getSchedulingMode ( )This method is from 'SparkContext' class.
getSparkHome ( )This method is from 'SparkContext' class.
hadoopConfiguration ( )This method is from 'SparkContext' class.
hadoopFile ( java.lang.String, int, scala.reflect.ClassTag<K>, scala.reflect.ClassTag<V>, scala.reflect.ClassTag<F> )This method is from 'SparkContext' class.
hadoopFile ( java.lang.String, java.lang.Class<? extends org.apache.hadoop.mapred.InputFormat<K,V>>, java.lang.Class<K>, java.lang.Class<V>, int )This method is from 'SparkContext' class.
hadoopFile ( java.lang.String, scala.reflect.ClassTag<K>, scala.reflect.ClassTag<V>, scala.reflect.ClassTag<F> )This method is from 'SparkContext' class.
hadoopRDD ( org.apache.hadoop.mapred.JobConf, java.lang.Class<? extends org.apache.hadoop.mapred.InputFormat<K,V>>, java.lang.Class<K>, java.lang.Class<V>, int )This method is from 'SparkContext' class.
intToIntWritable ( int )This method is from 'SparkContext' class.
intWritableConverter ( )This method is from 'SparkContext' class.
isLocal ( )This method is from 'SparkContext' class.
isTraceEnabled ( )This method is from 'SparkContext' class.
jarOfClass ( java.lang.Class<?> )This method is from 'SparkContext' class.
jarOfObject ( java.lang.Object )This method is from 'SparkContext' class.
jars ( )This method is from 'SparkContext' class.
listenerBus ( )This method is from 'SparkContext' class.
log ( )This method is from 'SparkContext' class.
logDebug ( scala.Function0<java.lang.String> )This method is from 'SparkContext' class.
logDebug ( scala.Function0<java.lang.String>, java.lang.Throwable )This method is from 'SparkContext' class.
logError ( scala.Function0<java.lang.String> )This method is from 'SparkContext' class.
logError ( scala.Function0<java.lang.String>, java.lang.Throwable )This method is from 'SparkContext' class.
logInfo ( scala.Function0<java.lang.String> )This method is from 'SparkContext' class.
logInfo ( scala.Function0<java.lang.String>, java.lang.Throwable )This method is from 'SparkContext' class.
logTrace ( scala.Function0<java.lang.String> )This method is from 'SparkContext' class.
logTrace ( scala.Function0<java.lang.String>, java.lang.Throwable )This method is from 'SparkContext' class.
logWarning ( scala.Function0<java.lang.String> )This method is from 'SparkContext' class.
logWarning ( scala.Function0<java.lang.String>, java.lang.Throwable )This method is from 'SparkContext' class.
longToLongWritable ( long )This method is from 'SparkContext' class.
longWritableConverter ( )This method is from 'SparkContext' class.
makeRDD ( scala.collection.Seq<scala.Tuple2<T,scala.collection.Seq<java.lang.String>>>, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
makeRDD ( scala.collection.Seq<T>, int, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
master ( )This method is from 'SparkContext' class.
metadataCleaner ( )This method is from 'SparkContext' class.
newAPIHadoopFile ( java.lang.String, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V>, org.apache.hadoop.conf.Configuration )This method is from 'SparkContext' class.
newAPIHadoopFile ( java.lang.String, scala.reflect.ClassTag<K>, scala.reflect.ClassTag<V>, scala.reflect.ClassTag<F> )This method is from 'SparkContext' class.
newAPIHadoopRDD ( org.apache.hadoop.conf.Configuration, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V> )This method is from 'SparkContext' class.
newRddId ( )This method is from 'SparkContext' class.
newShuffleId ( )This method is from 'SparkContext' class.
numericRDDToDoubleRDDFunctions ( rdd.RDD<T>, scala.math.Numeric<T> )This method is from 'SparkContext' class.
objectFile ( java.lang.String, int, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
Logging..log_ ( )This method is from 'SparkContext' class.
Logging..log__.eq ( org.slf4j.Logger )This method is from 'SparkContext' class.
SparkContext..warnSparkMem ( java.lang.String )This method is from 'SparkContext' class.
parallelize ( scala.collection.Seq<T>, int, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
persistentRdds ( )This method is from 'SparkContext' class.
persistRDD ( rdd.RDD<?> )This method is from 'SparkContext' class.
preferredNodeLocationData ( )This method is from 'SparkContext' class.
preferredNodeLocationData_.eq ( scala.collection.Map<java.lang.String,scala.collection.Set<scheduler.SplitInfo>> )This method is from 'SparkContext' class.
rddToAsyncRDDActions ( rdd.RDD<T>, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
rddToOrderedRDDFunctions ( rdd.RDD<scala.Tuple2<K,V>>, scala.math.Ordering<K>, scala.reflect.ClassTag<K>, scala.reflect.ClassTag<V> )This method is from 'SparkContext' class.
rddToPairRDDFunctions ( rdd.RDD<scala.Tuple2<K,V>>, scala.reflect.ClassTag<K>, scala.reflect.ClassTag<V>, scala.math.Ordering<K> )This method is from 'SparkContext' class.
rddToSequenceFileRDDFunctions ( rdd.RDD<scala.Tuple2<K,V>>, scala.Function1<K,org.apache.hadoop.io.Writable>, scala.reflect.ClassTag<K>, scala.Function1<V,org.apache.hadoop.io.Writable>, scala.reflect.ClassTag<V> )This method is from 'SparkContext' class.
runApproximateJob ( rdd.RDD<T>, scala.Function2<TaskContext,scala.collection.Iterator<T>,U>, partial.ApproximateEvaluator<U,R>, long )This method is from 'SparkContext' class.
runJob ( rdd.RDD<T>, scala.Function1<scala.collection.Iterator<T>,U>, scala.collection.Seq<java.lang.Object>, boolean, scala.reflect.ClassTag<U> )This method is from 'SparkContext' class.
runJob ( rdd.RDD<T>, scala.Function1<scala.collection.Iterator<T>,U>, scala.Function2<java.lang.Object,U,scala.runtime.BoxedUnit>, scala.reflect.ClassTag<U> )This method is from 'SparkContext' class.
runJob ( rdd.RDD<T>, scala.Function1<scala.collection.Iterator<T>,U>, scala.reflect.ClassTag<U> )This method is from 'SparkContext' class.
runJob ( rdd.RDD<T>, scala.Function2<TaskContext,scala.collection.Iterator<T>,U>, scala.collection.Seq<java.lang.Object>, boolean, scala.Function2<java.lang.Object,U,scala.runtime.BoxedUnit>, scala.reflect.ClassTag<U> )This method is from 'SparkContext' class.
runJob ( rdd.RDD<T>, scala.Function2<TaskContext,scala.collection.Iterator<T>,U>, scala.collection.Seq<java.lang.Object>, boolean, scala.reflect.ClassTag<U> )This method is from 'SparkContext' class.
runJob ( rdd.RDD<T>, scala.Function2<TaskContext,scala.collection.Iterator<T>,U>, scala.Function2<java.lang.Object,U,scala.runtime.BoxedUnit>, scala.reflect.ClassTag<U> )This method is from 'SparkContext' class.
runJob ( rdd.RDD<T>, scala.Function2<TaskContext,scala.collection.Iterator<T>,U>, scala.reflect.ClassTag<U> )This method is from 'SparkContext' class.
sequenceFile ( java.lang.String, int, scala.reflect.ClassTag<K>, scala.reflect.ClassTag<V>, scala.Function0<WritableConverter<K>>, scala.Function0<WritableConverter<V>> )This method is from 'SparkContext' class.
sequenceFile ( java.lang.String, java.lang.Class<K>, java.lang.Class<V> )This method is from 'SparkContext' class.
sequenceFile ( java.lang.String, java.lang.Class<K>, java.lang.Class<V>, int )This method is from 'SparkContext' class.
setCallSite ( java.lang.String )This method is from 'SparkContext' class.
setCheckpointDir ( java.lang.String )This method is from 'SparkContext' class.
setJobDescription ( java.lang.String )This method is from 'SparkContext' class.
setJobGroup ( java.lang.String, java.lang.String, boolean )This method is from 'SparkContext' class.
setLocalProperties ( java.util.Properties )This method is from 'SparkContext' class.
setLocalProperty ( java.lang.String, java.lang.String )This method is from 'SparkContext' class.
SparkContext ( )This constructor is from 'SparkContext' class.
SparkContext ( java.lang.String, java.lang.String )This constructor is from 'SparkContext' class.
SparkContext ( java.lang.String, java.lang.String, java.lang.String )This constructor is from 'SparkContext' class.
SparkContext ( java.lang.String, java.lang.String, java.lang.String, scala.collection.Seq<java.lang.String> )This constructor is from 'SparkContext' class.
SparkContext ( java.lang.String, java.lang.String, java.lang.String, scala.collection.Seq<java.lang.String>, scala.collection.Map<java.lang.String,java.lang.String>, scala.collection.Map<java.lang.String,scala.collection.Set<scheduler.SplitInfo>> )This constructor is from 'SparkContext' class.
SparkContext ( java.lang.String, java.lang.String, SparkConf )This constructor is from 'SparkContext' class.
SparkContext ( SparkConf )This constructor is from 'SparkContext' class.
SparkContext ( SparkConf, scala.collection.Map<java.lang.String,scala.collection.Set<scheduler.SplitInfo>> )This constructor is from 'SparkContext' class.
sparkUser ( )This method is from 'SparkContext' class.
startTime ( )This method is from 'SparkContext' class.
stop ( )This method is from 'SparkContext' class.
stringToText ( java.lang.String )This method is from 'SparkContext' class.
stringWritableConverter ( )This method is from 'SparkContext' class.
submitJob ( rdd.RDD<T>, scala.Function1<scala.collection.Iterator<T>,U>, scala.collection.Seq<java.lang.Object>, scala.Function2<java.lang.Object,U,scala.runtime.BoxedUnit>, scala.Function0<R> )This method is from 'SparkContext' class.
taskScheduler ( )This method is from 'SparkContext' class.
taskScheduler_.eq ( scheduler.TaskScheduler )This method is from 'SparkContext' class.
textFile ( java.lang.String, int )This method is from 'SparkContext' class.
union ( rdd.RDD<T>, scala.collection.Seq<rdd.RDD<T>>, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
union ( scala.collection.Seq<rdd.RDD<T>>, scala.reflect.ClassTag<T> )This method is from 'SparkContext' class.
unpersistRDD ( int, boolean )This method is from 'SparkContext' class.
version ( )This method is from 'SparkContext' class.
wholeTextFiles ( java.lang.String, int )This method is from 'SparkContext' class.
writableWritableConverter ( )This method is from 'SparkContext' class.
sparkContext ( )Return value of this method has type 'SparkContext'.
SQLContext ( SparkContext )1st parameter 'sparkContext' of this method has type 'SparkContext'.
[+] TaskContext (9)
| Change | Effect |
---|
1 | Abstract method addTaskCompletionListener ( util.TaskCompletionListener ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
2 | Abstract method addTaskCompletionListener ( scala.Function1<TaskContext,scala.runtime.BoxedUnit> ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
3 | Abstract method attemptNumber ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
4 | Abstract method isCompleted ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
5 | Abstract method isInterrupted ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
6 | Abstract method isRunningLocally ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
7 | Abstract method taskAttemptId ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
8 | Abstract method taskMemoryManager ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
9 | Removed super-interface java.io.Serializable. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (3)
partitionId ( )This abstract method is from 'TaskContext' abstract class.
stageId ( )This abstract method is from 'TaskContext' abstract class.
taskMetrics ( )This abstract method is from 'TaskContext' abstract class.
[+] TaskContextImpl (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (5)
attemptNumber ( )This method is from 'TaskContextImpl' class.
partitionId ( )This method is from 'TaskContextImpl' class.
stageId ( )This method is from 'TaskContextImpl' class.
taskAttemptId ( )This method is from 'TaskContextImpl' class.
TaskContextImpl ( int, int, long, int, unsafe.memory.TaskMemoryManager, boolean, executor.TaskMetrics )This constructor is from 'TaskContextImpl' class.
package org.apache.spark.api.java
[+] JavaSparkContext (1)
| Change | Effect |
---|
1 | Removed super-interface java.io.Closeable. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (70)
accumulable ( T, org.apache.spark.AccumulableParam<T,R> )This method is from 'JavaSparkContext' class.
accumulator ( double )This method is from 'JavaSparkContext' class.
accumulator ( int )This method is from 'JavaSparkContext' class.
accumulator ( T, org.apache.spark.AccumulatorParam<T> )This method is from 'JavaSparkContext' class.
addFile ( java.lang.String )This method is from 'JavaSparkContext' class.
addJar ( java.lang.String )This method is from 'JavaSparkContext' class.
appName ( )This method is from 'JavaSparkContext' class.
broadcast ( T )This method is from 'JavaSparkContext' class.
cancelAllJobs ( )This method is from 'JavaSparkContext' class.
cancelJobGroup ( java.lang.String )This method is from 'JavaSparkContext' class.
checkpointFile ( java.lang.String )This method is from 'JavaSparkContext' class.
clearCallSite ( )This method is from 'JavaSparkContext' class.
clearJobGroup ( )This method is from 'JavaSparkContext' class.
defaultMinPartitions ( )This method is from 'JavaSparkContext' class.
defaultParallelism ( )This method is from 'JavaSparkContext' class.
doubleAccumulator ( double )This method is from 'JavaSparkContext' class.
env ( )This method is from 'JavaSparkContext' class.
fromSparkContext ( org.apache.spark.SparkContext )Return value of this method has type 'JavaSparkContext'.
getCheckpointDir ( )This method is from 'JavaSparkContext' class.
getConf ( )This method is from 'JavaSparkContext' class.
getLocalProperty ( java.lang.String )This method is from 'JavaSparkContext' class.
getSparkHome ( )This method is from 'JavaSparkContext' class.
hadoopConfiguration ( )This method is from 'JavaSparkContext' class.
hadoopFile ( java.lang.String, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V> )This method is from 'JavaSparkContext' class.
hadoopFile ( java.lang.String, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V>, int )This method is from 'JavaSparkContext' class.
hadoopRDD ( org.apache.hadoop.mapred.JobConf, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V> )This method is from 'JavaSparkContext' class.
hadoopRDD ( org.apache.hadoop.mapred.JobConf, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V>, int )This method is from 'JavaSparkContext' class.
intAccumulator ( int )This method is from 'JavaSparkContext' class.
isLocal ( )This method is from 'JavaSparkContext' class.
jarOfClass ( java.lang.Class<?> )This method is from 'JavaSparkContext' class.
jarOfObject ( java.lang.Object )This method is from 'JavaSparkContext' class.
jars ( )This method is from 'JavaSparkContext' class.
JavaSparkContext ( )This constructor is from 'JavaSparkContext' class.
JavaSparkContext ( java.lang.String, java.lang.String )This constructor is from 'JavaSparkContext' class.
JavaSparkContext ( java.lang.String, java.lang.String, java.lang.String, java.lang.String )This constructor is from 'JavaSparkContext' class.
JavaSparkContext ( java.lang.String, java.lang.String, java.lang.String, java.lang.String[ ] )This constructor is from 'JavaSparkContext' class.
JavaSparkContext ( java.lang.String, java.lang.String, java.lang.String, java.lang.String[ ], java.util.Map<java.lang.String,java.lang.String> )This constructor is from 'JavaSparkContext' class.
JavaSparkContext ( java.lang.String, java.lang.String, org.apache.spark.SparkConf )This constructor is from 'JavaSparkContext' class.
JavaSparkContext ( org.apache.spark.SparkConf )This constructor is from 'JavaSparkContext' class.
JavaSparkContext ( org.apache.spark.SparkContext )This constructor is from 'JavaSparkContext' class.
master ( )This method is from 'JavaSparkContext' class.
newAPIHadoopFile ( java.lang.String, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V>, org.apache.hadoop.conf.Configuration )This method is from 'JavaSparkContext' class.
newAPIHadoopRDD ( org.apache.hadoop.conf.Configuration, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V> )This method is from 'JavaSparkContext' class.
objectFile ( java.lang.String )This method is from 'JavaSparkContext' class.
objectFile ( java.lang.String, int )This method is from 'JavaSparkContext' class.
parallelize ( java.util.List<T> )This method is from 'JavaSparkContext' class.
parallelize ( java.util.List<T>, int )This method is from 'JavaSparkContext' class.
parallelizeDoubles ( java.util.List<java.lang.Double> )This method is from 'JavaSparkContext' class.
parallelizeDoubles ( java.util.List<java.lang.Double>, int )This method is from 'JavaSparkContext' class.
parallelizePairs ( java.util.List<scala.Tuple2<K,V>> )This method is from 'JavaSparkContext' class.
parallelizePairs ( java.util.List<scala.Tuple2<K,V>>, int )This method is from 'JavaSparkContext' class.
sc ( )This method is from 'JavaSparkContext' class.
sequenceFile ( java.lang.String, java.lang.Class<K>, java.lang.Class<V> )This method is from 'JavaSparkContext' class.
sequenceFile ( java.lang.String, java.lang.Class<K>, java.lang.Class<V>, int )This method is from 'JavaSparkContext' class.
setCallSite ( java.lang.String )This method is from 'JavaSparkContext' class.
setCheckpointDir ( java.lang.String )This method is from 'JavaSparkContext' class.
setJobGroup ( java.lang.String, java.lang.String )This method is from 'JavaSparkContext' class.
setJobGroup ( java.lang.String, java.lang.String, boolean )This method is from 'JavaSparkContext' class.
setLocalProperty ( java.lang.String, java.lang.String )This method is from 'JavaSparkContext' class.
sparkUser ( )This method is from 'JavaSparkContext' class.
startTime ( )This method is from 'JavaSparkContext' class.
stop ( )This method is from 'JavaSparkContext' class.
textFile ( java.lang.String )This method is from 'JavaSparkContext' class.
textFile ( java.lang.String, int )This method is from 'JavaSparkContext' class.
toSparkContext ( JavaSparkContext )1st parameter 'p1' of this method has type 'JavaSparkContext'.
union ( JavaDoubleRDD, java.util.List<JavaDoubleRDD> )This method is from 'JavaSparkContext' class.
union ( JavaPairRDD<K,V>, java.util.List<JavaPairRDD<K,V>> )This method is from 'JavaSparkContext' class.
union ( JavaRDD<T>, java.util.List<JavaRDD<T>> )This method is from 'JavaSparkContext' class.
wholeTextFiles ( java.lang.String )This method is from 'JavaSparkContext' class.
wholeTextFiles ( java.lang.String, int )This method is from 'JavaSparkContext' class.
package org.apache.spark.broadcast
[+] Broadcast<T> (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.Logging. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (2)
broadcast ( T )Return value of this method has type 'Broadcast<T>'.
broadcast ( T, scala.reflect.ClassTag<T> )Return value of this method has type 'Broadcast<T>'.
package org.apache.spark.mapreduce
[+] SparkHadoopMapReduceUtil (1)
| Change | Effect |
---|
1 | This interface has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (3)
newJobContext ( org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.JobID )This abstract method is from 'SparkHadoopMapReduceUtil' interface.
newTaskAttemptContext ( org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.TaskAttemptID )This abstract method is from 'SparkHadoopMapReduceUtil' interface.
newTaskAttemptID ( java.lang.String, int, boolean, int, int )This abstract method is from 'SparkHadoopMapReduceUtil' interface.
package org.apache.spark.rdd
[+] PairRDDFunctions<K,V> (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.mapreduce.SparkHadoopMapReduceUtil. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (1)
rddToPairRDDFunctions ( RDD<scala.Tuple2<K,V>>, scala.reflect.ClassTag<K>, scala.reflect.ClassTag<V>, scala.math.Ordering<K> )Return value of this method has type 'PairRDDFunctions<K,V>'.
spark-sql_2.10-1.4.0.jar
package org.apache.spark.sql
[+] DataFrame (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (120)
agg ( java.util.Map<java.lang.String,java.lang.String> )This method is from 'DataFrame' class.
agg ( Column, Column... )This method is from 'DataFrame' class.
agg ( Column, scala.collection.Seq<Column> )This method is from 'DataFrame' class.
agg ( scala.collection.immutable.Map<java.lang.String,java.lang.String> )This method is from 'DataFrame' class.
agg ( scala.Tuple2<java.lang.String,java.lang.String>, scala.collection.Seq<scala.Tuple2<java.lang.String,java.lang.String>> )This method is from 'DataFrame' class.
apply ( java.lang.String )This method is from 'DataFrame' class.
as ( java.lang.String )This method is from 'DataFrame' class.
as ( scala.Symbol )This method is from 'DataFrame' class.
cache ( )This method is from 'DataFrame' class.
cache ( )This method is from 'DataFrame' class.
coalesce ( int )This method is from 'DataFrame' class.
col ( java.lang.String )This method is from 'DataFrame' class.
collect ( )This method is from 'DataFrame' class.
collect ( )This method is from 'DataFrame' class.
collectAsList ( )This method is from 'DataFrame' class.
columns ( )This method is from 'DataFrame' class.
count ( )This method is from 'DataFrame' class.
cube ( java.lang.String, java.lang.String... )This method is from 'DataFrame' class.
cube ( java.lang.String, scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
cube ( Column... )This method is from 'DataFrame' class.
cube ( scala.collection.Seq<Column> )This method is from 'DataFrame' class.
DataFrame ( SQLContext, catalyst.plans.logical.LogicalPlan )This constructor is from 'DataFrame' class.
DataFrame ( SQLContext, SQLContext.QueryExecution )This constructor is from 'DataFrame' class.
describe ( java.lang.String... )This method is from 'DataFrame' class.
describe ( scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
distinct ( )This method is from 'DataFrame' class.
drop ( java.lang.String )This method is from 'DataFrame' class.
dropDuplicates ( )This method is from 'DataFrame' class.
dropDuplicates ( java.lang.String[ ] )This method is from 'DataFrame' class.
dropDuplicates ( scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
dtypes ( )This method is from 'DataFrame' class.
except ( DataFrame )This method is from 'DataFrame' class.
explain ( )This method is from 'DataFrame' class.
explain ( boolean )This method is from 'DataFrame' class.
explode ( java.lang.String, java.lang.String, scala.Function1<A,scala.collection.TraversableOnce<B>>, scala.reflect.api.TypeTags.TypeTag<B> )This method is from 'DataFrame' class.
explode ( scala.collection.Seq<Column>, scala.Function1<Row,scala.collection.TraversableOnce<A>>, scala.reflect.api.TypeTags.TypeTag<A> )This method is from 'DataFrame' class.
filter ( java.lang.String )This method is from 'DataFrame' class.
filter ( Column )This method is from 'DataFrame' class.
first ( )This method is from 'DataFrame' class.
first ( )This method is from 'DataFrame' class.
flatMap ( scala.Function1<Row,scala.collection.TraversableOnce<R>>, scala.reflect.ClassTag<R> )This method is from 'DataFrame' class.
foreach ( scala.Function1<Row,scala.runtime.BoxedUnit> )This method is from 'DataFrame' class.
foreachPartition ( scala.Function1<scala.collection.Iterator<Row>,scala.runtime.BoxedUnit> )This method is from 'DataFrame' class.
groupBy ( java.lang.String, java.lang.String... )This method is from 'DataFrame' class.
groupBy ( java.lang.String, scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
groupBy ( Column... )This method is from 'DataFrame' class.
groupBy ( scala.collection.Seq<Column> )This method is from 'DataFrame' class.
head ( )This method is from 'DataFrame' class.
head ( int )This method is from 'DataFrame' class.
intersect ( DataFrame )This method is from 'DataFrame' class.
isLocal ( )This method is from 'DataFrame' class.
javaRDD ( )This method is from 'DataFrame' class.
javaToPython ( )This method is from 'DataFrame' class.
join ( DataFrame )This method is from 'DataFrame' class.
join ( DataFrame, java.lang.String )This method is from 'DataFrame' class.
join ( DataFrame, Column )This method is from 'DataFrame' class.
join ( DataFrame, Column, java.lang.String )This method is from 'DataFrame' class.
limit ( int )This method is from 'DataFrame' class.
logicalPlan ( )This method is from 'DataFrame' class.
map ( scala.Function1<Row,R>, scala.reflect.ClassTag<R> )This method is from 'DataFrame' class.
mapPartitions ( scala.Function1<scala.collection.Iterator<Row>,scala.collection.Iterator<R>>, scala.reflect.ClassTag<R> )This method is from 'DataFrame' class.
na ( )This method is from 'DataFrame' class.
numericColumns ( )This method is from 'DataFrame' class.
orderBy ( java.lang.String, java.lang.String... )This method is from 'DataFrame' class.
orderBy ( java.lang.String, scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
orderBy ( Column... )This method is from 'DataFrame' class.
orderBy ( scala.collection.Seq<Column> )This method is from 'DataFrame' class.
DataFrame..logicalPlanToDataFrame ( catalyst.plans.logical.LogicalPlan )This method is from 'DataFrame' class.
persist ( )This method is from 'DataFrame' class.
persist ( )This method is from 'DataFrame' class.
persist ( org.apache.spark.storage.StorageLevel )This method is from 'DataFrame' class.
persist ( org.apache.spark.storage.StorageLevel )This method is from 'DataFrame' class.
printSchema ( )This method is from 'DataFrame' class.
queryExecution ( )This method is from 'DataFrame' class.
randomSplit ( double[ ] )This method is from 'DataFrame' class.
randomSplit ( double[ ], long )This method is from 'DataFrame' class.
randomSplit ( scala.collection.immutable.List<java.lang.Object>, long )This method is from 'DataFrame' class.
rdd ( )This method is from 'DataFrame' class.
registerTempTable ( java.lang.String )This method is from 'DataFrame' class.
repartition ( int )This method is from 'DataFrame' class.
resolve ( java.lang.String )This method is from 'DataFrame' class.
rollup ( java.lang.String, java.lang.String... )This method is from 'DataFrame' class.
rollup ( java.lang.String, scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
rollup ( Column... )This method is from 'DataFrame' class.
rollup ( scala.collection.Seq<Column> )This method is from 'DataFrame' class.
sample ( boolean, double )This method is from 'DataFrame' class.
sample ( boolean, double, long )This method is from 'DataFrame' class.
schema ( )This method is from 'DataFrame' class.
select ( java.lang.String, java.lang.String... )This method is from 'DataFrame' class.
select ( java.lang.String, scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
select ( Column... )This method is from 'DataFrame' class.
select ( scala.collection.Seq<Column> )This method is from 'DataFrame' class.
selectExpr ( java.lang.String... )This method is from 'DataFrame' class.
selectExpr ( scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
show ( )This method is from 'DataFrame' class.
show ( int )This method is from 'DataFrame' class.
showString ( int )This method is from 'DataFrame' class.
sort ( java.lang.String, java.lang.String... )This method is from 'DataFrame' class.
sort ( java.lang.String, scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
sort ( Column... )This method is from 'DataFrame' class.
sort ( scala.collection.Seq<Column> )This method is from 'DataFrame' class.
sqlContext ( )This method is from 'DataFrame' class.
stat ( )This method is from 'DataFrame' class.
take ( int )This method is from 'DataFrame' class.
take ( int )This method is from 'DataFrame' class.
toDF ( )This method is from 'DataFrame' class.
toDF ( java.lang.String... )This method is from 'DataFrame' class.
toDF ( scala.collection.Seq<java.lang.String> )This method is from 'DataFrame' class.
toJavaRDD ( )This method is from 'DataFrame' class.
toJSON ( )This method is from 'DataFrame' class.
toString ( )This method is from 'DataFrame' class.
unionAll ( DataFrame )This method is from 'DataFrame' class.
unpersist ( )This method is from 'DataFrame' class.
unpersist ( )This method is from 'DataFrame' class.
unpersist ( boolean )This method is from 'DataFrame' class.
unpersist ( boolean )This method is from 'DataFrame' class.
where ( Column )This method is from 'DataFrame' class.
withColumn ( java.lang.String, Column )This method is from 'DataFrame' class.
withColumnRenamed ( java.lang.String, java.lang.String )This method is from 'DataFrame' class.
write ( )This method is from 'DataFrame' class.
[+] SQLConf (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (72)
AUTO_BROADCASTJOIN_THRESHOLD ( )This method is from 'SQLConf' class.
autoBroadcastJoinThreshold ( )This method is from 'SQLConf' class.
BROADCAST_TIMEOUT ( )This method is from 'SQLConf' class.
broadcastTimeout ( )This method is from 'SQLConf' class.
CASE_SENSITIVE ( )This method is from 'SQLConf' class.
caseSensitiveAnalysis ( )This method is from 'SQLConf' class.
clear ( )This method is from 'SQLConf' class.
CODEGEN_ENABLED ( )This method is from 'SQLConf' class.
codegenEnabled ( )This method is from 'SQLConf' class.
COLUMN_BATCH_SIZE ( )This method is from 'SQLConf' class.
COLUMN_NAME_OF_CORRUPT_RECORD ( )This method is from 'SQLConf' class.
columnBatchSize ( )This method is from 'SQLConf' class.
columnNameOfCorruptRecord ( )This method is from 'SQLConf' class.
COMPRESS_CACHED ( )This method is from 'SQLConf' class.
DATAFRAME_EAGER_ANALYSIS ( )This method is from 'SQLConf' class.
DATAFRAME_RETAIN_GROUP_COLUMNS ( )This method is from 'SQLConf' class.
DATAFRAME_SELF_JOIN_AUTO_RESOLVE_AMBIGUITY ( )This method is from 'SQLConf' class.
dataFrameEagerAnalysis ( )This method is from 'SQLConf' class.
dataFrameRetainGroupColumns ( )This method is from 'SQLConf' class.
dataFrameSelfJoinAutoResolveAmbiguity ( )This method is from 'SQLConf' class.
DEFAULT_DATA_SOURCE_NAME ( )This method is from 'SQLConf' class.
DEFAULT_SIZE_IN_BYTES ( )This method is from 'SQLConf' class.
defaultDataSourceName ( )This method is from 'SQLConf' class.
defaultSizeInBytes ( )This method is from 'SQLConf' class.
dialect ( )This method is from 'SQLConf' class.
DIALECT ( )This method is from 'SQLConf' class.
EXTERNAL_SORT ( )This method is from 'SQLConf' class.
externalSortEnabled ( )This method is from 'SQLConf' class.
getAllConfs ( )This method is from 'SQLConf' class.
getConf ( java.lang.String )This method is from 'SQLConf' class.
getConf ( java.lang.String, java.lang.String )This method is from 'SQLConf' class.
HIVE_VERIFY_PARTITIONPATH ( )This method is from 'SQLConf' class.
IN_MEMORY_PARTITION_PRUNING ( )This method is from 'SQLConf' class.
inMemoryPartitionPruning ( )This method is from 'SQLConf' class.
isParquetBinaryAsString ( )This method is from 'SQLConf' class.
isParquetINT96AsTimestamp ( )This method is from 'SQLConf' class.
numShufflePartitions ( )This method is from 'SQLConf' class.
ORC_FILTER_PUSHDOWN_ENABLED ( )This method is from 'SQLConf' class.
orcFilterPushDown ( )This method is from 'SQLConf' class.
OUTPUT_COMMITTER_CLASS ( )This method is from 'SQLConf' class.
PARQUET_BINARY_AS_STRING ( )This method is from 'SQLConf' class.
PARQUET_CACHE_METADATA ( )This method is from 'SQLConf' class.
PARQUET_COMPRESSION ( )This method is from 'SQLConf' class.
PARQUET_FILTER_PUSHDOWN_ENABLED ( )This method is from 'SQLConf' class.
PARQUET_INT96_AS_TIMESTAMP ( )This method is from 'SQLConf' class.
PARQUET_USE_DATA_SOURCE_API ( )This method is from 'SQLConf' class.
parquetCompressionCodec ( )This method is from 'SQLConf' class.
parquetFilterPushDown ( )This method is from 'SQLConf' class.
parquetUseDataSourceApi ( )This method is from 'SQLConf' class.
PARTITION_DISCOVERY_ENABLED ( )This method is from 'SQLConf' class.
partitionDiscoveryEnabled ( )This method is from 'SQLConf' class.
SCHEMA_STRING_LENGTH_THRESHOLD ( )This method is from 'SQLConf' class.
schemaStringLengthThreshold ( )This method is from 'SQLConf' class.
setConf ( java.lang.String, java.lang.String )This method is from 'SQLConf' class.
setConf ( java.util.Properties )This method is from 'SQLConf' class.
settings ( )This method is from 'SQLConf' class.
SHUFFLE_PARTITIONS ( )This method is from 'SQLConf' class.
SORTMERGE_JOIN ( )This method is from 'SQLConf' class.
sortMergeJoinEnabled ( )This method is from 'SQLConf' class.
SQLConf ( )This constructor is from 'SQLConf' class.
THRIFTSERVER_POOL ( )This method is from 'SQLConf' class.
THRIFTSERVER_UI_SESSION_LIMIT ( )This method is from 'SQLConf' class.
THRIFTSERVER_UI_STATEMENT_LIMIT ( )This method is from 'SQLConf' class.
UNSAFE_ENABLED ( )This method is from 'SQLConf' class.
unsafeEnabled ( )This method is from 'SQLConf' class.
unsetConf ( java.lang.String )This method is from 'SQLConf' class.
USE_JACKSON_STREAMING_API ( )This method is from 'SQLConf' class.
USE_SQL_SERIALIZER2 ( )This method is from 'SQLConf' class.
useCompression ( )This method is from 'SQLConf' class.
useJacksonStreamingAPI ( )This method is from 'SQLConf' class.
useSqlSerializer2 ( )This method is from 'SQLConf' class.
verifyPartitionPath ( )This method is from 'SQLConf' class.
[+] SQLContext (1)
| Change | Effect |
---|
1 | Removed super-interface org.apache.spark.Logging. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (11)
analyzer ( )This method is from 'SQLContext' class.
cacheTable ( java.lang.String )This method is from 'SQLContext' class.
catalog ( )This method is from 'SQLContext' class.
executePlan ( catalyst.plans.logical.LogicalPlan )This method is from 'SQLContext' class.
executeSql ( java.lang.String )This method is from 'SQLContext' class.
parseSql ( java.lang.String )This method is from 'SQLContext' class.
planner ( )This method is from 'SQLContext' class.
prepareForExecution ( )This method is from 'SQLContext' class.
sparkContext ( )This method is from 'SQLContext' class.
SQLContext ( org.apache.spark.SparkContext )This constructor is from 'SQLContext' class.
uncacheTable ( java.lang.String )This method is from 'SQLContext' class.
[+] SQLContext.QueryExecution (1)
| Change | Effect |
---|
1 | This class became abstract. | A client program may be interrupted by InstantiationError exception. |
[+] affected methods (2)
executePlan ( catalyst.plans.logical.LogicalPlan )Return value of this method has type 'SQLContext.QueryExecution'.
executeSql ( java.lang.String )Return value of this method has type 'SQLContext.QueryExecution'.
package org.apache.spark.sql.execution
[+] GeneratedAggregate (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (1)
GeneratedAggregate ( boolean, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression>, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression>, boolean, SparkPlan )This constructor is from 'GeneratedAggregate' class.
[+] RunnableCommand (1)
| Change | Effect |
---|
1 | This interface has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (3)
children ( )This abstract method is from 'RunnableCommand' interface.
output ( )This abstract method is from 'RunnableCommand' interface.
run ( org.apache.spark.sql.SQLContext )This abstract method is from 'RunnableCommand' interface.
[+] SparkPlan (4)
| Change | Effect |
---|
1 | Method execute ( ) became abstract. | A client program may be interrupted by InstantiationError exception. |
2 | Abstract method doExecute ( ) has been removed from this class. | A client program may be interrupted by NoSuchMethodError exception. |
3 | Removed super-interface org.apache.spark.Logging. | A client program may be interrupted by NoSuchMethodError exception. |
4 | Removed super-interface scala.Serializable. | A client program may be interrupted by NoSuchMethodError exception. |
[+] affected methods (6)
child ( )Return value of this method has type 'SparkPlan'.
Project ( scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression>, SparkPlan )2nd parameter 'child' of this method has type 'SparkPlan'.
execute ( )This method is from 'SparkPlan' abstract class.
outputPartitioning ( )This method is from 'SparkPlan' abstract class.
requiredChildDistribution ( )This method is from 'SparkPlan' abstract class.
SparkPlan ( )This constructor is from 'SparkPlan' abstract class.
package org.apache.spark.sql.sources
[+] BaseRelation (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (5)
BaseRelation ( )This constructor is from 'BaseRelation' abstract class.
needConversion ( )This method is from 'BaseRelation' abstract class.
schema ( )This abstract method is from 'BaseRelation' abstract class.
sizeInBytes ( )This method is from 'BaseRelation' abstract class.
sqlContext ( )This abstract method is from 'BaseRelation' abstract class.
[+] CatalystScan (1)
| Change | Effect |
---|
1 | This interface has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (1)
buildScan ( scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute>, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> )This abstract method is from 'CatalystScan' interface.
[+] InsertableRelation (1)
| Change | Effect |
---|
1 | This interface has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (1)
insert ( org.apache.spark.sql.DataFrame, boolean )This abstract method is from 'InsertableRelation' interface.
[+] LogicalRelation (1)
| Change | Effect |
---|
1 | This class has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (19)
andThen ( scala.Function1<LogicalRelation,A> )This method is from 'LogicalRelation' class.
attributeMap ( )This method is from 'LogicalRelation' class.
canEqual ( java.lang.Object )This method is from 'LogicalRelation' class.
compose ( scala.Function1<A,BaseRelation> )This method is from 'LogicalRelation' class.
copy ( BaseRelation )This method is from 'LogicalRelation' class.
equals ( java.lang.Object )This method is from 'LogicalRelation' class.
hashCode ( )This method is from 'LogicalRelation' class.
LogicalRelation ( BaseRelation )This constructor is from 'LogicalRelation' class.
newInstance ( )This method is from 'LogicalRelation' class.
newInstance ( )This method is from 'LogicalRelation' class.
output ( )This method is from 'LogicalRelation' class.
productArity ( )This method is from 'LogicalRelation' class.
productElement ( int )This method is from 'LogicalRelation' class.
productIterator ( )This method is from 'LogicalRelation' class.
productPrefix ( )This method is from 'LogicalRelation' class.
relation ( )This method is from 'LogicalRelation' class.
sameResult ( org.apache.spark.sql.catalyst.plans.logical.LogicalPlan )This method is from 'LogicalRelation' class.
simpleString ( )This method is from 'LogicalRelation' class.
statistics ( )This method is from 'LogicalRelation' class.
[+] RelationProvider (1)
| Change | Effect |
---|
1 | This interface has been removed. | A client program may be interrupted by NoClassDefFoundError exception. |
[+] affected methods (1)
createRelation ( org.apache.spark.sql.SQLContext, scala.collection.immutable.Map<java.lang.String,java.lang.String> )This abstract method is from 'RelationProvider' interface.
to the top
Problems with Methods, High Severity (1)
spark-sql_2.10-1.4.0.jar, SparkPlan
package org.apache.spark.sql.execution
[+] SparkPlan.execute ( ) : org.apache.spark.rdd.RDD<org.apache.spark.sql.Row> (1)
[mangled: org/apache/spark/sql/execution/SparkPlan.execute:()Lorg/apache/spark/rdd/RDD;]
| Change | Effect |
---|
1 | Method became abstract.
| A client program trying to create an instance of the method's class may be interrupted by InstantiationError exception. |
to the top
Problems with Data Types, Medium Severity (14)
spark-catalyst_2.10-1.4.0.jar
package org.apache.spark.sql.catalyst
[+] SqlParser (1)
| Change | Effect |
---|
1 | Superclass has been changed from AbstractSparkSQLParser to scala.util.parsing.combinator.syntactical.StandardTokenParsers. | 1) Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. 2) A static field from a super-interface of a client class may hide a field (with the same name) inherited from new super-class and cause IncompatibleClassChangeError exception. |
[+] affected methods (24)
andExpression ( )This method is from 'SqlParser' class.
assignAliases ( scala.collection.Seq<expressions.Expression> )This method is from 'SqlParser' class.
cast ( )This method is from 'SqlParser' class.
comparisonExpression ( )This method is from 'SqlParser' class.
dataType ( )This method is from 'SqlParser' class.
direction ( )This method is from 'SqlParser' class.
expression ( )This method is from 'SqlParser' class.
floatLit ( )This method is from 'SqlParser' class.
function ( )This method is from 'SqlParser' class.
insert ( )This method is from 'SqlParser' class.
joinConditions ( )This method is from 'SqlParser' class.
joinedRelation ( )This method is from 'SqlParser' class.
joinType ( )This method is from 'SqlParser' class.
literal ( )This method is from 'SqlParser' class.
ordering ( )This method is from 'SqlParser' class.
orExpression ( )This method is from 'SqlParser' class.
productExpression ( )This method is from 'SqlParser' class.
projection ( )This method is from 'SqlParser' class.
relation ( )This method is from 'SqlParser' class.
relationFactor ( )This method is from 'SqlParser' class.
relations ( )This method is from 'SqlParser' class.
select ( )This method is from 'SqlParser' class.
SqlParser ( )This constructor is from 'SqlParser' class.
termExpression ( )This method is from 'SqlParser' class.
package org.apache.spark.sql.catalyst.analysis
[+] Catalog (3)
| Change | Effect |
---|
1 | Abstract method lookupRelation ( scala.Option<java.lang.String>, java.lang.String, scala.Option<java.lang.String> ) has been added to this interface. | A client program may be interrupted by AbstractMethodError exception. Added abstract method is called in 2nd library version by the method uncacheTable ( java.lang.String ) and may not be implemented by old clients. |
2 | Abstract method registerTable ( scala.Option<java.lang.String>, java.lang.String, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan ) has been added to this interface. | A client program may be interrupted by AbstractMethodError exception. Added abstract method is called in 2nd library version by the method uncacheTable ( java.lang.String ) and may not be implemented by old clients. |
3 | Abstract method unregisterTable ( scala.Option<java.lang.String>, java.lang.String ) has been added to this interface. | A client program may be interrupted by AbstractMethodError exception. Added abstract method is called in 2nd library version by the method uncacheTable ( java.lang.String ) and may not be implemented by old clients. |
[+] affected methods (2)
unregisterAllTables ( )This abstract method is from 'Catalog' interface.
catalog ( )Return value of this method has type 'Catalog'.
package org.apache.spark.sql.catalyst.expressions
[+] And (1)
| Change | Effect |
---|
1 | Superclass has been changed from BinaryExpression to BinaryPredicate. | 1) Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. 2) A static field from a super-interface of a client class may hide a field (with the same name) inherited from new super-class and cause IncompatibleClassChangeError exception. |
[+] affected methods (3)
And ( Expression, Expression )This constructor is from 'And' class.
left ( )This method is from 'And' class.
right ( )This method is from 'And' class.
[+] BoundReference (1)
| Change | Effect |
---|
1 | Superclass has been changed from NamedExpression to Attribute. | 1) Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. 2) A static field from a super-interface of a client class may hide a field (with the same name) inherited from new super-class and cause IncompatibleClassChangeError exception. |
[+] affected methods (1)
ordinal ( )This method is from 'BoundReference' class.
[+] Expression (2)
| Change | Effect |
---|
1 | Abstract method dataType ( ) has been added to this class. | A client program may be interrupted by AbstractMethodError exception. Added abstract method is called in 2nd library version by the method AverageFunction ( Expression, AggregateExpression ) and may not be implemented by old clients. |
2 | Abstract method references ( ) has been added to this class. | A client program may be interrupted by AbstractMethodError exception. Added abstract method is called in 2nd library version by the method canEvaluate ( PredicateHelper, Expression, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan ) and may not be implemented by old clients. |
[+] affected methods (32)
And ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
childrenResolved ( )This method is from 'Expression' abstract class.
Expression ( )This constructor is from 'Expression' abstract class.
foldable ( )This method is from 'Expression' abstract class.
nullable ( )This abstract method is from 'Expression' abstract class.
resolved ( )This method is from 'Expression' abstract class.
GreaterThan ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
GreaterThanOrEqual ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
falseValue ( )Return value of this method has type 'Expression'.
predicate ( )Return value of this method has type 'Expression'.
trueValue ( )Return value of this method has type 'Expression'.
In ( Expression, scala.collection.Seq<Expression> )1st parameter 'value' of this method has type 'Expression'.
value ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
LessThan ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
LessThanOrEqual ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
Not ( Expression )1st parameter 'child' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
Or ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
[+] Or (1)
| Change | Effect |
---|
1 | Superclass has been changed from BinaryExpression to BinaryPredicate. | 1) Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. 2) A static field from a super-interface of a client class may hide a field (with the same name) inherited from new super-class and cause IncompatibleClassChangeError exception. |
[+] affected methods (3)
left ( )This method is from 'Or' class.
Or ( Expression, Expression )This constructor is from 'Or' class.
right ( )This method is from 'Or' class.
spark-core_2.10-1.4.0.jar
package org.apache.spark.api.java
[+] JavaDoubleRDD (1)
| Change | Effect |
---|
1 | Removed super-class AbstractJavaRDDLike<java.lang.Double,JavaDoubleRDD>. | Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. |
[+] affected methods (3)
parallelizeDoubles ( java.util.List<java.lang.Double> )Return value of this method has type 'JavaDoubleRDD'.
parallelizeDoubles ( java.util.List<java.lang.Double>, int )Return value of this method has type 'JavaDoubleRDD'.
union ( JavaDoubleRDD, java.util.List<JavaDoubleRDD> )1st parameter 'first' of this method has type 'JavaDoubleRDD'.
[+] JavaPairRDD<K,V> (1)
| Change | Effect |
---|
1 | Removed super-class AbstractJavaRDDLike<scala.Tuple2<K,V>,JavaPairRDD<K,V>>. | Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. |
[+] affected methods (11)
hadoopFile ( java.lang.String, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V> )Return value of this method has type 'JavaPairRDD<K,V>'.
hadoopFile ( java.lang.String, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V>, int )Return value of this method has type 'JavaPairRDD<K,V>'.
hadoopRDD ( org.apache.hadoop.mapred.JobConf, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V> )Return value of this method has type 'JavaPairRDD<K,V>'.
hadoopRDD ( org.apache.hadoop.mapred.JobConf, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V>, int )Return value of this method has type 'JavaPairRDD<K,V>'.
newAPIHadoopFile ( java.lang.String, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V>, org.apache.hadoop.conf.Configuration )Return value of this method has type 'JavaPairRDD<K,V>'.
newAPIHadoopRDD ( org.apache.hadoop.conf.Configuration, java.lang.Class<F>, java.lang.Class<K>, java.lang.Class<V> )Return value of this method has type 'JavaPairRDD<K,V>'.
parallelizePairs ( java.util.List<scala.Tuple2<K,V>> )Return value of this method has type 'JavaPairRDD<K,V>'.
parallelizePairs ( java.util.List<scala.Tuple2<K,V>>, int )Return value of this method has type 'JavaPairRDD<K,V>'.
sequenceFile ( java.lang.String, java.lang.Class<K>, java.lang.Class<V> )Return value of this method has type 'JavaPairRDD<K,V>'.
sequenceFile ( java.lang.String, java.lang.Class<K>, java.lang.Class<V>, int )Return value of this method has type 'JavaPairRDD<K,V>'.
union ( JavaPairRDD<K,V>, java.util.List<JavaPairRDD<K,V>> )1st parameter 'first' of this method has type 'JavaPairRDD<K,V>'.
[+] JavaRDD<T> (1)
| Change | Effect |
---|
1 | Removed super-class AbstractJavaRDDLike<T,JavaRDD<T>>. | Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. |
[+] affected methods (6)
checkpointFile ( java.lang.String )Return value of this method has type 'JavaRDD<T>'.
objectFile ( java.lang.String )Return value of this method has type 'JavaRDD<T>'.
objectFile ( java.lang.String, int )Return value of this method has type 'JavaRDD<T>'.
parallelize ( java.util.List<T> )Return value of this method has type 'JavaRDD<T>'.
parallelize ( java.util.List<T>, int )Return value of this method has type 'JavaRDD<T>'.
union ( JavaRDD<T>, java.util.List<JavaRDD<T>> )1st parameter 'first' of this method has type 'JavaRDD<T>'.
package org.apache.spark.scheduler
[+] LiveListenerBus (1)
| Change | Effect |
---|
1 | Removed super-class org.apache.spark.util.AsynchronousListenerBus<SparkListener,SparkListenerEvent>. | Access of a client program to the fields or methods of the old super-class may be interrupted by NoSuchFieldError or NoSuchMethodError exceptions. |
[+] affected methods (1)
listenerBus ( )Return value of this method has type 'LiveListenerBus'.
spark-sql_2.10-1.4.0.jar
package org.apache.spark.sql.execution
[+] SparkPlan (1)
| Change | Effect |
---|
1 | Added super-interface com.typesafe.scalalogging.slf4j.Logging. | If abstract methods from an added super-interface must be implemented by client then it may be interrupted by AbstractMethodError exception. Abstract method toString () from the added super-interface is called by the method cachedColumnBuffers.lzycompute ( ) in 2nd library version and may not be implemented by old clients. |
[+] affected methods (6)
child ( )Return value of this method has type 'SparkPlan'.
Project ( scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression>, SparkPlan )2nd parameter 'child' of this method has type 'SparkPlan'.
execute ( )This method is from 'SparkPlan' abstract class.
outputPartitioning ( )This method is from 'SparkPlan' abstract class.
requiredChildDistribution ( )This method is from 'SparkPlan' abstract class.
SparkPlan ( )This constructor is from 'SparkPlan' abstract class.
to the top
Problems with Data Types, Low Severity (6)
spark-catalyst_2.10-1.4.0.jar
package org.apache.spark.sql.catalyst.expressions
[+] AttributeReference (1)
| Change | Effect |
---|
1 | Method dataType ( ) has been moved up type hierarchy to dataType ( ) | Method dataType ( ) will be called instead of dataType ( ) in a client program. |
[+] affected methods (1)
dataType ( )Method 'dataType ( )' will be called instead of this method in a client program.
[+] GenericMutableRow (1)
| Change | Effect |
---|
1 | Method copy ( ) has been moved up type hierarchy to copy ( ) | Method copy ( ) will be called instead of copy ( ) in a client program. |
[+] affected methods (1)
copy ( )Method 'copy ( )' will be called instead of this method in a client program.
[+] Literal (1)
| Change | Effect |
---|
1 | Method dataType ( ) has been moved up type hierarchy to dataType ( ) | Method dataType ( ) will be called instead of dataType ( ) in a client program. |
[+] affected methods (1)
dataType ( )Method 'dataType ( )' will be called instead of this method in a client program.
spark-core_2.10-1.4.0.jar
package org.apache.spark
[+] TaskContext (3)
| Change | Effect |
---|
1 | Abstract method partitionId ( ) became non-abstract. | Some methods in this class may change behavior. |
2 | Abstract method stageId ( ) became non-abstract. | Some methods in this class may change behavior. |
3 | Abstract method taskMetrics ( ) became non-abstract. | Some methods in this class may change behavior. |
[+] affected methods (3)
partitionId ( )This abstract method is from 'TaskContext' abstract class.
stageId ( )This abstract method is from 'TaskContext' abstract class.
taskMetrics ( )This abstract method is from 'TaskContext' abstract class.
to the top
Problems with Methods, Low Severity (3)
spark-core_2.10-1.4.0.jar, TaskContext
package org.apache.spark
[+] TaskContext.partitionId ( ) [abstract] : int (1)
[mangled: org/apache/spark/TaskContext.partitionId:()I]
| Change | Effect |
---|
1 | Method became non-abstract.
| A client program may change behavior. |
[+] TaskContext.stageId ( ) [abstract] : int (1)
[mangled: org/apache/spark/TaskContext.stageId:()I]
| Change | Effect |
---|
1 | Method became non-abstract.
| A client program may change behavior. |
[+] TaskContext.taskMetrics ( ) [abstract] : executor.TaskMetrics (1)
[mangled: org/apache/spark/TaskContext.taskMetrics:()Lorg/apache/spark/executor/TaskMetrics;]
| Change | Effect |
---|
1 | Method became non-abstract.
| A client program may change behavior. |
to the top
Other Changes in Data Types (4)
spark-catalyst_2.10-1.4.0.jar
package org.apache.spark.sql.catalyst.expressions
[+] Expression (1)
| Change | Effect |
---|
1 | Abstract method eval ( Row ) has been added to this class. | No effect. |
[+] affected methods (32)
And ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
childrenResolved ( )This method is from 'Expression' abstract class.
Expression ( )This constructor is from 'Expression' abstract class.
foldable ( )This method is from 'Expression' abstract class.
nullable ( )This abstract method is from 'Expression' abstract class.
resolved ( )This method is from 'Expression' abstract class.
GreaterThan ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
GreaterThanOrEqual ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
falseValue ( )Return value of this method has type 'Expression'.
predicate ( )Return value of this method has type 'Expression'.
trueValue ( )Return value of this method has type 'Expression'.
In ( Expression, scala.collection.Seq<Expression> )1st parameter 'value' of this method has type 'Expression'.
value ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
LessThan ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
LessThanOrEqual ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
child ( )Return value of this method has type 'Expression'.
Not ( Expression )1st parameter 'child' of this method has type 'Expression'.
left ( )Return value of this method has type 'Expression'.
Or ( Expression, Expression )2nd parameter 'right' of this method has type 'Expression'.
right ( )Return value of this method has type 'Expression'.
[+] MutableRow (2)
| Change | Effect |
---|
1 | Abstract method getStringBuilder ( int ) has been added to this interface. | No effect. |
2 | Added super-interface Row. | No effect. |
[+] affected methods (10)
setBoolean ( int, boolean )This abstract method is from 'MutableRow' interface.
setByte ( int, byte )This abstract method is from 'MutableRow' interface.
setDouble ( int, double )This abstract method is from 'MutableRow' interface.
setFloat ( int, float )This abstract method is from 'MutableRow' interface.
setInt ( int, int )This abstract method is from 'MutableRow' interface.
setLong ( int, long )This abstract method is from 'MutableRow' interface.
setNullAt ( int )This abstract method is from 'MutableRow' interface.
setShort ( int, short )This abstract method is from 'MutableRow' interface.
setString ( int, java.lang.String )This abstract method is from 'MutableRow' interface.
update ( int, java.lang.Object )This abstract method is from 'MutableRow' interface.
package org.apache.spark.sql.catalyst.plans.logical
[+] LogicalPlan (1)
| Change | Effect |
---|
1 | Abstract method references ( ) has been added to this class. | No effect. |
[+] affected methods (8)
child ( )Return value of this method has type 'LogicalPlan'.
childrenResolved ( )This method is from 'LogicalPlan' abstract class.
LogicalPlan ( )This constructor is from 'LogicalPlan' abstract class.
resolved ( )This method is from 'LogicalPlan' abstract class.
child ( )Return value of this method has type 'LogicalPlan'.
Subquery ( java.lang.String, LogicalPlan )2nd parameter 'child' of this method has type 'LogicalPlan'.
executePlan ( LogicalPlan )1st parameter 'plan' of this method has type 'LogicalPlan'.
parseSql ( java.lang.String )Return value of this method has type 'LogicalPlan'.
to the top
Java ARchives (3)
spark-catalyst_2.10-1.4.0.jar
spark-core_2.10-1.4.0.jar
spark-sql_2.10-1.4.0.jar
to the top
Generated on Fri Jul 17 23:30:44 2015 for Spark-SQL-on-HBase-1.0.0 by Java API Compliance Checker 1.4.1
A tool for checking backward compatibility of a Java library API