public class QueryExecutionErrors
extends Object
QueryCompilationErrors
.Constructor and Description |
---|
QueryExecutionErrors() |
Modifier and Type | Method and Description |
---|---|
static Throwable |
addFilesWithAbsolutePathUnsupportedError(String commitProtocol) |
static Throwable |
addNewFunctionMismatchedWithFunctionError(String funcName) |
static RuntimeException |
aesCryptoError(String detailMessage) |
static RuntimeException |
aesModeUnsupportedError(String mode,
String padding) |
static Throwable |
alterTableWithDropPartitionAndPurgeUnsupportedError() |
static java.time.DateTimeException |
ansiDateTimeError(java.time.DateTimeException e) |
static java.time.format.DateTimeParseException |
ansiDateTimeParseError(java.time.format.DateTimeParseException e) |
static IllegalArgumentException |
ansiIllegalArgumentError(IllegalArgumentException e) |
static IllegalArgumentException |
ansiIllegalArgumentError(String message) |
static java.text.ParseException |
ansiParseError(java.text.ParseException e) |
static ArithmeticException |
arithmeticOverflowError(ArithmeticException e) |
static ArithmeticException |
arithmeticOverflowError(String message,
String hint,
String errorContext) |
static Throwable |
attributesForTypeUnsupportedError(org.apache.spark.sql.catalyst.ScalaReflection.Schema schema) |
static Throwable |
batchMetadataFileNotFoundError(org.apache.hadoop.fs.Path batchMetadataFile) |
static ArithmeticException |
binaryArithmeticCauseOverflowError(short eval1,
String symbol,
short eval2) |
static Throwable |
buildReaderUnsupportedForFileFormatError(String format) |
static Throwable |
cannotAcquireMemoryToBuildLongHashedRelationError(long size,
long got) |
static Throwable |
cannotAcquireMemoryToBuildUnsafeHashedRelationError() |
static Throwable |
cannotAddMultiPartitionsOnNonatomicPartitionTableError(String tableName) |
static Throwable |
cannotAllocateMemoryToGrowBytesToBytesMapError() |
static Throwable |
cannotBroadcastTableOverMaxTableBytesError(long maxBroadcastTableBytes,
long dataSize) |
static Throwable |
cannotBroadcastTableOverMaxTableRowsError(long maxBroadcastTableRows,
long numRows) |
static Throwable |
cannotBuildHashedRelationLargerThan8GError() |
static Throwable |
cannotBuildHashedRelationWithUniqueKeysExceededError() |
static Throwable |
cannotCastError(DataType from,
DataType to) |
static Throwable |
cannotCastFromNullTypeError(DataType to) |
static Throwable |
cannotCastToDateTimeError(Object value,
DataType from,
DataType to,
String errorContext) |
static ArithmeticException |
cannotChangeDecimalPrecisionError(Decimal value,
int decimalPrecision,
int decimalScale,
String context) |
static Throwable |
cannotClearOutputDirectoryError(org.apache.hadoop.fs.Path staticPrefixPath) |
static Throwable |
cannotClearPartitionDirectoryError(org.apache.hadoop.fs.Path path) |
static Throwable |
cannotCloneOrCopyReadOnlySQLConfError() |
static Throwable |
cannotCompareCostWithTargetCostError(String cost) |
static Throwable |
cannotConvertColumnToJSONError(String name,
DataType dataType) |
static Throwable |
cannotConvertOrcTimestampNTZToTimestampLTZError() |
static Throwable |
cannotConvertOrcTimestampToTimestampNTZError() |
static Throwable |
cannotCreateArrayWithElementsExceedLimitError(long numElements,
String additionalErrorMessage) |
static Throwable |
cannotCreateColumnarReaderError() |
static Throwable |
cannotCreateJDBCTableWithPartitionsError() |
static Throwable |
cannotCreateParquetConverterForDataTypeError(DataType t,
String parquetType) |
static Throwable |
cannotCreateParquetConverterForDecimalTypeError(DecimalType t,
String parquetType) |
static Throwable |
cannotCreateParquetConverterForTypeError(DecimalType t,
String parquetType) |
static Throwable |
cannotCreateStagingDirError(String message,
java.io.IOException e) |
static Throwable |
cannotDropMultiPartitionsOnNonatomicPartitionTableError(String tableName) |
static Throwable |
cannotEvaluateExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression) |
static Throwable |
cannotExecuteStreamingRelationExecError() |
static Throwable |
cannotFetchTablesOfDatabaseError(String dbName,
Exception e) |
static Throwable |
cannotFindConstructorForTypeError(String tpe) |
static Throwable |
cannotFindEncoderForTypeError(String tpe,
org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath) |
static Throwable |
cannotGenerateCodeForExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression) |
static Throwable |
cannotGenerateCodeForUncomparableTypeError(String codeType,
DataType dataType) |
static Throwable |
cannotGenerateCodeForUnsupportedTypeError(DataType dataType) |
static Throwable |
cannotGetEventTimeWatermarkError() |
static Throwable |
cannotGetJdbcTypeError(DataType dt) |
static Throwable |
cannotGetOuterPointerForInnerClassError(Class<?> innerCls) |
static Throwable |
cannotGetSQLConfInSchedulerEventLoopThreadError() |
static Throwable |
cannotHaveCircularReferencesInBeanClassError(Class<?> clazz) |
static Throwable |
cannotHaveCircularReferencesInClassError(String t) |
static Throwable |
cannotInstantiateAbstractCatalogPluginClassError(String name,
String pluginClassName,
Exception e) |
static Throwable |
cannotInterpolateClassIntoCodeBlockError(Object arg) |
static Throwable |
cannotLoadUserDefinedTypeError(String name,
String userClass) |
static Throwable |
cannotMergeClassWithOtherClassError(String className,
String otherClass) |
static Throwable |
cannotMergeDecimalTypesWithIncompatiblePrecisionAndScaleError(int leftPrecision,
int rightPrecision,
int leftScale,
int rightScale) |
static Throwable |
cannotMergeDecimalTypesWithIncompatiblePrecisionError(int leftPrecision,
int rightPrecision) |
static Throwable |
cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale,
int rightScale) |
static Throwable |
cannotMergeIncompatibleDataTypesError(DataType left,
DataType right) |
static Throwable |
cannotMutateReadOnlySQLConfError() |
static Throwable |
cannotParseDecimalError() |
static Throwable |
cannotParseJsonArraysAsStructsError() |
static Throwable |
cannotParseStatisticAsPercentileError(String stats,
NumberFormatException e) |
static Throwable |
cannotParseStringAsDataTypeError(com.fasterxml.jackson.core.JsonParser parser,
com.fasterxml.jackson.core.JsonToken token,
DataType dataType) |
static Throwable |
cannotParseStringAsDataTypeError(String pattern,
String value,
DataType dataType) |
static Throwable |
cannotPurgeAsBreakInternalStateError() |
static Throwable |
cannotReadFilesError(Throwable e,
String path) |
static Throwable |
cannotReadFooterForFileError(org.apache.hadoop.fs.FileStatus file,
RuntimeException e) |
static Throwable |
cannotReadFooterForFileError(org.apache.hadoop.fs.Path file,
java.io.IOException e) |
static Throwable |
cannotRecognizeHiveTypeError(org.apache.spark.sql.catalyst.parser.ParseException e,
String fieldType,
String fieldName) |
static Throwable |
cannotRemovePartitionDirError(org.apache.hadoop.fs.Path partitionPath) |
static Throwable |
cannotRemoveReservedPropertyError(String property) |
static Throwable |
cannotRewriteDomainJoinWithConditionsError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> conditions,
org.apache.spark.sql.catalyst.plans.logical.DomainJoin d) |
static Throwable |
cannotSafelyMergeSerdePropertiesError(scala.collection.immutable.Map<String,String> props1,
scala.collection.immutable.Map<String,String> props2,
scala.collection.immutable.Set<String> conflictKeys) |
static Throwable |
cannotSetTimeoutDurationError() |
static Throwable |
cannotSetTimeoutTimestampError() |
static Throwable |
cannotSpecifyBothJdbcTableNameAndQueryError(String jdbcTableName,
String jdbcQueryString) |
static Throwable |
cannotTerminateGeneratorError(org.apache.spark.sql.catalyst.analysis.UnresolvedGenerator generator) |
static Throwable |
cannotTranslateNonNullValueForFieldError(int pos) |
static Throwable |
cannotUseInvalidJavaIdentifierAsFieldNameError(String fieldName,
org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath) |
static ArithmeticException |
castingCauseOverflowError(Object t,
DataType from,
DataType to) |
static ArithmeticException |
castingCauseOverflowErrorInTableInsert(DataType from,
DataType to,
String columnName) |
static Throwable |
catalogFailToCallPublicNoArgConstructorError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogFailToFindPublicNoArgConstructorError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogPluginClassNotFoundError(String name) |
static Throwable |
catalogPluginClassNotFoundForCatalogError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogPluginClassNotImplementedError(String name,
String pluginClassName) |
static Throwable |
classHasUnexpectedSerializerError(String clsName,
org.apache.spark.sql.catalyst.expressions.Expression objSerializer) |
static RuntimeException |
classUnsupportedByMapObjectsError(Class<?> cls) |
static Throwable |
cleanUpSourceFilesUnsupportedError() |
static Throwable |
commentOnTableUnsupportedError() |
static Throwable |
commitDeniedError(int partId,
long taskId,
int attemptId,
int stageId,
int stageAttempt) |
static Throwable |
compilerError(org.codehaus.commons.compiler.CompileException e) |
static Throwable |
concatArraysWithElementsExceedLimitError(long numberOfElements) |
static Throwable |
concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String className) |
static Throwable |
concurrentQueryInstanceError() |
static Throwable |
constructorNotFoundError(String cls) |
static Throwable |
continuousProcessingUnsupportedByDataSourceError(String sourceName) |
static Throwable |
convertHiveTableToCatalogTableError(SparkException e,
String dbName,
String tableName) |
static Throwable |
copyNullFieldNotAllowedError() |
static RuntimeException |
createArrayWithElementsExceedLimitError(Object count) |
static Throwable |
createStreamingSourceNotSpecifySchemaError() |
static Throwable |
customCollectionClsNotResolvedError() |
static Throwable |
databaseNameConflictWithSystemPreservedDatabaseError(String globalTempDB) |
static Throwable |
dataPathNotSpecifiedError() |
static Throwable |
dataTypeOperationUnsupportedError() |
static Throwable |
dataTypeUnexpectedError(DataType dataType) |
static Throwable |
dataTypeUnsupportedError(String dataType,
String failure) |
static Throwable |
dataTypeUnsupportedYetError(DataType dataType) |
static Throwable |
ddlUnsupportedTemporarilyError(String ddl) |
static Throwable |
decimalPrecisionExceedsMaxPrecisionError(int precision,
int maxPrecision) |
static Throwable |
decorrelateInnerQueryThroughPlanUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) |
static ArithmeticException |
divideByZeroError(String context) |
static Throwable |
doExecuteBroadcastNotImplementedError(String nodeName) |
static Throwable |
doGenCodeOfAliasShouldNotBeCalledError() |
static Throwable |
dropTableWithPurgeUnsupportedError() |
static Throwable |
duplicateMapKeyFoundError(Object key) |
static Throwable |
dynamicPartitionKeyNotAmongWrittenPartitionPathsError(String key) |
static Throwable |
dynamicPartitionOverwriteUnsupportedByTableError(Table table) |
static Throwable |
elementsOfTupleExceedLimitError() |
static Throwable |
emptyOptionError(String optionName) |
static Throwable |
endOfIteratorError() |
static Throwable |
endOfStreamError() |
static Throwable |
exceedMapSizeLimitError(int size) |
static Throwable |
executeBroadcastTimeoutError(long timeout,
scala.Option<java.util.concurrent.TimeoutException> ex) |
static Throwable |
executeCodePathUnsupportedError(String execName) |
static Throwable |
expressionDecodingError(Exception e,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) |
static Throwable |
expressionEncodingError(Exception e,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) |
static Throwable |
failedExecuteUserDefinedFunctionError(String funcCls,
String inputTypes,
String outputType,
Throwable e) |
static Throwable |
failedMergingFieldsError(String leftName,
String rightName,
Throwable e) |
static Throwable |
failedMergingSchemaError(StructType schema,
SparkException e) |
static Throwable |
failedParsingStructTypeError(String raw) |
static Throwable |
failedRenameTempFileError(org.apache.hadoop.fs.Path srcPath,
org.apache.hadoop.fs.Path dstPath) |
static Throwable |
failedSplitSubExpressionError(int length) |
static String |
failedSplitSubExpressionMsg(int length) |
static Throwable |
failedToCastValueToDataTypeForPartitionColumnError(String value,
DataType dataType,
String columnName) |
static String |
failedToCompileMsg(Exception e) |
static org.apache.spark.sql.execution.QueryExecutionException |
failedToExecuteQueryError(Throwable e) |
static Throwable |
failedToFindDataSourceError(String provider,
Throwable error) |
static Throwable |
failedToGenerateEpochMarkerError(Throwable failureReason) |
static Throwable |
failedToInstantiateConstructorForCatalogError(String name,
String pluginClassName,
Exception e) |
static Throwable |
failedToMergeIncompatibleSchemasError(StructType left,
StructType right,
Throwable e) |
static Throwable |
failedToPushRowIntoRowQueueError(String rowQueue) |
static Throwable |
failedToReadDataError(Throwable failureReason) |
static Throwable |
failedToReadDeltaFileError(org.apache.hadoop.fs.Path fileToRead,
String clazz,
int keySize) |
static Throwable |
failedToReadSnapshotFileError(org.apache.hadoop.fs.Path fileToRead,
String clazz,
String message) |
static Throwable |
failToConvertValueToJsonError(Object value,
Class<?> cls,
DataType dataType) |
static Throwable |
failToFormatDateTimeInNewFormatterError(String resultCandidate,
Throwable e) |
static Throwable |
failToParseDateTimeInNewParserError(String s,
Throwable e) |
static Throwable |
failToParseEmptyStringForDataTypeError(DataType dataType) |
static Throwable |
failToParseValueForDataTypeError(com.fasterxml.jackson.core.JsonParser parser,
com.fasterxml.jackson.core.JsonToken token,
DataType dataType) |
static Throwable |
failToRecognizePatternAfterUpgradeError(String pattern,
Throwable e) |
static Throwable |
failToRecognizePatternError(String pattern,
Throwable e) |
static Throwable |
failToSetOriginalACLBackError(String aclEntries,
org.apache.hadoop.fs.Path path,
Throwable e) |
static Throwable |
failToSetOriginalPermissionBackError(org.apache.hadoop.fs.permission.FsPermission permission,
org.apache.hadoop.fs.Path path,
Throwable e) |
static Throwable |
fallbackV1RelationReportsInconsistentSchemaError(StructType v2Schema,
StructType v1Schema) |
static RuntimeException |
fieldCannotBeNullError(int index,
String fieldName) |
static String |
fieldCannotBeNullMsg(int index,
String fieldName) |
static Throwable |
fieldDiffersFromDerivedLocalDateError(java.time.temporal.ChronoField field,
int actual,
int expected,
java.time.LocalDate candidate) |
static Throwable |
fieldIndexOnRowWithoutSchemaError() |
static Throwable |
fileLengthExceedsMaxLengthError(org.apache.hadoop.fs.FileStatus status,
int maxLength) |
static Throwable |
fileNotFoundError(java.io.FileNotFoundException e) |
static Throwable |
flattenArraysWithElementsExceedLimitError(long numberOfElements) |
static Throwable |
foreachWriterAbortedDueToTaskFailureError() |
static Throwable |
foundDuplicateFieldInCaseInsensitiveModeError(String requiredFieldName,
String matchedOrcFields) |
static Throwable |
foundDuplicateFieldInFieldIdLookupModeError(int requiredId,
String matchedFields) |
static Throwable |
foundNullValueForNotNullableFieldError(String name) |
static Throwable |
getParentLoggerNotImplementedError(String className) |
static Throwable |
getPartitionMetadataByFilterError(java.lang.reflect.InvocationTargetException e) |
static Throwable |
getTablesByTypeUnsupportedByHiveVersionError() |
static Throwable |
hiveTableWithAnsiIntervalsError(String tableName) |
static Throwable |
illegalLocationClauseForViewPartitionError() |
static Throwable |
incompatibleDataSourceRegisterError(Throwable e) |
static Throwable |
indexOutOfBoundsOfArrayDataError(int idx) |
static Throwable |
inferSchemaUnsupportedForHiveError() |
static Throwable |
initialTypeNotTargetDataTypeError(DataType dataType,
String target) |
static Throwable |
initialTypeNotTargetDataTypesError(DataType dataType) |
static RuntimeException |
inputExternalRowCannotBeNullError() |
static Throwable |
inputFilterNotFullyConvertibleError(String owner) |
static Throwable |
inputTypeUnsupportedError(DataType dataType) |
static Throwable |
integerOverflowError(String message) |
static Throwable |
internalCompilerError(org.codehaus.janino.InternalCompilerException e) |
static RuntimeException |
invalidAesKeyLengthError(int actualLength) |
static ArrayIndexOutOfBoundsException |
invalidArrayIndexError(int index,
int numElements) |
static Throwable |
invalidBucketFile(String path) |
static ArrayIndexOutOfBoundsException |
invalidElementAtIndexError(int index,
int numElements) |
static java.time.DateTimeException |
invalidFractionOfSecondError() |
static Throwable |
invalidInputInCastToDatetimeError(Object value,
DataType from,
DataType to,
String errorContext) |
static org.apache.spark.SparkNumberFormatException |
invalidInputInCastToNumberError(DataType to,
org.apache.spark.unsafe.types.UTF8String s,
String errorContext) |
static ArrayIndexOutOfBoundsException |
invalidInputIndexError(int index,
int numElements) |
static org.apache.spark.SparkRuntimeException |
invalidInputSyntaxForBooleanError(org.apache.spark.unsafe.types.UTF8String s,
String errorContext) |
static Throwable |
invalidJdbcNumPartitionsError(int n,
String jdbcNumPartitions) |
static Throwable |
invalidJdbcTxnIsolationLevelError(String jdbcTxnIsolationLevel,
String value) |
static Throwable |
invalidKerberosConfigForHiveServer2Error() |
static Throwable |
invalidNamespaceNameError(String[] namespace) |
static Throwable |
invalidNumberFormatError(org.apache.spark.unsafe.types.UTF8String input,
String format) |
static Throwable |
invalidPartitionFilterError() |
static RuntimeException |
invalidPatternError(String funcName,
String pattern) |
static Throwable |
invalidStartIndexError(int numRows,
int startIndex) |
static Throwable |
invalidStreamingOutputModeError(scala.Option<OutputMode> outputMode) |
static Throwable |
invalidUrlError(org.apache.spark.unsafe.types.UTF8String url,
java.net.URISyntaxException e) |
static Throwable |
jobAbortedError(Throwable cause) |
static Throwable |
latestOffsetNotCalledError() |
static Throwable |
legacyCheckpointDirectoryExistsError(org.apache.hadoop.fs.Path checkpointPath,
String legacyCheckpointDir) |
static Throwable |
legacyMetadataPathExistsError(org.apache.hadoop.fs.Path metadataPath,
org.apache.hadoop.fs.Path legacyMetadataPath) |
static RuntimeException |
literalTypeUnsupportedError(Object v) |
static Throwable |
loadHiveClientCausesNoClassDefFoundError(NoClassDefFoundError cnf,
scala.collection.Seq<java.net.URL> execJars,
String key,
java.lang.reflect.InvocationTargetException e) |
static Throwable |
malformedCSVRecordError() |
static Throwable |
malformedJSONError() |
static Throwable |
malformedRecordsDetectedInRecordParsingError(org.apache.spark.sql.catalyst.util.BadRecordException e) |
static Throwable |
malformedRecordsDetectedInSchemaInferenceError(DataType dataType) |
static Throwable |
malformedRecordsDetectedInSchemaInferenceError(Throwable e) |
static Throwable |
mapDataKeyArrayLengthDiffersFromValueArrayLengthError() |
static java.util.NoSuchElementException |
mapKeyNotExistError(Object key,
DataType dataType,
String context) |
static RuntimeException |
mapSizeExceedArraySizeWhenZipMapError(int size) |
static Throwable |
mergeUnsupportedByWindowFunctionError() |
static Throwable |
methodCalledInAnalyzerNotAllowedError() |
static Throwable |
methodNotDeclaredError(String name) |
static Throwable |
methodNotImplementedError(String methodName) |
static Throwable |
microBatchUnsupportedByDataSourceError(String srcName) |
static Throwable |
missingDatabaseLocationError() |
static Throwable |
missingJdbcTableNameAndQueryError(String jdbcTableName,
String jdbcQueryString) |
static Throwable |
multiFailuresInStageMaterializationError(Throwable error) |
static Throwable |
multipleBucketTransformsError() |
static Throwable |
multiplePathsSpecifiedError(scala.collection.Seq<String> allPaths) |
static Throwable |
multiStreamingQueriesUsingPathConcurrentlyError(String path,
org.apache.hadoop.fs.FileAlreadyExistsException e) |
static Throwable |
namespaceNotEmptyError(String[] namespace) |
static Throwable |
negativeValueUnexpectedError(org.apache.spark.sql.catalyst.expressions.Expression frequencyExpression) |
static Throwable |
nestedArraysUnsupportedError() |
static Throwable |
nestedFieldUnsupportedError(String colName) |
static RuntimeException |
noDefaultForDataTypeError(DataType dataType) |
static Throwable |
noRecordsFromEmptyDataReaderError() |
static Throwable |
noSuchElementExceptionError() |
static Throwable |
noSuchElementExceptionError(String key) |
static Throwable |
notADatasourceRDDPartitionError(Partition split) |
static Throwable |
notEnoughMemoryToBuildAndBroadcastTableError(OutOfMemoryError oe) |
static Throwable |
notExpectedUnresolvedEncoderError(org.apache.spark.sql.catalyst.expressions.AttributeReference attr) |
static Throwable |
notOverrideExpectedMethodsError(String className,
String m1,
String m2) |
static Throwable |
notPublicClassError(String name) |
static Throwable |
notSupportNonPrimitiveTypeError() |
static Throwable |
notSupportTypeError(DataType dataType) |
static Throwable |
notUserDefinedTypeError(String name,
String userClass) |
static RuntimeException |
nullAsMapKeyNotAllowedError() |
static Throwable |
nullComparisonResultError() |
static Throwable |
nullLiteralsCannotBeCastedError(String name) |
static <TreeType extends org.apache.spark.sql.catalyst.trees.TreeNode<?>> |
onceStrategyIdempotenceIsBrokenForBatchError(String batchName,
TreeType plan,
TreeType reOptimized) |
static Throwable |
onlySupportDataSourcesProvidingFileFormatError(String providingClass) |
static Throwable |
operatingOnCanonicalizationPlanError() |
static Throwable |
orderedOperationUnsupportedByDataTypeError(DataType dataType) |
static Throwable |
outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str) |
static Throwable |
outputDataTypeUnsupportedByNodeWithoutSerdeError(String nodeName,
DataType dt) |
static ArithmeticException |
overflowInIntegralDivideError(String context) |
static ArithmeticException |
overflowInSumOfDecimalError(String context) |
static Throwable |
overwriteTableByUnsupportedExpressionError(Table table) |
static Throwable |
pairUnsupportedAtFunctionError(org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r1,
org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r2,
String function) |
static Throwable |
paramExceedOneCharError(String paramName) |
static Throwable |
paramIsNotBooleanValueError(String paramName) |
static Throwable |
paramIsNotIntegerError(String paramName,
String value) |
static Throwable |
parentSparkUIToAttachTabNotFoundError() |
static Throwable |
partitionColumnNotFoundInSchemaError(String col,
StructType schema) |
static RuntimeException |
pivotColumnUnsupportedError(Object v,
DataType dataType) |
static Throwable |
pivotNotAfterGroupByUnsupportedError() |
static Throwable |
primaryConstructorNotFoundError(Class<?> cls) |
static Throwable |
primitiveTypesNotSupportedError() |
static Throwable |
readCurrentFileNotFoundError(java.io.FileNotFoundException e) |
static Throwable |
regexGroupIndexExceedGroupCountError(int groupCount,
int groupIndex) |
static Throwable |
regexGroupIndexLessThanZeroError() |
static Throwable |
registeringStreamingQueryListenerError(Exception e) |
static Throwable |
remoteOperationsUnsupportedError() |
static Throwable |
removedClassInSpark2Error(String className,
Throwable e) |
static Throwable |
renameAsExistsPathError(org.apache.hadoop.fs.Path dstPath) |
static Throwable |
renameColumnUnsupportedForOlderMySQLError() |
static Throwable |
renamePathAsExistsPathError(org.apache.hadoop.fs.Path srcPath,
org.apache.hadoop.fs.Path dstPath) |
static Throwable |
renameSrcPathNotFoundError(org.apache.hadoop.fs.Path srcPath) |
static Throwable |
repeatedPivotsUnsupportedError() |
static Throwable |
requestedPartitionsMismatchTablePartitionsError(org.apache.spark.sql.catalyst.catalog.CatalogTable table,
scala.collection.immutable.Map<String,scala.Option<String>> partition) |
static Throwable |
resolveCannotHandleNestedSchema(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) |
static Throwable |
rootConverterReturnNullError() |
static Throwable |
rowLargerThan256MUnsupportedError() |
static Throwable |
ruleIdNotFoundForRuleError(String ruleName) |
static Throwable |
saveModeUnsupportedError(Object saveMode,
boolean pathExists) |
static Throwable |
schemaForTypeUnsupportedError(String tpe) |
static Throwable |
serDeInterfaceNotFoundError(NoClassDefFoundError e) |
static org.apache.spark.SparkUpgradeException |
sparkUpgradeInReadingDatesError(String format,
String config,
String option) |
static org.apache.spark.SparkUpgradeException |
sparkUpgradeInWritingDatesError(String format,
String config) |
static ArrayIndexOutOfBoundsException |
sqlArrayIndexNotStartAtOneError() |
static Throwable |
stateNotDefinedOrAlreadyRemovedError() |
static Throwable |
statisticNotRecognizedError(String stats) |
static Throwable |
streamedOperatorUnsupportedByDataSourceError(String className,
String operator) |
static Throwable |
structuralIntegrityIsBrokenAfterApplyingRuleError(String ruleName,
String batchName) |
static Throwable |
structuralIntegrityOfInputPlanIsBrokenInClassError(String className) |
static Throwable |
subprocessExitedError(int exitCode,
org.apache.spark.util.CircularBuffer stderrBuffer,
Throwable cause) |
static Throwable |
tableStatsNotSpecifiedError() |
static Throwable |
taskFailedWhileWritingRowsError(Throwable cause) |
static ArithmeticException |
timestampAddOverflowError(long micros,
int amount,
String unit) |
static Throwable |
timeZoneIdNotSpecifiedForTimestampTypeError() |
static String |
toDSOption(String option) |
static String |
toSQLConf(String conf) |
static String |
toSQLExpr(org.apache.spark.sql.catalyst.expressions.Expression e) |
static String |
toSQLId(scala.collection.Seq<String> parts) |
static String |
toSQLId(String parts) |
static String |
toSQLStmt(String text) |
static String |
toSQLType(DataType t) |
static String |
toSQLType(String text) |
static String |
toSQLValue(Object v,
DataType t) |
static Throwable |
transactionUnsupportedByJdbcServerError() |
static Throwable |
transformationsAndActionsNotInvokedByDriverError() |
static Throwable |
truncateMultiPartitionUnsupportedError(String tableName) |
static Throwable |
typeUnsupportedError(DataType dataType) |
static Throwable |
unableToCreateDatabaseAsFailedToCreateDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition,
java.io.IOException e) |
static Throwable |
unableToCreatePartitionPathError(org.apache.hadoop.fs.Path partitionPath,
java.io.IOException e) |
static Throwable |
unableToCreateTableAsFailedToCreateDirectoryError(String table,
org.apache.hadoop.fs.Path defaultTableLocation,
java.io.IOException e) |
static Throwable |
unableToDeletePartitionPathError(org.apache.hadoop.fs.Path partitionPath,
java.io.IOException e) |
static Throwable |
unableToDropDatabaseAsFailedToDeleteDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition,
java.io.IOException e) |
static Throwable |
unableToDropTableAsFailedToDeleteDirectoryError(String table,
org.apache.hadoop.fs.Path dir,
java.io.IOException e) |
static Throwable |
unableToRenamePartitionPathError(org.apache.hadoop.fs.Path oldPartPath,
java.io.IOException e) |
static Throwable |
unableToRenameTableAsFailedToRenameDirectoryError(String oldName,
String newName,
org.apache.hadoop.fs.Path oldDir,
java.io.IOException e) |
static ArithmeticException |
unaryMinusCauseOverflowError(int originValue) |
static Throwable |
unexpectedAccumulableUpdateValueError(Object o) |
static Throwable |
unexpectedOperatorInCorrelatedSubquery(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan op,
String pos) |
static RuntimeException |
unexpectedValueForLengthInFunctionError(String prettyName) |
static RuntimeException |
unexpectedValueForStartInFunctionError(String prettyName) |
static Throwable |
unexpectedWindowFunctionFrameError(String frame) |
static Throwable |
unionArrayWithElementsExceedLimitError(int length) |
static Throwable |
unknownColumnError(String unknownColumn) |
static Throwable |
unreachableError(String err) |
static Throwable |
unrecognizedCompressionSchemaTypeIDError(int typeId) |
static Throwable |
unrecognizedFileFormatError(String format) |
static Throwable |
unrecognizedSqlTypeError(int sqlType) |
static Throwable |
unscaledValueTooLargeForPrecisionError() |
static Throwable |
unsupportedArrayElementTypeBasedOnBinaryError(DataType dt) |
static Throwable |
unsupportedArrayTypeError(Class<?> clazz) |
static Throwable |
unsupportedCreateNamespaceCommentError() |
static Throwable |
unsupportedDataTypeError(String dt) |
static Throwable |
unsupportedDropNamespaceRestrictError() |
static Throwable |
unsupportedEncoderError() |
static Throwable |
unsupportedFieldNameError(String fieldName) |
static Throwable |
unsupportedHiveMetastoreVersionError(String version,
String key) |
static Throwable |
unsupportedJavaTypeError(Class<?> clazz) |
static Throwable |
unsupportedJdbcTypeError(String content) |
static Throwable |
unsupportedNaturalJoinTypeError(org.apache.spark.sql.catalyst.plans.JoinType joinType) |
static Throwable |
unsupportedOperandTypeForSizeFunctionError(DataType dataType) |
static Throwable |
unsupportedOperationExceptionError() |
static Throwable |
unsupportedOperationForDataTypeError(DataType dataType) |
static Throwable |
unsupportedPartitionTransformError(Transform transform) |
static Throwable |
unsupportedRemoveNamespaceCommentError() |
static Throwable |
unsupportedRoundingMode(scala.Enumeration.Value roundMode) |
static Throwable |
unsupportedSchemaColumnConvertError(String filePath,
String column,
String logicalType,
String physicalType,
Exception e) |
static Throwable |
unsupportedTableChangeError(IllegalArgumentException e) |
static Throwable |
unsupportedTableWritesError(Identifier ident) |
static Throwable |
unsupportedTypeError(DataType dataType) |
static Throwable |
unsupportedUpdateColumnNullabilityError() |
static Throwable |
unsupportedUserSpecifiedSchemaError() |
static Throwable |
useDictionaryEncodingWhenDictionaryOverflowError() |
static Throwable |
userDefinedTypeNotAnnotatedAndRegisteredError(UserDefinedType<?> udt) |
static Throwable |
userSpecifiedSchemaUnsupportedByDataSourceError(TableProvider provider) |
static Throwable |
valueIsNullError(int index) |
static Throwable |
writePartitionExceedConfigSizeWhenDynamicPartitionError(int numWrittenParts,
int maxDynamicPartitions,
String maxDynamicPartitionsKey) |
static Throwable |
writeUnsupportedForBinaryFileDataSourceError() |
static Throwable |
writingJobAbortedError(Throwable e) |
static Throwable |
writingJobFailedError(Throwable cause) |
public static Throwable cannotEvaluateExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression)
public static Throwable cannotGenerateCodeForExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression)
public static Throwable cannotTerminateGeneratorError(org.apache.spark.sql.catalyst.analysis.UnresolvedGenerator generator)
public static ArithmeticException castingCauseOverflowError(Object t, DataType from, DataType to)
public static ArithmeticException castingCauseOverflowErrorInTableInsert(DataType from, DataType to, String columnName)
public static ArithmeticException cannotChangeDecimalPrecisionError(Decimal value, int decimalPrecision, int decimalScale, String context)
public static Throwable invalidInputInCastToDatetimeError(Object value, DataType from, DataType to, String errorContext)
public static org.apache.spark.SparkRuntimeException invalidInputSyntaxForBooleanError(org.apache.spark.unsafe.types.UTF8String s, String errorContext)
public static org.apache.spark.SparkNumberFormatException invalidInputInCastToNumberError(DataType to, org.apache.spark.unsafe.types.UTF8String s, String errorContext)
public static Throwable cannotCastFromNullTypeError(DataType to)
public static Throwable cannotParseDecimalError()
public static Throwable dataTypeUnsupportedError(String dataType, String failure)
public static Throwable failedExecuteUserDefinedFunctionError(String funcCls, String inputTypes, String outputType, Throwable e)
public static ArithmeticException divideByZeroError(String context)
public static ArrayIndexOutOfBoundsException invalidArrayIndexError(int index, int numElements)
public static ArrayIndexOutOfBoundsException invalidInputIndexError(int index, int numElements)
public static ArrayIndexOutOfBoundsException invalidElementAtIndexError(int index, int numElements)
public static java.util.NoSuchElementException mapKeyNotExistError(Object key, DataType dataType, String context)
public static Throwable inputTypeUnsupportedError(DataType dataType)
public static java.time.DateTimeException invalidFractionOfSecondError()
public static java.time.format.DateTimeParseException ansiDateTimeParseError(java.time.format.DateTimeParseException e)
public static java.time.DateTimeException ansiDateTimeError(java.time.DateTimeException e)
public static java.text.ParseException ansiParseError(java.text.ParseException e)
public static IllegalArgumentException ansiIllegalArgumentError(String message)
public static IllegalArgumentException ansiIllegalArgumentError(IllegalArgumentException e)
public static ArithmeticException overflowInSumOfDecimalError(String context)
public static ArithmeticException overflowInIntegralDivideError(String context)
public static RuntimeException mapSizeExceedArraySizeWhenZipMapError(int size)
public static Throwable copyNullFieldNotAllowedError()
public static RuntimeException literalTypeUnsupportedError(Object v)
public static RuntimeException pivotColumnUnsupportedError(Object v, DataType dataType)
public static RuntimeException noDefaultForDataTypeError(DataType dataType)
public static Throwable doGenCodeOfAliasShouldNotBeCalledError()
public static Throwable orderedOperationUnsupportedByDataTypeError(DataType dataType)
public static Throwable regexGroupIndexLessThanZeroError()
public static Throwable regexGroupIndexExceedGroupCountError(int groupCount, int groupIndex)
public static Throwable invalidUrlError(org.apache.spark.unsafe.types.UTF8String url, java.net.URISyntaxException e)
public static Throwable dataTypeOperationUnsupportedError()
public static Throwable mergeUnsupportedByWindowFunctionError()
public static Throwable dataTypeUnexpectedError(DataType dataType)
public static Throwable typeUnsupportedError(DataType dataType)
public static Throwable negativeValueUnexpectedError(org.apache.spark.sql.catalyst.expressions.Expression frequencyExpression)
public static Throwable addNewFunctionMismatchedWithFunctionError(String funcName)
public static Throwable cannotGenerateCodeForUncomparableTypeError(String codeType, DataType dataType)
public static Throwable cannotGenerateCodeForUnsupportedTypeError(DataType dataType)
public static Throwable cannotInterpolateClassIntoCodeBlockError(Object arg)
public static Throwable customCollectionClsNotResolvedError()
public static RuntimeException classUnsupportedByMapObjectsError(Class<?> cls)
public static RuntimeException nullAsMapKeyNotAllowedError()
public static Throwable methodNotDeclaredError(String name)
public static Throwable constructorNotFoundError(String cls)
public static Throwable primaryConstructorNotFoundError(Class<?> cls)
public static Throwable unsupportedNaturalJoinTypeError(org.apache.spark.sql.catalyst.plans.JoinType joinType)
public static Throwable notExpectedUnresolvedEncoderError(org.apache.spark.sql.catalyst.expressions.AttributeReference attr)
public static Throwable unsupportedEncoderError()
public static Throwable notOverrideExpectedMethodsError(String className, String m1, String m2)
public static Throwable failToConvertValueToJsonError(Object value, Class<?> cls, DataType dataType)
public static Throwable unexpectedOperatorInCorrelatedSubquery(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan op, String pos)
public static Throwable unreachableError(String err)
public static Throwable unsupportedRoundingMode(scala.Enumeration.Value roundMode)
public static Throwable resolveCannotHandleNestedSchema(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan)
public static RuntimeException inputExternalRowCannotBeNullError()
public static String fieldCannotBeNullMsg(int index, String fieldName)
public static RuntimeException fieldCannotBeNullError(int index, String fieldName)
public static Throwable unableToCreateDatabaseAsFailedToCreateDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, java.io.IOException e)
public static Throwable unableToDropDatabaseAsFailedToDeleteDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, java.io.IOException e)
public static Throwable unableToCreateTableAsFailedToCreateDirectoryError(String table, org.apache.hadoop.fs.Path defaultTableLocation, java.io.IOException e)
public static Throwable unableToDeletePartitionPathError(org.apache.hadoop.fs.Path partitionPath, java.io.IOException e)
public static Throwable unableToDropTableAsFailedToDeleteDirectoryError(String table, org.apache.hadoop.fs.Path dir, java.io.IOException e)
public static Throwable unableToRenameTableAsFailedToRenameDirectoryError(String oldName, String newName, org.apache.hadoop.fs.Path oldDir, java.io.IOException e)
public static Throwable unableToCreatePartitionPathError(org.apache.hadoop.fs.Path partitionPath, java.io.IOException e)
public static Throwable unableToRenamePartitionPathError(org.apache.hadoop.fs.Path oldPartPath, java.io.IOException e)
public static Throwable methodNotImplementedError(String methodName)
public static Throwable tableStatsNotSpecifiedError()
public static ArithmeticException arithmeticOverflowError(ArithmeticException e)
public static ArithmeticException arithmeticOverflowError(String message, String hint, String errorContext)
public static ArithmeticException unaryMinusCauseOverflowError(int originValue)
public static ArithmeticException binaryArithmeticCauseOverflowError(short eval1, String symbol, short eval2)
public static String failedSplitSubExpressionMsg(int length)
public static Throwable failedSplitSubExpressionError(int length)
public static String failedToCompileMsg(Exception e)
public static Throwable internalCompilerError(org.codehaus.janino.InternalCompilerException e)
public static Throwable compilerError(org.codehaus.commons.compiler.CompileException e)
public static Throwable unsupportedTableChangeError(IllegalArgumentException e)
public static Throwable notADatasourceRDDPartitionError(Partition split)
public static Throwable dataPathNotSpecifiedError()
public static Throwable createStreamingSourceNotSpecifySchemaError()
public static Throwable streamedOperatorUnsupportedByDataSourceError(String className, String operator)
public static Throwable multiplePathsSpecifiedError(scala.collection.Seq<String> allPaths)
public static Throwable failedToFindDataSourceError(String provider, Throwable error)
public static Throwable removedClassInSpark2Error(String className, Throwable e)
public static Throwable incompatibleDataSourceRegisterError(Throwable e)
public static Throwable unrecognizedFileFormatError(String format)
public static org.apache.spark.SparkUpgradeException sparkUpgradeInReadingDatesError(String format, String config, String option)
public static org.apache.spark.SparkUpgradeException sparkUpgradeInWritingDatesError(String format, String config)
public static Throwable buildReaderUnsupportedForFileFormatError(String format)
public static Throwable jobAbortedError(Throwable cause)
public static Throwable taskFailedWhileWritingRowsError(Throwable cause)
public static Throwable readCurrentFileNotFoundError(java.io.FileNotFoundException e)
public static Throwable saveModeUnsupportedError(Object saveMode, boolean pathExists)
public static Throwable cannotClearOutputDirectoryError(org.apache.hadoop.fs.Path staticPrefixPath)
public static Throwable cannotClearPartitionDirectoryError(org.apache.hadoop.fs.Path path)
public static Throwable failedToCastValueToDataTypeForPartitionColumnError(String value, DataType dataType, String columnName)
public static Throwable endOfStreamError()
public static Throwable fallbackV1RelationReportsInconsistentSchemaError(StructType v2Schema, StructType v1Schema)
public static Throwable noRecordsFromEmptyDataReaderError()
public static Throwable fileNotFoundError(java.io.FileNotFoundException e)
public static Throwable unsupportedSchemaColumnConvertError(String filePath, String column, String logicalType, String physicalType, Exception e)
public static Throwable cannotReadFilesError(Throwable e, String path)
public static Throwable cannotCreateColumnarReaderError()
public static Throwable invalidNamespaceNameError(String[] namespace)
public static Throwable unsupportedPartitionTransformError(Transform transform)
public static Throwable missingDatabaseLocationError()
public static Throwable cannotRemoveReservedPropertyError(String property)
public static Throwable namespaceNotEmptyError(String[] namespace)
public static Throwable writingJobFailedError(Throwable cause)
public static Throwable writingJobAbortedError(Throwable e)
public static Throwable commitDeniedError(int partId, long taskId, int attemptId, int stageId, int stageAttempt)
public static Throwable unsupportedTableWritesError(Identifier ident)
public static Throwable cannotCreateJDBCTableWithPartitionsError()
public static Throwable unsupportedUserSpecifiedSchemaError()
public static Throwable writeUnsupportedForBinaryFileDataSourceError()
public static Throwable fileLengthExceedsMaxLengthError(org.apache.hadoop.fs.FileStatus status, int maxLength)
public static Throwable unsupportedFieldNameError(String fieldName)
public static Throwable cannotSpecifyBothJdbcTableNameAndQueryError(String jdbcTableName, String jdbcQueryString)
public static Throwable missingJdbcTableNameAndQueryError(String jdbcTableName, String jdbcQueryString)
public static Throwable emptyOptionError(String optionName)
public static Throwable invalidJdbcTxnIsolationLevelError(String jdbcTxnIsolationLevel, String value)
public static Throwable cannotGetJdbcTypeError(DataType dt)
public static Throwable unrecognizedSqlTypeError(int sqlType)
public static Throwable unsupportedJdbcTypeError(String content)
public static Throwable unsupportedArrayElementTypeBasedOnBinaryError(DataType dt)
public static Throwable nestedArraysUnsupportedError()
public static Throwable cannotTranslateNonNullValueForFieldError(int pos)
public static Throwable invalidJdbcNumPartitionsError(int n, String jdbcNumPartitions)
public static Throwable transactionUnsupportedByJdbcServerError()
public static Throwable dataTypeUnsupportedYetError(DataType dataType)
public static Throwable unsupportedOperationForDataTypeError(DataType dataType)
public static Throwable inputFilterNotFullyConvertibleError(String owner)
public static Throwable cannotReadFooterForFileError(org.apache.hadoop.fs.Path file, java.io.IOException e)
public static Throwable cannotReadFooterForFileError(org.apache.hadoop.fs.FileStatus file, RuntimeException e)
public static Throwable foundDuplicateFieldInCaseInsensitiveModeError(String requiredFieldName, String matchedOrcFields)
public static Throwable foundDuplicateFieldInFieldIdLookupModeError(int requiredId, String matchedFields)
public static Throwable failedToMergeIncompatibleSchemasError(StructType left, StructType right, Throwable e)
public static Throwable ddlUnsupportedTemporarilyError(String ddl)
public static Throwable operatingOnCanonicalizationPlanError()
public static Throwable executeBroadcastTimeoutError(long timeout, scala.Option<java.util.concurrent.TimeoutException> ex)
public static Throwable cannotCompareCostWithTargetCostError(String cost)
public static Throwable unsupportedDataTypeError(String dt)
public static Throwable notSupportTypeError(DataType dataType)
public static Throwable notSupportNonPrimitiveTypeError()
public static Throwable unsupportedTypeError(DataType dataType)
public static Throwable useDictionaryEncodingWhenDictionaryOverflowError()
public static Throwable endOfIteratorError()
public static Throwable cannotAllocateMemoryToGrowBytesToBytesMapError()
public static Throwable cannotAcquireMemoryToBuildLongHashedRelationError(long size, long got)
public static Throwable cannotAcquireMemoryToBuildUnsafeHashedRelationError()
public static Throwable rowLargerThan256MUnsupportedError()
public static Throwable cannotBuildHashedRelationWithUniqueKeysExceededError()
public static Throwable cannotBuildHashedRelationLargerThan8GError()
public static Throwable failedToPushRowIntoRowQueueError(String rowQueue)
public static Throwable unexpectedWindowFunctionFrameError(String frame)
public static Throwable cannotParseStatisticAsPercentileError(String stats, NumberFormatException e)
public static Throwable statisticNotRecognizedError(String stats)
public static Throwable unknownColumnError(String unknownColumn)
public static Throwable unexpectedAccumulableUpdateValueError(Object o)
public static Throwable unscaledValueTooLargeForPrecisionError()
public static Throwable decimalPrecisionExceedsMaxPrecisionError(int precision, int maxPrecision)
public static Throwable outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str)
public static Throwable unsupportedArrayTypeError(Class<?> clazz)
public static Throwable unsupportedJavaTypeError(Class<?> clazz)
public static Throwable failedParsingStructTypeError(String raw)
public static Throwable failedMergingFieldsError(String leftName, String rightName, Throwable e)
public static Throwable cannotMergeDecimalTypesWithIncompatiblePrecisionAndScaleError(int leftPrecision, int rightPrecision, int leftScale, int rightScale)
public static Throwable cannotMergeDecimalTypesWithIncompatiblePrecisionError(int leftPrecision, int rightPrecision)
public static Throwable cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale, int rightScale)
public static Throwable cannotMergeIncompatibleDataTypesError(DataType left, DataType right)
public static Throwable exceedMapSizeLimitError(int size)
public static Throwable duplicateMapKeyFoundError(Object key)
public static Throwable mapDataKeyArrayLengthDiffersFromValueArrayLengthError()
public static Throwable fieldDiffersFromDerivedLocalDateError(java.time.temporal.ChronoField field, int actual, int expected, java.time.LocalDate candidate)
public static Throwable failToParseDateTimeInNewParserError(String s, Throwable e)
public static Throwable failToFormatDateTimeInNewFormatterError(String resultCandidate, Throwable e)
public static Throwable failToRecognizePatternAfterUpgradeError(String pattern, Throwable e)
public static Throwable failToRecognizePatternError(String pattern, Throwable e)
public static Throwable cannotCastToDateTimeError(Object value, DataType from, DataType to, String errorContext)
public static Throwable registeringStreamingQueryListenerError(Exception e)
public static Throwable concurrentQueryInstanceError()
public static Throwable cannotParseJsonArraysAsStructsError()
public static Throwable cannotParseStringAsDataTypeError(com.fasterxml.jackson.core.JsonParser parser, com.fasterxml.jackson.core.JsonToken token, DataType dataType)
public static Throwable cannotParseStringAsDataTypeError(String pattern, String value, DataType dataType)
public static Throwable failToParseEmptyStringForDataTypeError(DataType dataType)
public static Throwable failToParseValueForDataTypeError(com.fasterxml.jackson.core.JsonParser parser, com.fasterxml.jackson.core.JsonToken token, DataType dataType)
public static Throwable rootConverterReturnNullError()
public static Throwable cannotHaveCircularReferencesInBeanClassError(Class<?> clazz)
public static Throwable cannotHaveCircularReferencesInClassError(String t)
public static Throwable cannotUseInvalidJavaIdentifierAsFieldNameError(String fieldName, org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath)
public static Throwable cannotFindEncoderForTypeError(String tpe, org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath)
public static Throwable attributesForTypeUnsupportedError(org.apache.spark.sql.catalyst.ScalaReflection.Schema schema)
public static Throwable schemaForTypeUnsupportedError(String tpe)
public static Throwable cannotFindConstructorForTypeError(String tpe)
public static Throwable paramExceedOneCharError(String paramName)
public static Throwable paramIsNotIntegerError(String paramName, String value)
public static Throwable paramIsNotBooleanValueError(String paramName)
public static Throwable foundNullValueForNotNullableFieldError(String name)
public static Throwable malformedCSVRecordError()
public static Throwable elementsOfTupleExceedLimitError()
public static Throwable expressionDecodingError(Exception e, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions)
public static Throwable expressionEncodingError(Exception e, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions)
public static Throwable classHasUnexpectedSerializerError(String clsName, org.apache.spark.sql.catalyst.expressions.Expression objSerializer)
public static Throwable cannotGetOuterPointerForInnerClassError(Class<?> innerCls)
public static Throwable userDefinedTypeNotAnnotatedAndRegisteredError(UserDefinedType<?> udt)
public static Throwable unsupportedOperandTypeForSizeFunctionError(DataType dataType)
public static RuntimeException unexpectedValueForStartInFunctionError(String prettyName)
public static RuntimeException unexpectedValueForLengthInFunctionError(String prettyName)
public static ArrayIndexOutOfBoundsException sqlArrayIndexNotStartAtOneError()
public static Throwable concatArraysWithElementsExceedLimitError(long numberOfElements)
public static Throwable flattenArraysWithElementsExceedLimitError(long numberOfElements)
public static RuntimeException createArrayWithElementsExceedLimitError(Object count)
public static Throwable unionArrayWithElementsExceedLimitError(int length)
public static Throwable initialTypeNotTargetDataTypeError(DataType dataType, String target)
public static Throwable initialTypeNotTargetDataTypesError(DataType dataType)
public static Throwable cannotConvertColumnToJSONError(String name, DataType dataType)
public static Throwable malformedRecordsDetectedInSchemaInferenceError(Throwable e)
public static Throwable malformedJSONError()
public static Throwable malformedRecordsDetectedInSchemaInferenceError(DataType dataType)
public static Throwable cannotRewriteDomainJoinWithConditionsError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> conditions, org.apache.spark.sql.catalyst.plans.logical.DomainJoin d)
public static Throwable decorrelateInnerQueryThroughPlanUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan)
public static Throwable methodCalledInAnalyzerNotAllowedError()
public static Throwable cannotSafelyMergeSerdePropertiesError(scala.collection.immutable.Map<String,String> props1, scala.collection.immutable.Map<String,String> props2, scala.collection.immutable.Set<String> conflictKeys)
public static Throwable pairUnsupportedAtFunctionError(org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r1, org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r2, String function)
public static <TreeType extends org.apache.spark.sql.catalyst.trees.TreeNode<?>> Throwable onceStrategyIdempotenceIsBrokenForBatchError(String batchName, TreeType plan, TreeType reOptimized)
public static Throwable structuralIntegrityOfInputPlanIsBrokenInClassError(String className)
public static Throwable structuralIntegrityIsBrokenAfterApplyingRuleError(String ruleName, String batchName)
public static Throwable ruleIdNotFoundForRuleError(String ruleName)
public static Throwable cannotCreateArrayWithElementsExceedLimitError(long numElements, String additionalErrorMessage)
public static Throwable indexOutOfBoundsOfArrayDataError(int idx)
public static Throwable malformedRecordsDetectedInRecordParsingError(org.apache.spark.sql.catalyst.util.BadRecordException e)
public static Throwable remoteOperationsUnsupportedError()
public static Throwable invalidKerberosConfigForHiveServer2Error()
public static Throwable parentSparkUIToAttachTabNotFoundError()
public static Throwable inferSchemaUnsupportedForHiveError()
public static Throwable requestedPartitionsMismatchTablePartitionsError(org.apache.spark.sql.catalyst.catalog.CatalogTable table, scala.collection.immutable.Map<String,scala.Option<String>> partition)
public static Throwable dynamicPartitionKeyNotAmongWrittenPartitionPathsError(String key)
public static Throwable cannotRemovePartitionDirError(org.apache.hadoop.fs.Path partitionPath)
public static Throwable cannotCreateStagingDirError(String message, java.io.IOException e)
public static Throwable serDeInterfaceNotFoundError(NoClassDefFoundError e)
public static Throwable convertHiveTableToCatalogTableError(SparkException e, String dbName, String tableName)
public static Throwable cannotRecognizeHiveTypeError(org.apache.spark.sql.catalyst.parser.ParseException e, String fieldType, String fieldName)
public static Throwable getTablesByTypeUnsupportedByHiveVersionError()
public static Throwable dropTableWithPurgeUnsupportedError()
public static Throwable alterTableWithDropPartitionAndPurgeUnsupportedError()
public static Throwable invalidPartitionFilterError()
public static Throwable getPartitionMetadataByFilterError(java.lang.reflect.InvocationTargetException e)
public static Throwable unsupportedHiveMetastoreVersionError(String version, String key)
public static Throwable loadHiveClientCausesNoClassDefFoundError(NoClassDefFoundError cnf, scala.collection.Seq<java.net.URL> execJars, String key, java.lang.reflect.InvocationTargetException e)
public static Throwable cannotFetchTablesOfDatabaseError(String dbName, Exception e)
public static Throwable illegalLocationClauseForViewPartitionError()
public static Throwable renamePathAsExistsPathError(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath)
public static Throwable renameAsExistsPathError(org.apache.hadoop.fs.Path dstPath)
public static Throwable renameSrcPathNotFoundError(org.apache.hadoop.fs.Path srcPath)
public static Throwable failedRenameTempFileError(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath)
public static Throwable legacyMetadataPathExistsError(org.apache.hadoop.fs.Path metadataPath, org.apache.hadoop.fs.Path legacyMetadataPath)
public static Throwable partitionColumnNotFoundInSchemaError(String col, StructType schema)
public static Throwable stateNotDefinedOrAlreadyRemovedError()
public static Throwable cannotSetTimeoutDurationError()
public static Throwable cannotGetEventTimeWatermarkError()
public static Throwable cannotSetTimeoutTimestampError()
public static Throwable batchMetadataFileNotFoundError(org.apache.hadoop.fs.Path batchMetadataFile)
public static Throwable multiStreamingQueriesUsingPathConcurrentlyError(String path, org.apache.hadoop.fs.FileAlreadyExistsException e)
public static Throwable addFilesWithAbsolutePathUnsupportedError(String commitProtocol)
public static Throwable microBatchUnsupportedByDataSourceError(String srcName)
public static Throwable cannotExecuteStreamingRelationExecError()
public static Throwable invalidStreamingOutputModeError(scala.Option<OutputMode> outputMode)
public static Throwable catalogPluginClassNotFoundError(String name)
public static Throwable catalogPluginClassNotImplementedError(String name, String pluginClassName)
public static Throwable catalogPluginClassNotFoundForCatalogError(String name, String pluginClassName, Exception e)
public static Throwable catalogFailToFindPublicNoArgConstructorError(String name, String pluginClassName, Exception e)
public static Throwable catalogFailToCallPublicNoArgConstructorError(String name, String pluginClassName, Exception e)
public static Throwable cannotInstantiateAbstractCatalogPluginClassError(String name, String pluginClassName, Exception e)
public static Throwable failedToInstantiateConstructorForCatalogError(String name, String pluginClassName, Exception e)
public static Throwable noSuchElementExceptionError()
public static Throwable noSuchElementExceptionError(String key)
public static Throwable cannotMutateReadOnlySQLConfError()
public static Throwable cannotCloneOrCopyReadOnlySQLConfError()
public static Throwable cannotGetSQLConfInSchedulerEventLoopThreadError()
public static Throwable unsupportedOperationExceptionError()
public static Throwable nullLiteralsCannotBeCastedError(String name)
public static Throwable notUserDefinedTypeError(String name, String userClass)
public static Throwable cannotLoadUserDefinedTypeError(String name, String userClass)
public static Throwable timeZoneIdNotSpecifiedForTimestampTypeError()
public static Throwable notPublicClassError(String name)
public static Throwable primitiveTypesNotSupportedError()
public static Throwable fieldIndexOnRowWithoutSchemaError()
public static Throwable valueIsNullError(int index)
public static Throwable onlySupportDataSourcesProvidingFileFormatError(String providingClass)
public static Throwable failToSetOriginalPermissionBackError(org.apache.hadoop.fs.permission.FsPermission permission, org.apache.hadoop.fs.Path path, Throwable e)
public static Throwable failToSetOriginalACLBackError(String aclEntries, org.apache.hadoop.fs.Path path, Throwable e)
public static Throwable multiFailuresInStageMaterializationError(Throwable error)
public static Throwable unrecognizedCompressionSchemaTypeIDError(int typeId)
public static Throwable getParentLoggerNotImplementedError(String className)
public static Throwable cannotCreateParquetConverterForTypeError(DecimalType t, String parquetType)
public static Throwable cannotCreateParquetConverterForDecimalTypeError(DecimalType t, String parquetType)
public static Throwable cannotCreateParquetConverterForDataTypeError(DataType t, String parquetType)
public static Throwable cannotAddMultiPartitionsOnNonatomicPartitionTableError(String tableName)
public static Throwable userSpecifiedSchemaUnsupportedByDataSourceError(TableProvider provider)
public static Throwable cannotDropMultiPartitionsOnNonatomicPartitionTableError(String tableName)
public static Throwable truncateMultiPartitionUnsupportedError(String tableName)
public static Throwable overwriteTableByUnsupportedExpressionError(Table table)
public static Throwable dynamicPartitionOverwriteUnsupportedByTableError(Table table)
public static Throwable failedMergingSchemaError(StructType schema, SparkException e)
public static Throwable cannotBroadcastTableOverMaxTableRowsError(long maxBroadcastTableRows, long numRows)
public static Throwable cannotBroadcastTableOverMaxTableBytesError(long maxBroadcastTableBytes, long dataSize)
public static Throwable notEnoughMemoryToBuildAndBroadcastTableError(OutOfMemoryError oe)
public static Throwable executeCodePathUnsupportedError(String execName)
public static Throwable cannotMergeClassWithOtherClassError(String className, String otherClass)
public static Throwable continuousProcessingUnsupportedByDataSourceError(String sourceName)
public static Throwable failedToReadDataError(Throwable failureReason)
public static Throwable failedToGenerateEpochMarkerError(Throwable failureReason)
public static Throwable foreachWriterAbortedDueToTaskFailureError()
public static Throwable integerOverflowError(String message)
public static Throwable failedToReadDeltaFileError(org.apache.hadoop.fs.Path fileToRead, String clazz, int keySize)
public static Throwable failedToReadSnapshotFileError(org.apache.hadoop.fs.Path fileToRead, String clazz, String message)
public static Throwable cannotPurgeAsBreakInternalStateError()
public static Throwable cleanUpSourceFilesUnsupportedError()
public static Throwable latestOffsetNotCalledError()
public static Throwable legacyCheckpointDirectoryExistsError(org.apache.hadoop.fs.Path checkpointPath, String legacyCheckpointDir)
public static Throwable subprocessExitedError(int exitCode, org.apache.spark.util.CircularBuffer stderrBuffer, Throwable cause)
public static Throwable outputDataTypeUnsupportedByNodeWithoutSerdeError(String nodeName, DataType dt)
public static Throwable invalidStartIndexError(int numRows, int startIndex)
public static Throwable concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String className)
public static Throwable doExecuteBroadcastNotImplementedError(String nodeName)
public static Throwable databaseNameConflictWithSystemPreservedDatabaseError(String globalTempDB)
public static Throwable commentOnTableUnsupportedError()
public static Throwable unsupportedUpdateColumnNullabilityError()
public static Throwable renameColumnUnsupportedForOlderMySQLError()
public static org.apache.spark.sql.execution.QueryExecutionException failedToExecuteQueryError(Throwable e)
public static Throwable nestedFieldUnsupportedError(String colName)
public static Throwable transformationsAndActionsNotInvokedByDriverError()
public static Throwable repeatedPivotsUnsupportedError()
public static Throwable pivotNotAfterGroupByUnsupportedError()
public static RuntimeException invalidAesKeyLengthError(int actualLength)
public static RuntimeException aesModeUnsupportedError(String mode, String padding)
public static RuntimeException aesCryptoError(String detailMessage)
public static Throwable hiveTableWithAnsiIntervalsError(String tableName)
public static Throwable cannotConvertOrcTimestampToTimestampNTZError()
public static Throwable cannotConvertOrcTimestampNTZToTimestampLTZError()
public static Throwable writePartitionExceedConfigSizeWhenDynamicPartitionError(int numWrittenParts, int maxDynamicPartitions, String maxDynamicPartitionsKey)
public static Throwable invalidNumberFormatError(org.apache.spark.unsafe.types.UTF8String input, String format)
public static Throwable multipleBucketTransformsError()
public static Throwable unsupportedCreateNamespaceCommentError()
public static Throwable unsupportedRemoveNamespaceCommentError()
public static Throwable unsupportedDropNamespaceRestrictError()
public static ArithmeticException timestampAddOverflowError(long micros, int amount, String unit)
public static Throwable nullComparisonResultError()
public static Throwable invalidBucketFile(String path)
public static RuntimeException invalidPatternError(String funcName, String pattern)
public static String toSQLValue(Object v, DataType t)
public static String toSQLStmt(String text)
public static String toSQLId(scala.collection.Seq<String> parts)
public static String toSQLId(String parts)
public static String toSQLType(DataType t)
public static String toSQLType(String text)
public static String toSQLConf(String conf)
public static String toDSOption(String option)
public static String toSQLExpr(org.apache.spark.sql.catalyst.expressions.Expression e)