Interface HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
-
- All Superinterfaces:
org.apache.camel.builder.EndpointConsumerBuilder
,org.apache.camel.EndpointConsumerResolver
- All Known Subinterfaces:
HdfsEndpointBuilderFactory.AdvancedHdfsEndpointBuilder
- Enclosing interface:
- HdfsEndpointBuilderFactory
public static interface HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder extends org.apache.camel.builder.EndpointConsumerBuilder
Advanced builder for endpoint consumers for the HDFS component.
-
-
Method Summary
All Methods Instance Methods Default Methods Modifier and Type Method Description default HdfsEndpointBuilderFactory.HdfsEndpointConsumerBuilder
basic()
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
blockSize(long blockSize)
The size of the HDFS blocks.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
blockSize(String blockSize)
The size of the HDFS blocks.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
bufferSize(int bufferSize)
The buffer size used by HDFS.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
bufferSize(String bufferSize)
The buffer size used by HDFS.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
checkIdleInterval(int checkIdleInterval)
How often (time in millis) in to run the idle checker background task.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
checkIdleInterval(String checkIdleInterval)
How often (time in millis) in to run the idle checker background task.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
chunkSize(int chunkSize)
When reading a normal file, this is split into chunks producing a message per chunk.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
chunkSize(String chunkSize)
When reading a normal file, this is split into chunks producing a message per chunk.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
compressionCodec(String compressionCodec)
The compression codec to use.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
compressionCodec(HdfsEndpointBuilderFactory.HdfsCompressionCodec compressionCodec)
The compression codec to use.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
compressionType(String compressionType)
The compression type to use (is default not in use).default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
compressionType(HdfsEndpointBuilderFactory.CompressionType compressionType)
The compression type to use (is default not in use).default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
exceptionHandler(String exceptionHandler)
To let the consumer use a custom ExceptionHandler.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
exceptionHandler(org.apache.camel.spi.ExceptionHandler exceptionHandler)
To let the consumer use a custom ExceptionHandler.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
exchangePattern(String exchangePattern)
Sets the exchange pattern when the consumer creates an exchange.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
exchangePattern(org.apache.camel.ExchangePattern exchangePattern)
Sets the exchange pattern when the consumer creates an exchange.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
openedSuffix(String openedSuffix)
When a file is opened for reading/writing the file is renamed with this suffix to avoid to read it during the writing phase.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
pollStrategy(String pollStrategy)
A pluggable org.apache.camel.PollingConsumerPollingStrategy allowing you to provide your custom implementation to control error handling usually occurred during the poll operation before an Exchange have been created and being routed in Camel.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
pollStrategy(org.apache.camel.spi.PollingConsumerPollStrategy pollStrategy)
A pluggable org.apache.camel.PollingConsumerPollingStrategy allowing you to provide your custom implementation to control error handling usually occurred during the poll operation before an Exchange have been created and being routed in Camel.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
readSuffix(String readSuffix)
Once the file has been read is renamed with this suffix to avoid to read it again.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
replication(short replication)
The HDFS replication factor.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
replication(String replication)
The HDFS replication factor.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
splitStrategy(String splitStrategy)
In the current version of Hadoop opening a file in append mode is disabled since it's not very reliable.default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
synchronous(boolean synchronous)
Sets whether synchronous processing should be strictly used, or Camel is allowed to use asynchronous processing (if supported).default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder
synchronous(String synchronous)
Sets whether synchronous processing should be strictly used, or Camel is allowed to use asynchronous processing (if supported).
-
-
-
Method Detail
-
basic
default HdfsEndpointBuilderFactory.HdfsEndpointConsumerBuilder basic()
-
exceptionHandler
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder exceptionHandler(org.apache.camel.spi.ExceptionHandler exceptionHandler)
To let the consumer use a custom ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this option is not in use. By default the consumer will deal with exceptions, that will be logged at WARN or ERROR level and ignored. The option is a:org.apache.camel.spi.ExceptionHandler
type. Group: consumer (advanced)
-
exceptionHandler
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder exceptionHandler(String exceptionHandler)
To let the consumer use a custom ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this option is not in use. By default the consumer will deal with exceptions, that will be logged at WARN or ERROR level and ignored. The option will be converted to aorg.apache.camel.spi.ExceptionHandler
type. Group: consumer (advanced)
-
exchangePattern
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder exchangePattern(org.apache.camel.ExchangePattern exchangePattern)
Sets the exchange pattern when the consumer creates an exchange. The option is a:org.apache.camel.ExchangePattern
type. Group: consumer (advanced)
-
exchangePattern
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder exchangePattern(String exchangePattern)
Sets the exchange pattern when the consumer creates an exchange. The option will be converted to aorg.apache.camel.ExchangePattern
type. Group: consumer (advanced)
-
pollStrategy
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder pollStrategy(org.apache.camel.spi.PollingConsumerPollStrategy pollStrategy)
A pluggable org.apache.camel.PollingConsumerPollingStrategy allowing you to provide your custom implementation to control error handling usually occurred during the poll operation before an Exchange have been created and being routed in Camel. The option is a:org.apache.camel.spi.PollingConsumerPollStrategy
type. Group: consumer (advanced)
-
pollStrategy
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder pollStrategy(String pollStrategy)
A pluggable org.apache.camel.PollingConsumerPollingStrategy allowing you to provide your custom implementation to control error handling usually occurred during the poll operation before an Exchange have been created and being routed in Camel. The option will be converted to aorg.apache.camel.spi.PollingConsumerPollStrategy
type. Group: consumer (advanced)
-
blockSize
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder blockSize(long blockSize)
The size of the HDFS blocks. The option is a:long
type. Default: 67108864 Group: advanced
-
blockSize
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder blockSize(String blockSize)
The size of the HDFS blocks. The option will be converted to along
type. Default: 67108864 Group: advanced
-
bufferSize
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder bufferSize(int bufferSize)
The buffer size used by HDFS. The option is a:int
type. Default: 4096 Group: advanced
-
bufferSize
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder bufferSize(String bufferSize)
The buffer size used by HDFS. The option will be converted to aint
type. Default: 4096 Group: advanced
-
checkIdleInterval
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder checkIdleInterval(int checkIdleInterval)
How often (time in millis) in to run the idle checker background task. This option is only in use if the splitter strategy is IDLE. The option is a:int
type. Default: 500 Group: advanced
-
checkIdleInterval
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder checkIdleInterval(String checkIdleInterval)
How often (time in millis) in to run the idle checker background task. This option is only in use if the splitter strategy is IDLE. The option will be converted to aint
type. Default: 500 Group: advanced
-
chunkSize
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder chunkSize(int chunkSize)
When reading a normal file, this is split into chunks producing a message per chunk. The option is a:int
type. Default: 4096 Group: advanced
-
chunkSize
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder chunkSize(String chunkSize)
When reading a normal file, this is split into chunks producing a message per chunk. The option will be converted to aint
type. Default: 4096 Group: advanced
-
compressionCodec
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder compressionCodec(HdfsEndpointBuilderFactory.HdfsCompressionCodec compressionCodec)
The compression codec to use. The option is a:org.apache.camel.component.hdfs.HdfsCompressionCodec
type. Default: DEFAULT Group: advanced
-
compressionCodec
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder compressionCodec(String compressionCodec)
The compression codec to use. The option will be converted to aorg.apache.camel.component.hdfs.HdfsCompressionCodec
type. Default: DEFAULT Group: advanced
-
compressionType
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder compressionType(HdfsEndpointBuilderFactory.CompressionType compressionType)
The compression type to use (is default not in use). The option is a:org.apache.hadoop.io.SequenceFile$CompressionType
type. Default: NONE Group: advanced
-
compressionType
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder compressionType(String compressionType)
The compression type to use (is default not in use). The option will be converted to aorg.apache.hadoop.io.SequenceFile$CompressionType
type. Default: NONE Group: advanced
-
openedSuffix
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder openedSuffix(String openedSuffix)
When a file is opened for reading/writing the file is renamed with this suffix to avoid to read it during the writing phase. The option is a:java.lang.String
type. Default: opened Group: advanced
-
readSuffix
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder readSuffix(String readSuffix)
Once the file has been read is renamed with this suffix to avoid to read it again. The option is a:java.lang.String
type. Default: read Group: advanced
-
replication
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder replication(short replication)
The HDFS replication factor. The option is a:short
type. Default: 3 Group: advanced
-
replication
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder replication(String replication)
The HDFS replication factor. The option will be converted to ashort
type. Default: 3 Group: advanced
-
splitStrategy
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder splitStrategy(String splitStrategy)
In the current version of Hadoop opening a file in append mode is disabled since it's not very reliable. So, for the moment, it's only possible to create new files. The Camel HDFS endpoint tries to solve this problem in this way: If the split strategy option has been defined, the hdfs path will be used as a directory and files will be created using the configured UuidGenerator. Every time a splitting condition is met, a new file is created. The splitStrategy option is defined as a string with the following syntax: splitStrategy=ST:value,ST:value,... where ST can be: BYTES a new file is created, and the old is closed when the number of written bytes is more than value MESSAGES a new file is created, and the old is closed when the number of written messages is more than value IDLE a new file is created, and the old is closed when no writing happened in the last value milliseconds. The option is a:java.lang.String
type. Group: advanced
-
synchronous
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder synchronous(boolean synchronous)
Sets whether synchronous processing should be strictly used, or Camel is allowed to use asynchronous processing (if supported). The option is a:boolean
type. Default: false Group: advanced
-
synchronous
default HdfsEndpointBuilderFactory.AdvancedHdfsEndpointConsumerBuilder synchronous(String synchronous)
Sets whether synchronous processing should be strictly used, or Camel is allowed to use asynchronous processing (if supported). The option will be converted to aboolean
type. Default: false Group: advanced
-
-