The text_with_headers serializer (HDFS sink serializer) allows to save the Flume event headers rather than discarding them. The output format consists of the headers, followed by a space, then the body payload. We would like to drop the body and retain the headers only. For the HBase sink, the "RegexHbaseEventSerializer" allows us to transform the events. But I am unable to find such a provision for the HDFS sink.
相关问题
- UserGroupInformation: No groups available for user
- UnknownHostException while formatting HDFS
- Curl download to HDFS
- How does hive/hadoop assures that each mapper work
- Cannot run Flume because of JAR conflict
相关文章
- Java写文件至HDFS失败
- Is there a way to control number of part files in
- Writing to HDFS : File is overwritten
- Where Mapper output in Hadoop is saved?
- Calculate the percentage of categories in a column
- Content of the fsimage hdfs
- How can I change HDFS replication factor for my Sp
- How to pass Jar files to shell script in OOZIE she
You can set serializer property to header_and_text, which outputs both the headers and the body.
For example: