hadoop - How to filter multiple source data using Apache Flume? -


i using flume handle multiple sources data , stored in hdfs not understand how filter data before storing in hdfs.

you have 2 options:

  • use flume interceptor, check answer here.
  • use streaming based solution(apache spark, apache heron/storm) filter records store in hdfs,

2nd option give more flexibility write different types of streaming patterns. add comment if have more queries.


Comments

Popular posts from this blog

Qt QGraphicsScene is not accessable from QGraphicsView (on Qt 5.6.1) -

php - Cannot override Laravel Spark authentication with own implementation -

What is happening when Matlab is starting a "parallel pool"? -