flume学习(三)

flume使用实例

实例一

说明

监听客户端的连接然后打印

使用

netcat source---------》监听端口
memory channel  ----》数据通道
logger sink-----------》打印监听到的数据

配置

  • 在conf中新建agent目录
    mkdir -p conf/agent
  • 在agent目录中新建example-net-mem-log.conf
    touch example-net-mem-log.conf
  • 在example-net-mem-log.conf中配置source channel sink

    a1.sources = r1
    a1.channels = c1 
    a1.sinks = k1
    
    a1.sources.r1.type = netcat                              
    a1.sources.r1.bind = dev-hadoop-single.com               
    a1.sources.r1.port = 4444
    
    a1.channels.c1.type                  = memory
        #channel 能够容纳的event数量
    a1.channels.c1.capacity              = 1000
        #channel 一个事务中允许的最大的evnet数量
    a1.channels.c1.transactionCapacity = 100                 
    
    a1.sinks.k1.type = logger
    
    a1.sources.r1.channels = c1                              
    a1.sinks.k1.channel = c1   

    测试

    • 启动flume
      $ flume-ng agent --conf /opt/modules/apache-flume-1.5.0-cdh5.3.6-bin/conf/agent/ --conf-file /opt/modules/apache-flume-1.5.0-cdh5.3.6-bin/conf/agent/example-net-mem-log.conf --name a1 -Dflume.root.logger=INFO,console
    • 使用telnet发送消息

      $ telnet dev-hadoop-single.com 4444
      Trying 192.168.56.101...
      Connected to dev-hadoop-single.com.
      Escape character is '^]'.
      test
      OK
      message
      OK

      flume输出如下

      16/10/19 18:40:23 INFO sink.LoggerSink: Event: { headers:{} body: 74 65 73 74 0D test. }
      16/10/19 18:40:26 INFO sink.LoggerSink: Event: { headers:{} body: 6D 65 73 73 61 67 65 0D message. }

实例二

说明

Nginx + Flume + Hdfs

使用

exec source     1个
memory channel  1个
hdfs sink       1个

配置

flume配置
agent.sources = r1
agent.channels = c1
agent.sinks = k1

##
agent.sources.r1.channels = c1
agent.sinks.k1.channel    = c1

#exec source
agent.sources.r1.type          = exec
agent.sources.r1.command       = tail -F /home/hadoop/access.log
agent.sources.r1.selector.type = replicating


#memory channel
agent.channels.c1.type                = memory
agent.channels.c1.capacity            = 1000
agent.channels.c1.transactionCapacity = 100 
#agent.channels.c1.byteCapacityBufferPercentage = 60
#agent.channels.c1.byteCapacity        = 12800000000
agent.channels.c1.keep-alive          = 60


agent.sinks.k1.type = hdfs
agent.sinks.k1.hdfs.path = hdfs://dev-hadoop-single.com:8020/flume/events-01/%Y-%m-%d
agent.sinks.k1.hdfs.fileType=DataStream
#default:FlumeData
agent.sinks.k1.hdfs.filePrefix = log-spool
agent.sinks.k1.hdfs.fileShuffix = .log
agent.sinks.k1.hdfs.minBlockReplicas = 1 
flume测试

启动flume

$ flume-ng agent --conf /opt/modules/apache-flume-1.5.0-cdh5.3.6-bin/conf/agent/ --conf-file /opt/modules/apache-flume-1.5.0-cdh5.3.6-bin/conf/agent/example-exec-mem-hdfs.conf --name agent -Dflume.root.logger=INFO,console

写入数据
access_log >> access.log
测试结果

16/10/19 19:03:46 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026338.tmp
16/10/19 19:05:36 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026338.tmp
16/10/19 19:05:36 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026338.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026338
16/10/19 19:05:36 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026339.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026339.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026339.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026339
16/10/19 19:05:37 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026340.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026340.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026340.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026340
16/10/19 19:05:37 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026341.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026341.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026341.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026341
16/10/19 19:05:37 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026342.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026342.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026342.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026342
16/10/19 19:05:37 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026343.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026343.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026343.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026343
16/10/19 19:05:37 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026344.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026344.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026344.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026344
16/10/19 19:05:37 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026345.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026345.tmp
16/10/19 19:05:37 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026345.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026345
16/10/19 19:05:38 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026346.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026346.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026346.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026346
16/10/19 19:05:38 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026347.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026347.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026347.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026347
16/10/19 19:05:38 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026348.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026348.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026348.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026348
16/10/19 19:05:38 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026349.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026349.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026349.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026349
16/10/19 19:05:38 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026350.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026350.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026350.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026350
16/10/19 19:05:38 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026351.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026351.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026351.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026351
16/10/19 19:05:38 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026352.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026352.tmp
16/10/19 19:05:38 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026352.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026352
16/10/19 19:05:38 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026353.tmp
16/10/19 19:05:39 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026353.tmp
16/10/19 19:05:39 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026353.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026353
16/10/19 19:05:39 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-01/2016-10-19/log-spool.1476875026354.tmp
资源下载链接为: https://pan.quark.cn/s/22ca96b7bd39 在 IT 领域,文档格式转换是常见需求,尤其在处理多种文件类型时。本文将聚焦于利用 Java 技术栈,尤其是 Apache POI 和 iTextPDF 库,实现 doc、xls(涵盖 Excel 2003 及 Excel 2007+)以及 txt、图片等格式文件向 PDF 的转换,并实现在线浏览功能。 先从 Apache POI 说起,它是一个强大的 Java 库,专注于处理 Microsoft Office 格式文件,比如 doc 和 xls。Apache POI 提供了 HSSF 和 XSSF 两个 API,其中 HSSF 用于读写老版本的 BIFF8 格式(Excel 97-2003),XSSF 则针对新的 XML 格式(Excel 2007+)。这两个 API 均具备读取和写入工作表、单元格、公式、样式等功能。读取 Excel 文件时,可通过创建 HSSFWorkbook 或 XSSFWorkbook 对象来打开相应格式的文件,进而遍历工作簿中的每个 Sheet,获取行和列数据。写入 Excel 文件时,创建新的 Workbook 对象,添加 Sheet、Row 和 Cell,即可构建新 Excel 文件。 再看 iTextPDF,它是一个用于生成和修改 PDF 文档的 Java 库,拥有丰富的 API。创建 PDF 文档时,借助 Document 对象,可定义页面尺寸、边距等属性来定制 PDF 外观。添加内容方面,可使用 Paragraph、List、Table 等元素将文本、列表和表格加入 PDF,图片可通过 Image 类加载插入。iTextPDF 支持多种字体和样式,可设置文本颜色、大小、样式等。此外,iTextPDF 的 TextRenderer 类能将 HTML、
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值