flume学习(四)

本文介绍了一个使用Nginx配置Web服务器及Flume进行日志收集的实例,详细展示了Nginx的基本配置文件,如何启动并测试Nginx服务,以及通过浏览器访问特定URL触发日志记录的过程。最后验证了Flume成功将日志数据写入HDFS。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

接上章

nginx配置
worker_processes  1;

events {
    worker_connections  1024;
}

http {
    include       mime.types;
    default_type  application/octet-stream;

    log_format  main  '$remote_addr - $remote_user [$time_local] "$request" '
                      '$status $body_bytes_sent "$http_referer" '
                      '"$http_user_agent" "$http_x_forwarded_for"';


    log_format  log_format   '$remote_addr^A$msec^A$http_host^A$request_uri';

    access_log /home/hadoop/access.log log_format;
    sendfile        on;
    keepalive_timeout  65;
    #include /etc/nginx/conf.d/*.conf;

server {
    listen       80;
    server_name  hh 0.0.0.0;

    location ~ .*(BfImg)\.(gif)$ {
      default_type image/gif;
      root /usr/local/nginx/www/source;  
   }
}
}
运行nginx
sbin/nginx
测试nginx
# curl http://dev-hadoop-single.com:80
<!DOCTYPE html>
<html>
<head>
<title>Welcome to nginx!</title>
<style>
    body {
        width: 35em;
        margin-left: 100px;
        margin-top:100px;
        font-family: Tahoma, Verdana, Arial, sans-serif;
    }
</style>
</head>
<body>
<h1>Welcome to nginx! backup</h1>
<p>this is the first page</p>
<p>If you see this page, the nginx web server is successfully installed and
working. Further configuration is required.</p>

<p>For online documentation and support please refer to
<a href="http://nginx.org/">nginx.org</a>.<br/>
Commercial support is available at
<a href="http://nginx.com/">nginx.com</a>.</p>

<p><em>Thank you for using nginx.</em></p>
</body>
</html>
启动flume进行测试

浏览器访问
http://dev-hadoop-single.com/a.gif?userid=xxx&ctime=xxxx&url=YYY
进行多次访问
查看/home/hadoop/access.log
“`
# tail -f /home/hadoop/access.log
192.168.56.1^A1476876997.488^Adev-hadoop-single.com^A/a.gif
192.168.56.1^A1476876999.153^Adev-hadoop-single.com^A/a.gif
192.168.56.1^A1476877028.421^Adev-hadoop-single.com^A/a.gif
192.168.56.1^A1476877030.573^Adev-hadoop-single.com^A/a.gif
192.168.56.1^A1476877064.052^Adev-hadoop-single.com^A/a.gif?userid=xxx&ctime=xxxx&url=YYY
192.168.56.1^A1476877065.791^Adev-hadoop-single.com^A/a.gif?userid=xxx&ctime=xxxx&url=YYY
192.168.56.1^A1476877067.537^Adev-hadoop-single.com^A/a.gif?userid=xxx&ctime=xxxx&url=YYY

查看flume日志

16/10/19 19:43:21 INFO hdfs.BucketWriter: Closing hdfs://dev-hadoop-single.com:8020/flume/events-02/2016-10-19/log-spool.1476876160609.tmp
16/10/19 19:43:21 INFO hdfs.BucketWriter: Renaming hdfs://dev-hadoop-single.com:8020/flume/events-02/2016-10-19/log-spool.1476876160609.tmp to hdfs://dev-hadoop-single.com:8020/flume/events-02/2016-10-19/log-spool.1476876160609
16/10/19 19:43:22 INFO hdfs.BucketWriter: Creating hdfs://dev-hadoop-single.com:8020/flume/events-02/2016-10-19/log-spool.1476876160610.tmp

查看hdfs目录

$ hdfs dfs -ls /flume/events-02/2016-10-19
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/opt/modules/hadoop-2.5.0-cdh5.3.6/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/modules/hbase-0.98.6-cdh5.3.6/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
16/10/19 19:46:20 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform… using builtin-java classes where applicable
Found 33 items
-rw-r–r– 1 hadoop supergroup 10397 2016-10-19 19:43 /flume/events-02/2016-10-19/log-spool.1476876160609
-rw-r–r– 1 hadoop supergroup 10439 2016-10-19 19:43 /flume/events-02/2016-10-19/log-spool.1476876160610
-rw-r–r– 1 hadoop supergroup 10439 2016-10-19 19:43 /flume/events-02/2016-10-19/log-spool.1476876160611
-rw-r–r– 1 hadoop supergroup 10439 2016-10-19 19:44 /flume/events-02/2016-10-19/log-spool.1476876160612
“`

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值