1.使用jieba分词对中文文档进行分词
# -*- coding: utf-8 -*-
# @Time : 17-8-4 上午9:26
# @Author : 未来战士biubiu!!
# @FileName: test.py
# @Software: PyCharm Community Edition
# @Blog :http://blog.youkuaiyun.com/u010105243/article/
# Python3
import jieba
# jieba.load_userdict('userdict.txt')
# 创建停用词list
def stopwordslist(filepath):
stopwords = [line.strip() for line in open(filepath, 'r', encoding='utf-8').readlines()]
return stopwords
# 对句子进行分词
def seg_sentence(sentence):
sentence_seged = jieba.cut(sentence.strip())
stopwords = stopwordslist('./test/stopwords.txt') # 这里加载停用词的路径
outstr = ''
for word in sentence_seged:
if word not in stopwords:
if word != '\t':
outstr += word
outstr += " "
return outstr
inputs = open('./test/input.txt', 'r', encoding='utf-8')
outputs = open('./test/output