jieba分词(聊斋相关分词)
import jieba
excludes = {"什么","一个","我们","那里","你们","如今","说道","知道","起来","姑娘","这里","出来","他们","众人","自己",
"一面","只见","怎么","两个","没有","不是","不知","这个","听见","这样","进来","咱们","告诉","就是",
"东西","袭人","回来","只是","大家","只得","老爷","丫头","这些","不敢","出去","所以","不过","的话","不好",
"姐姐","探春","鸳鸯","一时","不能","过来","心里","如此","今日","银子","几个","答应","二人","还有","只管",
"这么","说话","一回","那边","这话","外头","打发","自然","今儿","罢了","屋里","那些","听说","小丫头","不用","如何"}
txt = open("聊斋.txt","r",encoding='utf-8').read()
'''
不写明路径的话,默认和保存的python文件在同一目录下 注意打开格式是utf-8,这个可以打开txt文件,选择另存为,注意界面右下角的格式
'''
words = jieba.lcut(txt)
'''
利用jieba库将聊斋的所有语句分成词汇
'''
counts = {}
'''
创建的一个空的字典
'''
for word in words:
if len(word) == 1: #如果长度是一,可能是语气词之类的,应该删除掉
continue
else:
counts[word] = counts.get(word,0) + 1
'''
如果字典中没有这个健(名字)则创建,如果有这个健那么就给他的计数加一
[姓名:数量],这里是数量加一
'''
for word in excludes:
del(counts[word])
'''
#这一步:如果列出的干扰词汇在分完词后的所有词汇中那么删除
'''
items = list(counts.items())
'''
把保存[姓名:个数]的字典转换成列表
'''
items.sort(key=lambda x:x[1],reverse = True)
'''
对上述列表进行排序,'True'是降序排列
'''
for i in range(20):
word,count = items[i]
print("{0:<10}{1:>5}".format(word,count))
学号:3039
posted on 2021-11-14 11:27 蓝螃蟹Karry0921 阅读(62) 评论(0) 编辑 收藏 举报