-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathchunking.py
45 lines (32 loc) · 1.29 KB
/
chunking.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
#Chunking - The idea is to group nouns with the words that are in relation to them.
#Regular expressions used to chunk:
#+ = match 1 or more
#? = match 0 or 1 repetitions.
#* = match 0 or MORE repetitions
#. = Any character except a new line
import nltk
from nltk.corpus import state_union
from nltk.tokenize import PunktSentenceTokenizer
train_text = state_union.raw("2005-GWBush.txt")
sample_text = state_union.raw("2006-GWBush.txt")
custom_sent_tokenizer = PunktSentenceTokenizer(train_text)
tokenized = custom_sent_tokenizer.tokenize(sample_text)
def process_content():
try:
for i in tokenized:
words = nltk.word_tokenize(i)
tagged = nltk.pos_tag(words)
chunkGram = r"""Chunk: {<RB.?>*<VB.?>*<NNP>+<NN>?}"""
chunkParser = nltk.RegexpParser(chunkGram)
chunked = chunkParser.parse(tagged)
print(chunked)
for subtree in chunked.subtrees(filter=lambda t: t.label() == 'Chunk'):
print(subtree)
chunked.draw()
except Exception as e:
print(str(e))
process_content()
#<RB.?>* = "0 or more of any tense of adverb," followed by:
#<VB.?>* = "0 or more of any tense of verb," followed by:
#<NNP>+ = "One or more proper nouns," followed by
#<NN>? = "zero or one singular noun."