Tokenisasi Kelompok3
Tokenisasi Kelompok3
In [5]:
# -------------------KELOMPOK 3 ------------------------
#Tokenisasi
from nltk.tokenize import RegexpTokenizer
tokenizer = RegexpTokenizer(r'\w+')
filterdText=tokenizer.tokenize('Sarana laboratorium. Laboratorium ilmu komunikasi merup
print(filterdText)
In [1]:
#Stemming
cri
cri
cri
In [6]:
#POS Tagging
import re
import nltk
from nltk import word_tokenize
from nltk import pos_tag
tokens = word_tokenize(clean_words)