import nltk | |
from sumy.parsers.plaintext import PlaintextParser | |
from sumy.summarizers.luhn import LuhnSummarizer | |
from sumy.nlp.tokenizers import Tokenizer | |
nltk.download('punkt') | |
def summarize(file, SENTENCES_COUNT): | |
sumarizer = LuhnSummarizer() | |
with open(file.name) as f: | |
doc = f.read() | |
sentences_ = [] | |
doc_ = PlaintextParser(doc, Tokenizer("en")).document | |
for sentence in sumarizer(doc_, SENTENCES_COUNT): | |
sentences_.append(str(sentence)) | |
summm_ = " ".join(sentences_) | |
return summm_ |