Datasets:
Tasks:
Summarization
#!/usr/bin/env python3 | |
import argparse | |
import os | |
from pathlib import Path | |
if __name__ == "__main__": | |
parser = argparse.ArgumentParser(description="article scraper") | |
parser.add_argument("--dataset", "-d", | |
help="the dataset to be made after articles are scraped (choices: 3000, 5000)", | |
default="3000") | |
args = parser.parse_args() | |
path = os.getcwd() | |
if args.dataset == "3000": | |
path += "/3000.html" | |
if not Path(path).is_file(): | |
os.system("wget https://raw.githubusercontent.com/jamesesguerra/misc/main/3000.html") | |
os.system("sed -i '' 's/5000/3000/' spiders/kami_spider.py") | |
elif args.dataset == "5000": | |
path += "/5000.html" | |
if not Path(path).is_file(): | |
os.system("wget https://raw.githubusercontent.com/jamesesguerra/misc/main/5000.html") | |
os.system("sed -i '' 's/3000/5000/' spiders/kami_spider.py") | |
else: | |
print(f"Dataset {args.dataset} isn't a valid option.") | |
os.system("scrapy crawl kami -O ../csv_files/kami.csv") | |