def get_notices_crawled(): source_code = requests.get('http://www.ssu.ac.kr/web/kor/plaza_d_01') plain_text = source_code.text soup = BeautifulSoup(plain_text, 'lxml') notices = [] for notice in soup.select('table.bbs-list > tbody > tr.trNotice'): (categories, title) = split_category_title(notice.contents[2].a.string) url = notice.contents[2].a['href'] owner = notice.contents[4].string date = notice.contents[5].string.replace('.', '-') hits = notice.contents[6].string notices.append(Notice.create(title, url, date, hits, categories, owner)) notices.reverse() return notices