def get_posts_by_user(username, number, detail, debug): if username: ins_crawler = InsCrawler(has_screen=debug) if settings.login: ins_crawler.login() return ins_crawler.get_user_posts(username, number, detail) else: pass
def get_posts_by_user(username, number, detail, debug): ins_crawler = InsCrawler(has_screen=debug) return ins_crawler.get_user_posts(username, number, detail)
target_path = 'result_username' debug = False current_timestamp = datetime.now().strftime('%Y-%m-%d_%H-%M-%S') target_img_path = os.path.join(target_path, '%s_%s' % (username, current_timestamp)) output_filename = '%s_%s.csv' % (username, current_timestamp) output_path = os.path.join(target_path, output_filename) os.makedirs(target_path, exist_ok=True) os.makedirs(target_img_path, exist_ok=True) ins_crawler = InsCrawler(has_screen=debug) results = ins_crawler.get_user_posts(username, number, detail=False) print('[*] %d results' % len(results)) df = pd.DataFrame(columns=['key', 'caption', 'img_url']) for result in results: # key, caption, img_url if '1 person' in result['caption'] and 'closeup' in result['caption']: parsed = urlparse(result['img_url']) filename = parsed.path.split('/')[-1] result['filename'] = filename urllib.request.urlretrieve(result['img_url'], os.path.join(target_img_path, filename))
def get_posts_by_user(username, number, detail, debug, ins_crawler=None): if ins_crawler is None: ins_crawler = InsCrawler(has_screen=debug) ins_crawler.login() return ins_crawler.get_user_posts(username, number, detail)
def get_posts_by_user(username, number): ins_crawler = InsCrawler() return ins_crawler.get_user_posts(username, number)
def get_posts_by_user(debug): ins_crawler = InsCrawler(has_screen=debug) return ins_crawler.get_user_posts()