import asyncio import signal from pysitemap.base_crawler import Crawler def crawler( root_url, out_file, out_format='xml', maxtasks=10, exclude_urls=[], exclude_imgs=[], image_root_urls=[], verifyssl=True, findimages=True, images_this_domain=True, headers=None, timezone_offset=0, changefreq=None, priorities=None): """ run crowler :param root_url: Site root url :param out_file: path to the out file :param out_format: format of out file [xml, txt] :param maxtasks: max count of tasks :param exclude_urls: excludable url paths :param exclude_imgs: excludable img url paths :param image_root_urls: recognized image root urls on the domain :param verifyssl: verify website certificate? :param findimages: Find images references? :param images_this_domain: Find images which refer to this domain only? :param headers: Send these headers in every request :param timezone_offset: timezone offset for lastmod tags :param changefreq: dictionary, where key is site sub url regex, and value is changefreq :param priorities: dictionary, where key is site sub url regex, and value is priority float :return: """ loop = asyncio.get_event_loop() c = Crawler(root_url, out_file=out_file, out_format=out_format, maxtasks=maxtasks, exclude_urls=exclude_urls, exclude_imgs=exclude_imgs, image_root_urls=image_root_urls, verifyssl=verifyssl, findimages=findimages, images_this_domain=images_this_domain, headers=headers, timezone_offset=timezone_offset, changefreq=changefreq, priorities=priorities) loop.run_until_complete(c.run()) try: loop.add_signal_handler(signal.SIGINT, loop.stop) except RuntimeError: pass print('todo_queue:', len(c.todo_queue)) print('busy:', len(c.busy)) print('done:', len(c.done), '; ok:', sum(list(zip(*c.done.values()))[0]) ) print('tasks:', len(c.tasks))