def test_first_file(self): fd = FilenameDeduplicator() assert_equals(fd.deduplicate('index.html'), 'index.html')
def test_unicode_filename(self): fd = FilenameDeduplicator() assert_equals(fd.deduplicate(u'\xa3'), u'\xa3') assert_equals(fd.deduplicate(u'\xa3'), u'\xa31')
def test_reset(self): fd = FilenameDeduplicator() assert_equals(fd.deduplicate('index.html'), 'index.html') fd.reset() assert_equals(fd.deduplicate('index.html'), 'index.html')
def test_no_extension(self): fd = FilenameDeduplicator() assert_equals(fd.deduplicate('index'), 'index') assert_equals(fd.deduplicate('index'), 'index1')
def test_second_file(self): fd = FilenameDeduplicator() assert_equals(fd.deduplicate('index.html'), 'index.html') assert_equals(fd.deduplicate('robots.txt'), 'robots.txt')
'name': pkg['name'], 'title': pkg['title'], 'license': LICENSE_LOOKUP.get(pkg['license_id'], ''), 'resources': [], } if pkg['notes']: datapackage['description'] = pkg['notes'] try: package_zip = PackageZip.get_for_package(pkg['id']) datapackage['filepath'] = package_zip.filepath except Exception, ex: pass fd = FilenameDeduplicator() for res in pkg['resources']: archival = Archival.get_for_resource(res['id']) if archival and archival.cache_filepath: # We have archived it, and we have a path. _, resource_id, filename = archival.cache_filepath.rsplit('/', 2) cache_filepath = archival.cache_filepath else: # Try and work out the filename from the URL. try: _, filename = res['url'].rsplit('/', 1) except ValueError: filename = res['id'] cache_filepath = '' filename = fd.deduplicate(filename)