Ejemplo n.º 1
0
def download_images(query, tag):
    if os.path.isdir(IMG_DIR + "/" + tag) and len(
        [f for f in os.listdir(IMG_DIR + "/" + tag)]) >= 3:
        print("It looks like you've already downloaded images for '" + query +
              "'.")
        print("Skipping download.")
        print("To re-download images, remove all images in the " + query +
              " folder.")
        return

    settings = {}
    settings["keywords"] = query
    settings["limit"] = 30  # how many images to download
    settings["output_directory"] = IMG_DIR
    settings["no_numbering"] = True
    settings["image_directory"] = tag
    settings["format"] = "jpg"
    settings["color_type"] = "full-color"
    settings["size"] = "medium"
    settings["type"] = "photo"

    downloader = googleimagesdownload()
    downloader.download(settings)

    # make sure images can be loaded
    imglist = [
        IMG_DIR + "/" + tag + "/" + f for f in os.listdir(IMG_DIR + "/" + tag)
        if not f.startswith('.')
    ]
    for img in imglist:
        try:
            Image.open(img)
        except OSError:
            print(img, "can't be opened--will delete.")
            os.remove(img)
Ejemplo n.º 2
0
def main():
    dloader = googleimagesdownload()
    dloader.download({
        "keywords": "cake round",
        "limit": 70000,
        "chromedriver": "/usr/bin/chromedriver",
        "no_numbering": True,
        "output_directory": "data_raw"
    })
    print("test")
Ejemplo n.º 3
0
def img():
    global E2
    global E3
    string = E2.get()
    num = E3.get()
    response = googleimagesdownload()
    absolute_image_paths = response.download({
        "keywords": string,
        "limit": num
    })
    print("\n downloaded!!")
Ejemplo n.º 4
0
def main():
    ssl._create_default_https_context = ssl._create_unverified_context
    response = google_images_download.googleimagesdownload()

    # download by keyword
    arguments = {
        'keywords': 'kermit',
        'limit': 100,
        'output_directory': 'data/google-images'
    }
    paths = response.download(arguments)
    print(paths)
Ejemplo n.º 5
0
def im_id(s):
    response = google_images_download.googleimagesdownload(
    )  #class instantiation
    arguments = {"keywords": s, "limit": 1}
    paths = response.download(arguments)[s][0]
    url = "https://dialogs.yandex.net/api/v1/skills/42bf7345-43a8-4ea7-8a6a-8f73f885fc13/images"
    headers = {
        'Authorization': "OAuth AQAAAAAc4bhNAAT7o9I-T9hCMkPXoL0nwVPPuDc"
    }
    response = requests.request("POST",
                                url,
                                files={'file': open(paths, 'rb')},
                                headers=headers)
    return (response.json()['image']['id'])
Ejemplo n.º 6
0
def imageCrawling(keyword, dir):
    response = gi.googleimagesdownload()
    format = "jpg"
    arguments = {
        "keywords": keyword,  # 검색 키워드
        "limit": 500,  # 크롤링 이미지 수
        "print_urls": True,  # 이미지 url 출력
        "no_directory": True,  #
        'output_directory': dir,  # 크롤링 이미지를 저장할 폴더
        'format': format
    }

    paths = response.download(arguments)
    print(paths)
Ejemplo n.º 7
0
def imageCrawling(keyword, dir):
    response = googleimagesdownload()

    arguments = {
        "keywords": keyword,
        "format": "jpg",
        "limit": 200,  #크롤링 이미지수
        "print_urls": True,  #이미지 url 출력
        "no_directory": True,
        'output_directory': dir
    }  #크롤링 이미지 저장 폴더

    paths = response.download(arguments)
    print paths
Ejemplo n.º 8
0
    def run(self):
        print("Searching for " + self.keyword)
        response = google_images_download.googleimagesdownload(
        )  #class instantiation
        arguments = {
            "keywords": self.keyword,
            "limit": 100,
            "print_urls": False,
            "print_paths": True,
            "no_directory": True,
            "no_download": True,
            "type": "photo",
            "format": "jpg"
        }
        arguments['size'] = self.size
        if (self.color != 'none'):
            arguments['color'] = self.color
        #"color_type":"black-and-white"} "usage_rights":"labeled-for-noncommercial-reuse-with-modification", "size": ">1024*768", "language":"French"   #creating list of arguments
        paths = response.download(
            arguments)  #passing the arguments to the function
        n = len(paths[self.keyword])
        if (n == 0):
            print("Searching for " + self.keyword + " - not found")
        else:
            #RANDOM IMAGES
            if (self.mode == 'random'):
                r = random.randint(0, n - 1)
                print("Searching for " + self.keyword + " + found : " +
                      str(r) + " / " + str(n))
                url = paths[self.keyword][r]
                self.osc_client.send('/keyword ' +
                                     unicode(self.keyword).encode('utf-8'))
                self.osc_client.send('/result ' + str(r + 1) + ' ' + str(n))
                self.osc_client.send('/path ' + unicode(url).encode('utf-8'))

            #ALL IMAGES
            elif (self.mode == 'all'):
                print("Searching for " + self.keyword + " + found : " + str(n))
                self.osc_client.send('/keyword ' +
                                     unicode(self.keyword).encode('utf-8'))
                for x in range(0, n):
                    url = paths[self.keyword][x]
                    self.osc_client.send('/result ' + str(x + 1) + ' ' +
                                         str(n))
                    self.osc_client.send('/path ' +
                                         unicode(url).encode('utf-8'))
def test_download_images_to_default_location():
    start_time = time.time()
    argumnets = {"keywords": "Polar bears", "limit": 5, "print_urls": False}
    try:
        temp = argumnets['output_folder']
    except KeyError:
        pass
    else:
        assert False, "This test checks download to default location yet an output folder was provided"

    output_folder_path = os.path.join(os.path.realpath('.'), 'downloads',
                                      '{}'.format(argumnets['keywords']))
    if os.path.exists(output_folder_path):
        start_amount_of_files_in_output_folder = len([
            name for name in os.listdir(output_folder_path)
            if os.path.isfile(os.path.join(output_folder_path, name)) and os.
            path.getctime(os.path.join(output_folder_path, name)) < start_time
        ])
    else:
        start_amount_of_files_in_output_folder = 0

    response = google_images_download.googleimagesdownload()
    response.download(argumnets)
    files_modified_after_test_started = [
        name for name in os.listdir(output_folder_path)
        if os.path.isfile(os.path.join(output_folder_path, name)) and
        os.path.getmtime(os.path.join(output_folder_path, name)) > start_time
    ]
    end_amount_of_files_in_output_folder = len(
        files_modified_after_test_started)
    print(f"Files downloaded by test {__name__}:")
    for file in files_modified_after_test_started:
        print(os.path.join(output_folder_path, file))

    # assert end_amount_of_files_in_output_folder - start_amount_of_files_in_output_folder == argumnets['limit']
    assert end_amount_of_files_in_output_folder == argumnets['limit']

    print(f"Cleaning up all files downloaded by test {__name__}...")
    for file in files_modified_after_test_started:
        if silent_remove_of_file(os.path.join(output_folder_path, file)):
            print(f"Deleted {os.path.join(output_folder_path, file)}")
        else:
            print(f"Failed to delete {os.path.join(output_folder_path, file)}")
Ejemplo n.º 10
0
def google_img(search_name, output_folder, num=10, download=False):
    response = google_images_download.googleimagesdownload()
    if download:
        no_download = False
    else:
        no_download = True
    data = response.download({
        "keywords": search_name,
        "color_type": "full-color",
        "format": "jpg",
        "size": "medium",
        "limit": num,
        'no_directory': True,
        'no_download': no_download,
        'silent_mode': True,
        "output_directory": output_folder
    })
    data = list(map(lambda x: x['image_link'], data[2]))
    return {search_name: data}
Ejemplo n.º 11
0
        def downloadimages(query):
            response = google_images_download.googleimagesdownload()
            search_queries = []
            # keywords is the search query
            # format is the image file format
            # limit is the number of images to be downloaded
            # print urs is to print the image file url
            # size is the image size which can
            # be specified manually ("large, medium, icon")
            # aspect ratio denotes the height width ratio
            # of images to download. ("tall, square, wide, panoramic")
            arguments = {
                "keywords": query,
                "format": "jpg",
                "limit": 1,
                "print_urls": True,
                "aspect_ratio": "panoramic",
                "no_directory": True,
                "prefix": query
            }
            try:
                response.download(arguments)

                # Handling File NotFound Error
            except FileNotFoundError:
                arguments = {
                    "keywords": query,
                    "format": "jpg",
                    "limit": 4,
                    "print_urls": True,
                    "size": "medium"
                }

            # Providing arguments for the searched query
            try:
                # Downloading the photos based
                # on the given arguments
                response.download(arguments)
            except:
                pass
import csv
import google_images_download  #importing the library

# Print Header
print('~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~')
print('*********************')
print('                      **********************')
print('            Histology_Image_Search          ')
print('*********************')
print('                      **********************')
print('~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~')

# set up the initial condition
search_terms = []
result = google_images_download.googleimagesdownload()

# open the csv file that contains at least 10 different body regions with search terms just create
with open('Histology_Image_SearchTerms.csv', 'r') as filename:
    data = csv.reader(filename)
    search_terms = list(data)

search_terms.pop(0)  # Removes the Header

# searching each term base for the key shows below
for i in search_terms:
    key = {'keywords': i[1], 'limit': '10', 'print_urls': True, 'prefix': i[0]}

    data = result.download(key)  # save the result data
Ejemplo n.º 13
0
 def Images(self, limit, name):
     # Gets limit(integer) and string and returns list of limit images.
     self.response = google_images_download.googleimagesdownload()
     arguments = {"keywords": name, "limit": limit, "silent_mode": True}
     self.paths = self.response.download(arguments)
     return self.paths[0].get(name)
Ejemplo n.º 14
0
    'x': 3,
    'y': 4,
    'shuffle': True
}

if len(sys.argv) == 2:
    """
    U can set port using command line
    """
    SETTINGS['port'] = sys.argv[1]

GAME_DATA = {'died': False}
sio = socketio.AsyncClient(
    logger=False, reconnection=False
)  # U can turn logger True, if u need to catch events that are not described in current version of program
response = google_images_download.googleimagesdownload(
)  # For this program to work u need modified google_images_download module
"""
Game palette
"""
palette = hitherdither.palette.Palette([
    0xFFFFFF, 0x000000, 0xC1C1C1, 0x4C4C4C, 0xEF130B, 0x740B07, 0xFF7100,
    0xC23800, 0xFFE400, 0xE8A200, 0x00CC00, 0x005510, 0x00B2FF, 0x00569E,
    0x231FD3, 0x0E0865, 0xA300BA, 0x550069, 0xD37CAA, 0xA75574, 0xA0522D,
    0x63300D
])


async def dither(word):
    """
    Here we are creating list of arguments for googleimagesdownload module,
    then we get response that is a list of links to images,
Ejemplo n.º 15
0
import google_images_download

response = google_images_download.googleimagesdownload()

args_male = {
    "keywords": "남자",
    "suffix_keywords": "호랑이상,곰상,공룡상,강아지상,메기상,돼지상",
    "limit": 50,
    "print_urls": True
}

args_female = {
    "keywords": "여자",
    "suffix_keywords": "토끼상,여우상,강아지상,고양이상,금붕어상,다람쥐상",
    "limit": 50,
    "print_urls": True
}

paths = response.download(args_male)
print(paths)

paths = response.download(args_female)
print(paths)
Ejemplo n.º 16
0
def list_add():
	# return "hello"
    #print("Before JSON")
	text = request.get_json()
	# print(text)
	# r="The tiger is the largest extant cat species and a member of the genus Panthera. It is most recognisable for its dark vertical stripes on orange-brown fur with a lighter underside. It is an apex predator, primarily preying on ungulates such as deer and wild boar."
	r=text['data']
	# print(r)
    #print("This is where i am")
    #sys.stdout.flush()
	a,b=generate_summ_key(r)
	response = google_images_download.googleimagesdownload() 
	search_queries = []

	for contents in b:
		search_queries.append(contents)



	def downloadimages(query): 
		# keywords is the search query 
		# format is the image file format 
		# limit is the number of images to be downloaded 
		# print urs is to print the image file url 
		# size is the image size which can 
		# be specified manually ("large, medium, icon") 
		# aspect ratio denotes the height width ratio 
		# of images to download. ("tall, square, wide, panoramic") 
		arguments = {"keywords": query, 
					"format": "jpg", 
					"limit":1, 
					"print_urls":True, 
					"size": "medium", 
					"aspect_ratio": "panoramic"} 
		try: 
			response.download(arguments) 
		
		# Handling File NotFound Error   
		except FileNotFoundError: 
			arguments = {"keywords": query, 
						"format": "jpg", 
						"limit":1, 
						"print_urls":True, 
						"size": "medium"} 
						
			# Providing arguments for the searched query 
			try: 
				# Downloading the photos based 
				# on the given arguments 
				response.download(arguments)
			except: 
				pass

	# Driver Code 
	# orig_stdout = sys.stdout
	# f = open('purang_dega.txt', 'w+')
	# sys.stdout = f
	# print (1234)
	links=[]
	for query in search_queries:
		downloadimages(query)
		links.append(query)
	f = open("keywords.txt", "r")
	
	# print (1234)
	# if f.mode=="r":
		# contents = f.read()
		# links.append(contents)
	p=''
	# print (1234)
	for i in a:
		# p = i + '.'+''.join(prev)
		p=p+'.'+i
	output= p+'  %42069420%  ' + ' '.join(links)
	# print (1234)
	return output
Ejemplo n.º 17
0
    print(f"{item_count}. Out of {all_items_length}")

    ingredient = item['ingredient']
    descriptor = item['descriptor'][0]

    search_query = f"{ingredient} {descriptor}".strip()
    folder_name = '_'.join(ingredient.split())

    print("Search Query", search_query, sep=": ")

    exists = os.path.exists(f"{dowload_path}/{folder_name}")

    if not exists:
        print("Folder Name", folder_name, sep=": ")

        response = googleimagesdownload()
        absolute_image_paths = response.download({
            "keywords":
            search_query,
            "limit":
            number_of_images,
            "image_directory":
            folder_name,
            "output_directory":
            dowload_path
        })

    else:  # folder exists skipping
        stmt = f"Skipping... Folder for {ingredient} exists - {dowload_path}/{folder_name}"
        print(stmt)
Ejemplo n.º 18
0
parser.add_argument('-l',
                    '--limit',
                    help='str type of limit size per keyword',
                    type=str,
                    default='10')

args = vars(parser.parse_args())

if args['add_alias'] == True:
    alias_list = codecs.open(ag.generate_alias(), 'r',
                             encoding='utf8').readlines()
else:
    alias_list = codecs.open(args['keywords_file'], 'r',
                             encoding='utf8').readlines()

downloader = gd.googleimagesdownload()
arguments = dict()

for aliases in alias_list:
    aliases = aliases.strip()
    splits = aliases.split(' ')
    dir_name = splits[0].strip()
    keywords = ', '.join([keyword.replace('_', ' ') for keyword in splits[1:]])
    pdb.set_trace()

    arguments['keywords'] = keywords
    arguments['image_directory'] = dir_name
    arguments['limit'] = args['limit']
    arguments['no_numbering'] = True

    paths, errors = downloader.download_executor(arguments)
Ejemplo n.º 19
0
import google_images_download  #importing the library

response = google_images_download.googleimagesdownload()  #class instantiation

arguments = {
    "keywords": "Street Art",
    "limit": 200,
    "print_urls": True
}  #creating list of arguments
paths = response.download(arguments)  #passing the arguments to the function
print(paths)  #printing absolut
Ejemplo n.º 20
0
def get_image_url(keyword):
    response = google_images_download.googleimagesdownload()
    absolute_image_paths = response.download({'keywords':keyword,'limit':1,'no_download':True,'aspect_ratio':'square','print_urls':True,'usage_rights':'labeled-for-reuse-with-modifications'})
    return absolute_image_paths.get(keyword)[0]