So I'm trying to do a small crawler to just pick a few Google-search images links and then download them. It's not going to be anything that needs to run 1000 times a day with 1000 queries, but just a simple script to download 10 of the first images for a certain search word.
For that I have the following code:
import requests
from bs4 import BeautifulSoup
import json
import urllib
s = requests.session()
s.headers.update({"User-Agent": "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/36.0.1985.125 Safari/537.36"})
URL = "https://www.google.dk/search"
def get_images(query, start):
images = []
screen_width = 1920
screen_height = 1080
params = {
"q": query,
"sa": "X",
"biw": screen_width,
"bih": screen_height,
"tbm": "isch",
"ijn": start/100,
"start": start,
#"ei": "" - This seems like a unique ID, you might want to use it to avoid getting banned. But you probably still are.
}
request = s.get(URL, params=params)
bs = BeautifulSoup(request.text, "lxml")
for img in bs.findAll("div", {"class": "rg_meta"}):
js = json.loads(img.text)
images.append(js['ou'])
return images
So basically I get a list of links I can then parse through and download via this code where it even names the images from 1 to how many there is now being crawled:
searchlist = ["cats"] #search strings
nr_img = 5 #number of images to be crawled
for k, searchstring in enumerate(searchlist):
k += 0
images = get_images("{}".format(searchstring), 0)
img_nr_list = []
for n, x in enumerate(images[0:nr_img]):
n += 1+k*nr_img
urllib.urlretrieve("{}".format(x), "\foo\bar\{}.jpg".format(n))
img_nr_list.append("{}.jpg".format(n))
In principle pretty straight forward. However, my problem is that some images are just thumbnails, or just have a low image size. So my question is: Is there a way in which I can say something like: "If width < 600px and height < 400px then skip" or something like that ?