document quick mode
[clinton/thingy_grabber.git] / thingy_grabber.py
dissimilarity index 61%
index 587d47e..981dd74 100755 (executable)
-#!/usr/bin/env python3
-"""
-Thingiverse bulk downloader
-"""
-
-import re
-import os
-import argparse
-import unicodedata
-import requests
-from bs4 import BeautifulSoup
-
-URL_BASE = "https://www.thingiverse.com"
-URL_COLLECTION = URL_BASE + "/ajax/thingcollection/list_collected_things"
-
-ID_REGEX = re.compile(r'"id":(\d*),')
-TOTAL_REGEX = re.compile(r'"total":(\d*),')
-LAST_PAGE_REGEX = re.compile(r'"last_page":(\d*),')
-# This appears to be fixed at 12, but if it changes would screw the rest up.
-PER_PAGE_REGEX = re.compile(r'"per_page":(\d*),')
-NO_WHITESPACE_REGEX = re.compile(r'[-\s]+')
-
-VERBOSE = False
-
-def strip_ws(value):
-    """ Remove whitespace from a string """
-    return str(NO_WHITESPACE_REGEX.sub('-', value))
-
-def slugify(value):
-    """
-    Normalizes string, converts to lowercase, removes non-alpha characters,
-    and converts spaces to hyphens.
-    """
-    value = unicodedata.normalize('NFKD', value).encode('ascii', 'ignore').decode()
-    value = str(re.sub(r'[^\w\s-]', '', value).strip())
-    value = str(NO_WHITESPACE_REGEX.sub('-', value))
-    #value = str(re.sub(r'[-\s]+', '-', value))
-    return value
-
-class Collection:
-    """ Holds details of a collection. """
-    def __init__(self, user, name):
-        self.user = user
-        self.name = name
-        self.things = []
-        self.total = 0
-        self.req_id = None
-        self.last_page = 0
-        self.per_page = None
-
-    def _get_small_collection(self, req):
-        """ Handle small collections """
-        soup = BeautifulSoup(req.text, features='lxml')
-        links = soup.find_all('a', {'class':'card-img-holder'})
-        self.things = [x['href'].split(':')[1] for x in links]
-
-        return self.things
-
-    def get_collection(self):
-        """ retrieve the things of the collection. """
-        if self.things:
-            # We've already done it.
-            return self.things
-
-        # Get the internal details of the collection.
-        c_url = "{}/{}/collections/{}".format(URL_BASE, self.user, strip_ws(self.name))
-        if VERBOSE:
-            print("Querying {}".format(c_url))
-        c_req = requests.get(c_url)
-        total = TOTAL_REGEX.search(c_req.text)
-        if total is None:
-            # This is a small (<13) items collection. Pull the list from this req.
-            return self._get_small_collection(c_req)
-        self.total = total.groups()[0]
-        self.req_id = ID_REGEX.search(c_req.text).groups()[0]
-        self.last_page = int(LAST_PAGE_REGEX.search(c_req.text).groups()[0])
-        self.per_page = PER_PAGE_REGEX.search(c_req.text).groups()[0]
-        parameters = {
-            'base_url':"{}/collections/{}".format(self.user, self.name),
-            'page':'1',
-            'per_page':'12',
-            'id':self.req_id
-        }
-        for current_page in range(1, self.last_page + 1):
-            parameters['page'] = current_page
-            req = requests.post(URL_COLLECTION, parameters)
-            soup = BeautifulSoup(req.text, features='lxml')
-            links = soup.find_all('a', {'class':'card-img-holder'})
-            self.things += [x['href'].split(':')[1] for x in links]
-
-        return self.things
-
-    def download(self):
-        """ Downloads all the files in a collection """
-        if not self.things:
-            self.get_collection()
-        base_dir = os.getcwd()
-        new_dir = "{}-{}".format(slugify(self.user), slugify(self.name))
-        target_dir = os.path.join(base_dir, new_dir)
-        try:
-            os.mkdir(target_dir)
-        except FileExistsError:
-            print("Target directory {} already exists. Assuming a resume.".format(new_dir))
-        os.chdir(target_dir)
-        for thing in self.things:
-            download_thing(thing)
-
-
-def download_thing(thing):
-    """ Downloads all the files for a given thing. """
-    file_url = "{}/thing:{}/files".format(URL_BASE, thing)
-    file_req = requests.get(file_url)
-    file_soup = BeautifulSoup(file_req.text, features='lxml')
-
-    title = slugify(file_soup.find_all('h1')[0].text.strip())
-    base_dir = os.getcwd()
-    try:
-        os.mkdir(title)
-    except FileExistsError:
-        pass
-
-    print("Downloading {} ({})".format(thing, title))
-    os.chdir(title)
-    last_time = None
-
-    try:
-        with open('timestamp.txt', 'r') as fh:
-            last_time = fh.readlines()[0]
-        if VERBOSE:
-            print("last downloaded version: {}".format(last_time))
-    except FileNotFoundError:
-        # Not run on this thing before.
-        if VERBOSE:
-            print('Directory for thing already exists, checking for update.')
-        last_time = None
-
-    file_links = file_soup.find_all('a', {'class':'file-download'})
-    new_last_time = last_time
-    new_file_links = []
-
-    for file_link in file_links:
-        timestamp = file_link.find_all('time')[0]['datetime']
-        if VERBOSE:
-            print("Checking {} (updated {})".format(file_link["title"], timestamp))
-        if not last_time or timestamp > last_time:
-            new_file_links.append(file_link)
-        if not new_last_time or timestamp > new_last_time:
-            new_last_time = timestamp
-
-    if last_time and new_last_time <= last_time:
-        print("Thing already downloaded. Skipping.")
-    files = [("{}{}".format(URL_BASE, x['href']), x["title"]) for x in new_file_links]
-
-    try:
-        for url, name in files:
-            if VERBOSE:
-                print("Downloading {} from {}".format(name, url))
-            data_req = requests.get(url)
-            with open(name, 'wb') as handle:
-                handle.write(data_req.content)
-        # now write timestamp
-        with open('timestamp.txt', 'w') as fh:
-            fh.write(new_last_time)
-    except Exception as exception:
-        print("Failed to download {} - {}".format(name, exception))
-        os.chdir(base_dir)
-        os.rename(title, "{}_failed".format(title))
-        return
-
-
-    os.chdir(base_dir)
-
-def main():
-    """ Entry point for script being run as a command. """
-    parser = argparse.ArgumentParser()
-    parser.add_argument("owner", help="The owner of the collection to get")
-    parser.add_argument("collection", help="The name of the collection to get")
-    parser.add_argument("-v", "--verbose", help="Be more verbose", action="store_true")
-    args = parser.parse_args()
-    global VERBOSE
-    VERBOSE = args.verbose
-
-    collection = Collection(args.owner, args.collection)
-    print(collection.get_collection())
-    collection.download()
-
-if __name__ == "__main__":
-    main()
+#!/usr/bin/env python3
+"""
+Thingiverse bulk downloader
+"""
+
+import re
+import sys
+import os
+import argparse
+import unicodedata
+import requests
+import logging
+import multiprocessing
+import enum
+from shutil import copyfile
+from bs4 import BeautifulSoup
+
+URL_BASE = "https://www.thingiverse.com"
+URL_COLLECTION = URL_BASE + "/ajax/thingcollection/list_collected_things"
+USER_COLLECTION = URL_BASE + "/ajax/user/designs"
+
+ID_REGEX = re.compile(r'"id":(\d*),')
+TOTAL_REGEX = re.compile(r'"total":(\d*),')
+LAST_PAGE_REGEX = re.compile(r'"last_page":(\d*),')
+# This appears to be fixed at 12, but if it changes would screw the rest up.
+PER_PAGE_REGEX = re.compile(r'"per_page":(\d*),')
+NO_WHITESPACE_REGEX = re.compile(r'[-\s]+')
+
+DOWNLOADER_COUNT = 1
+RETRY_COUNT = 3
+
+VERSION = "0.7.0"
+
+class State(enum.Enum):
+    OK = enum.auto()
+    FAILED = enum.auto()
+    ALREADY_DOWNLOADED = enum.auto()
+
+
+def strip_ws(value):
+    """ Remove whitespace from a string """
+    return str(NO_WHITESPACE_REGEX.sub('-', value))
+
+
+def slugify(value):
+    """
+    Normalizes string, converts to lowercase, removes non-alpha characters,
+    and converts spaces to hyphens.
+    """
+    value = unicodedata.normalize('NFKD', value).encode(
+        'ascii', 'ignore').decode()
+    value = str(re.sub(r'[^\w\s-]', '', value).strip())
+    value = str(NO_WHITESPACE_REGEX.sub('-', value))
+    #value = str(re.sub(r'[-\s]+', '-', value))
+    return value
+
+class Downloader(multiprocessing.Process):
+    """
+    Class to handle downloading the things we have found to get.
+    """
+
+    def __init__(self, thing_queue, download_directory):
+        multiprocessing.Process.__init__(self)
+        # TODO: add parameters
+        self.thing_queue = thing_queue
+        self.download_directory = download_directory
+
+    def run(self):
+        """ actual download loop.
+        """
+        while True:
+            thing_id = self.thing_queue.get()
+            if thing_id is None:
+                logging.info("Shutting download queue")
+                self.thing_queue.task_done()
+                break
+            logging.info("Handling id {}".format(thing_id))
+            Thing(thing_id).download(self.download_directory)
+            self.thing_queue.task_done()
+        return
+
+
+                
+
+
+class Grouping:
+    """ Holds details of a group of things for download
+        This is effectively (although not actually) an abstract class
+        - use Collection or Designs instead.
+    """
+
+    def __init__(self, quick):
+        self.things = []
+        self.total = 0
+        self.req_id = None
+        self.last_page = 0
+        self.per_page = None
+        # Should we stop downloading when we hit a known datestamp?
+        self.quick = quick 
+        # These should be set by child classes.
+        self.url = None
+        self.download_dir = None
+        self.collection_url = None
+
+    def _get_small_grouping(self, req):
+        """ Handle small groupings """
+        soup = BeautifulSoup(req.text, features='lxml')
+        links = soup.find_all('a', {'class': 'card-img-holder'})
+        self.things = [x['href'].split(':')[1] for x in links]
+        self.total = len(self.things)
+
+        return self.things
+
+    def get(self):
+        """ retrieve the things of the grouping. """
+        if self.things:
+            # We've already done it.
+            return self.things
+
+        # Check for initialisation:
+        if not self.url:
+            logging.error("No URL set - object not initialised properly?")
+            raise ValueError("No URL set - object not initialised properly?")
+
+        # Get the internal details of the grouping.
+        logging.debug("Querying {}".format(self.url))
+        c_req = requests.get(self.url)
+        total = TOTAL_REGEX.search(c_req.text)
+        if total is None:
+            # This is a small (<13) items grouping. Pull the list from this req.
+            return self._get_small_grouping(c_req)
+        self.total = total.groups()[0]
+        self.req_id = ID_REGEX.search(c_req.text).groups()[0]
+        self.last_page = int(LAST_PAGE_REGEX.search(c_req.text).groups()[0])
+        self.per_page = PER_PAGE_REGEX.search(c_req.text).groups()[0]
+        parameters = {
+            'base_url': self.url,
+            'page': '1',
+            'per_page': '12',
+            'id': self.req_id
+        }
+        for current_page in range(1, self.last_page + 1):
+            parameters['page'] = current_page
+            req = requests.post(self.collection_url, parameters)
+            soup = BeautifulSoup(req.text, features='lxml')
+            links = soup.find_all('a', {'class': 'card-img-holder'})
+            self.things += [x['href'].split(':')[1] for x in links]
+
+        return self.things
+
+    def download(self):
+        """ Downloads all the files in a collection """
+        if not self.things:
+            self.get()
+
+        if not self.download_dir:
+            raise ValueError(
+                "No download_dir set - invalidly initialised object?")
+
+        base_dir = os.getcwd()
+        try:
+            os.mkdir(self.download_dir)
+        except FileExistsError:
+            logging.info("Target directory {} already exists. Assuming a resume."
+                         .format(self.download_dir))
+        logging.info("Downloading {} thing(s).".format(self.total))
+        for idx, thing in enumerate(self.things):
+            logging.info("Downloading thing {}".format(idx))
+            RC = Thing(thing).download(self.download_dir)
+            if self.quick and RC==State.ALREADY_DOWNLOADED:
+                logging.info("Caught up, stopping.")
+                return
+
+
+class Collection(Grouping):
+    """ Holds details of a collection. """
+
+    def __init__(self, user, name, directory, quick):
+        Grouping.__init__(self, quick)
+        self.user = user
+        self.name = name
+        self.url = "{}/{}/collections/{}".format(
+            URL_BASE, self.user, strip_ws(self.name))
+        self.download_dir = os.path.join(directory,
+                                         "{}-{}".format(slugify(self.user), slugify(self.name)))
+        self.collection_url = URL_COLLECTION
+
+
+class Designs(Grouping):
+    """ Holds details of all of a users' designs. """
+
+    def __init__(self, user, directory, quick):
+        Grouping.__init__(self, quick)
+        self.user = user
+        self.url = "{}/{}/designs".format(URL_BASE, self.user)
+        self.download_dir = os.path.join(
+            directory, "{} designs".format(slugify(self.user)))
+        self.collection_url = USER_COLLECTION
+
+
+class Thing:
+    """ An individual design on thingiverse. """
+
+    def __init__(self, thing_id):
+        self.thing_id = thing_id
+        self.last_time = None
+        self._parsed = False
+        self._needs_download = True
+        self.text = None
+        self.title = None
+        self.download_dir = None
+
+    def _parse(self, base_dir):
+        """ Work out what, if anything needs to be done. """
+        if self._parsed:
+            return
+
+        url = "{}/thing:{}/files".format(URL_BASE, self.thing_id)
+        try:
+            req = requests.get(url)
+        except requests.exceptions.ConnectionError as error:
+            logging.error("Unable to connect for thing {}: {}".format(
+                self.thing_id, error))
+            return
+
+        self.text = req.text
+        soup = BeautifulSoup(self.text, features='lxml')
+        #import code
+        #code.interact(local=dict(globals(), **locals()))
+        try:
+            self.title = slugify(soup.find_all('h1')[0].text.strip())
+        except IndexError:
+            logging.warning(
+                "No title found for thing {}".format(self.thing_id))
+            self.title = self.thing_id
+
+        if req.status_code == 404:
+            logging.warning(
+                "404 for thing {} - DMCA or invalid number?".format(self.thing_id))
+            return
+
+        if req.status_code > 299:
+            logging.warning(
+                "bad status code {}  for thing {} - try again later?".format(req.status_code, self.thing_id))
+            return
+
+        self.download_dir = os.path.join(base_dir, self.title)
+
+        logging.debug("Parsing {} ({})".format(self.thing_id, self.title))
+
+        if not os.path.exists(self.download_dir):
+            # Not yet downloaded
+            self._parsed = True
+            return
+
+        timestamp_file = os.path.join(self.download_dir, 'timestamp.txt')
+        if not os.path.exists(timestamp_file):
+            # Old download from before
+            logging.warning(
+                "Old-style download directory found. Assuming update required.")
+            self._parsed = True
+            return
+
+        try:
+            with open(timestamp_file, 'r') as timestamp_handle:
+                self.last_time = timestamp_handle.readlines()[0]
+            logging.info("last downloaded version: {}".format(self.last_time))
+        except FileNotFoundError:
+            # Not run on this thing before.
+            logging.info(
+                "Old-style download directory found. Assuming update required.")
+            self.last_time = None
+            self._parsed = True
+            return
+
+        # OK, so we have a timestamp, lets see if there is anything new to get
+        file_links = soup.find_all('a', {'class': 'file-download'})
+        for file_link in file_links:
+            timestamp = file_link.find_all('time')[0]['datetime']
+            logging.debug("Checking {} (updated {})".format(
+                file_link["title"], timestamp))
+            if timestamp > self.last_time:
+                logging.info(
+                    "Found new/updated file {}".format(file_link["title"]))
+                self._needs_download = True
+                self._parsed = True
+                return
+        # Got here, so nope, no new files.
+        self._needs_download = False
+        self._parsed = True
+
+    def download(self, base_dir):
+        """ Download all files for a given thing. 
+            Returns True iff the thing is now downloaded (not iff it downloads the thing!)
+        """
+        if not self._parsed:
+            self._parse(base_dir)
+
+        if not self._parsed:
+            logging.error(
+                "Unable to parse {} - aborting download".format(self.thing_id))
+            return State.FAILED
+
+        if not self._needs_download:
+            print("{} - {} already downloaded - skipping.".format(self.thing_id, self.title))
+            return State.ALREADY_DOWNLOADED
+
+        # Have we already downloaded some things?
+        timestamp_file = os.path.join(self.download_dir, 'timestamp.txt')
+        prev_dir = None
+        if os.path.exists(self.download_dir):
+            if not os.path.exists(timestamp_file):
+                # edge case: old style dir w/out timestamp.
+                logging.warning(
+                    "Old style download dir found for {}".format(self.title))
+                prev_count = 0
+                target_dir = "{}_old".format(self.download_dir)
+                while os.path.exists(target_dir):
+                    prev_count = prev_count + 1
+                    target_dir = "{}_old_{}".format(self.download_dir, prev_count)
+                os.rename(self.download_dir, target_dir)
+            else:
+                prev_dir = "{}_{}".format(self.download_dir, self.last_time)
+                os.rename(self.download_dir, prev_dir)
+
+        # Get the list of files to download
+        soup = BeautifulSoup(self.text, features='lxml')
+        file_links = soup.find_all('a', {'class': 'file-download'})
+
+        new_file_links = []
+        old_file_links = []
+        new_last_time = None
+
+        if not self.last_time:
+            # If we don't have anything to copy from, then it is all new.
+            new_file_links = file_links
+            try:
+                new_last_time = file_links[0].find_all('time')[0]['datetime']
+            except:
+                import code
+                code.interact(local=dict(globals(), **locals()))
+
+            for file_link in file_links:
+                timestamp = file_link.find_all('time')[0]['datetime']
+                logging.debug("Found file {} from {}".format(
+                    file_link["title"], timestamp))
+                if timestamp > new_last_time:
+                    new_last_time = timestamp
+        else:
+            for file_link in file_links:
+                timestamp = file_link.find_all('time')[0]['datetime']
+                logging.debug("Checking {} (updated {})".format(
+                    file_link["title"], timestamp))
+                if timestamp > self.last_time:
+                    new_file_links.append(file_link)
+                else:
+                    old_file_links.append(file_link)
+                if not new_last_time or timestamp > new_last_time:
+                    new_last_time = timestamp
+
+        logging.debug("new timestamp {}".format(new_last_time))
+
+        # OK. Time to get to work.
+        logging.debug("Generating download_dir")
+        os.mkdir(self.download_dir)
+        # First grab the cached files (if any)
+        logging.info("Copying {} unchanged files.".format(len(old_file_links)))
+        for file_link in old_file_links:
+            old_file = os.path.join(prev_dir, file_link["title"])
+            new_file = os.path.join(self.download_dir, file_link["title"])
+            try:
+                logging.debug("Copying {} to {}".format(old_file, new_file))
+                copyfile(old_file, new_file)
+            except FileNotFoundError:
+                logging.warning(
+                    "Unable to find {} in old archive, redownloading".format(file_link["title"]))
+                new_file_links.append(file_link)
+
+        # Now download the new ones
+        files = [("{}{}".format(URL_BASE, x['href']), x["title"])
+                 for x in new_file_links]
+        logging.info("Downloading {} new files of {}".format(
+            len(new_file_links), len(file_links)))
+        try:
+            for url, name in files:
+                file_name = os.path.join(self.download_dir, name)
+                logging.debug("Downloading {} from {} to {}".format(
+                    name, url, file_name))
+                data_req = requests.get(url)
+                with open(file_name, 'wb') as handle:
+                    handle.write(data_req.content)
+        except Exception as exception:
+            logging.error("Failed to download {} - {}".format(name, exception))
+            os.rename(self.download_dir, "{}_failed".format(self.download_dir))
+            return State.FAILED
+
+        # People like images
+        image_dir = os.path.join(self.download_dir, 'images')
+        imagelinks = soup.find_all('span', {'class': 'gallery-slider'})[0] \
+                         .find_all('div', {'class': 'gallery-photo'})
+        logging.info("Downloading {} images.".format(len(imagelinks)))
+        try:
+            os.mkdir(image_dir)
+            for imagelink in imagelinks:
+                url = next(filter(None, [imagelink[x] for x in ['data-full',
+                                                                'data-large',
+                                                                'data-medium',
+                                                                'data-thumb']]), None)
+                if not url:
+                    logging.warning(
+                        "Unable to find any urls for {}".format(imagelink))
+                    continue
+
+                filename = os.path.basename(url)
+                if filename.endswith('stl'):
+                    filename = "{}.png".format(filename)
+                image_req = requests.get(url)
+                with open(os.path.join(image_dir, filename), 'wb') as handle:
+                    handle.write(image_req.content)
+        except Exception as exception:
+            print("Failed to download {} - {}".format(filename, exception))
+            os.rename(self.download_dir, "{}_failed".format(self.download_dir))
+            return State.FAILED
+
+        # instructions are good too.
+        logging.info("Downloading readme")
+        try:
+            readme_txt = soup.find('meta', property='og:description')[
+                'content']
+            with open(os.path.join(self.download_dir, 'readme.txt'), 'w') as readme_handle:
+                readme_handle.write("{}\n".format(readme_txt))
+        except (TypeError, KeyError) as exception:
+            logging.warning("No readme? {}".format(exception))
+        except IOError as exception:
+            logging.warning("Failed to write readme! {}".format(exception))
+
+        # Best get some licenses
+        logging.info("Downloading license")
+        try:
+            license_txt = soup.find('div', {'class': 'license-text'}).text
+            if license_txt:
+                with open(os.path.join(self.download_dir, 'license.txt'), 'w') as license_handle:
+                    license_handle.write("{}\n".format(license_txt))
+        except AttributeError as exception:
+            logging.warning("No license? {}".format(exception))
+        except IOError as exception:
+            logging.warning("Failed to write license! {}".format(exception))
+
+        try:
+            # Now write the timestamp
+            with open(timestamp_file, 'w') as timestamp_handle:
+                timestamp_handle.write(new_last_time)
+        except Exception as exception:
+            print("Failed to write timestamp file - {}".format(exception))
+            os.rename(self.download_dir, "{}_failed".format(self.download_dir))
+            return State.FAILED
+        self._needs_download = False
+        logging.debug("Download of {} finished".format(self.title))
+        return State.OK
+
+
+def do_batch(batch_file, download_dir, quick):
+    """ Read a file in line by line, parsing each as a set of calls to this script."""
+    with open(batch_file) as handle:
+        for line in handle:
+            line = line.strip()
+            logging.info("Handling instruction {}".format(line))
+            command_arr = line.split()
+            if command_arr[0] == "thing":
+                logging.debug(
+                    "Handling batch thing instruction: {}".format(line))
+                Thing(command_arr[1]).download(download_dir)
+                continue
+            if command_arr[0] == "collection":
+                logging.debug(
+                    "Handling batch collection instruction: {}".format(line))
+                Collection(command_arr[1], command_arr[2],
+                           download_dir, quick).download()
+                continue
+            if command_arr[0] == "user":
+                logging.debug(
+                    "Handling batch collection instruction: {}".format(line))
+                Designs(command_arr[1], download_dir, quick).download()
+                continue
+            logging.warning("Unable to parse current instruction. Skipping.")
+
+
+def main():
+    """ Entry point for script being run as a command. """
+    parser = argparse.ArgumentParser()
+    parser.add_argument("-l", "--log-level", choices=[
+                        'debug', 'info', 'warning'], default='info', help="level of logging desired")
+    parser.add_argument("-d", "--directory",
+                        help="Target directory to download into")
+    parser.add_argument("-f", "--log-file",
+                        help="Place to log debug information to")
+    parser.add_argument("-q", "--quick", action="store_true",
+                        help="Assume date ordering on posts")
+
+    subparsers = parser.add_subparsers(
+        help="Type of thing to download", dest="subcommand")
+    collection_parser = subparsers.add_parser(
+        'collection', help="Download one or more entire collection(s)")
+    collection_parser.add_argument(
+        "owner", help="The owner of the collection(s) to get")
+    collection_parser.add_argument(
+        "collections", nargs="+",  help="Space seperated list of the name(s) of collection to get")
+    thing_parser = subparsers.add_parser(
+        'thing', help="Download a single thing.")
+    thing_parser.add_argument(
+        "things", nargs="*", help="Space seperated list of thing ID(s) to download")
+    user_parser = subparsers.add_parser(
+        "user",  help="Download all things by one or more users")
+    user_parser.add_argument(
+        "users", nargs="+", help="A space seperated list of the user(s) to get the designs of")
+    batch_parser = subparsers.add_parser(
+        "batch", help="Perform multiple actions written in a text file")
+    batch_parser.add_argument(
+        "batch_file", help="The name of the file to read.")
+    subparsers.add_parser("version", help="Show the current version")
+
+    args = parser.parse_args()
+    if not args.subcommand:
+        parser.print_help()
+        sys.exit(1)
+    if not args.directory:
+        args.directory = os.getcwd()
+
+    logger = logging.getLogger()
+    formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')
+    logger.setLevel(logging.DEBUG)
+    console_handler = logging.StreamHandler()
+    console_handler.setLevel(args.log_level.upper())
+
+    logger.addHandler(console_handler)
+    if args.log_file:
+        file_handler = logging.FileHandler(args.log_file)
+        file_handler.setLevel(logging.DEBUG)
+        file_handler.setFormatter(formatter)
+        logger.addHandler(file_handler)
+
+
+    # Start downloader
+    thing_queue = multiprocessing.JoinableQueue()
+    logging.debug("starting {} downloader(s)".format(DOWNLOADER_COUNT))
+    downloaders = [Downloader(thing_queue, args.directory) for _ in range(DOWNLOADER_COUNT)]
+    for downloader in downloaders:
+        downloader.start()
+
+
+    if args.subcommand.startswith("collection"):
+        for collection in args.collections:
+            Collection(args.owner, collection, args.directory, args.quick).download()
+    if args.subcommand == "thing":
+        for thing in args.things:
+            thing_queue.put(thing)
+    if args.subcommand == "user":
+        for user in args.users:
+            Designs(user, args.directory, args.quick).download()
+    if args.subcommand == "version":
+        print("thingy_grabber.py version {}".format(VERSION))
+    if args.subcommand == "batch":
+        do_batch(args.batch_file, args.directory, args.quick)
+
+    # Stop the downloader processes
+    for downloader in downloaders:
+        thing_queue.put(None)
+
+if __name__ == "__main__":
+    main()