aboutsummaryrefslogtreecommitdiff
path: root/license_protected_downloads/management/commands/s3_flatten.py
blob: c642b6170ea1965619107d09ab8eecf18235707f (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
from django.conf import settings
from django.core.management.base import BaseCommand

import logging
import datetime
from fnmatch import fnmatch
from boto.s3.connection import S3Connection
from boto.s3 import deletemarker,key,prefix
import sys
import httplib
from time import sleep

logging.getLogger().setLevel(logging.INFO)


class Command(BaseCommand):

    help = 'Mark files as deleted or delete files for good, which are older \
            than X days'

    bucket = None

    @staticmethod
    def add_arguments(parser):
        parser.add_argument('--dryrun', action='store_true',
                            help='Do not perform any actions, just report')
        parser.add_argument('--markdays', default=90,
                            help='Number of days to mark files as deleted')
        parser.add_argument('--deletedays', default=180,
                            help='Number of days to delete files for good')
        parser.add_argument('--prefix', default='snapshots/',
                            help='Custom prefix path')
        parser.add_argument('-V', '--verbose', action='store_true',
                            help='log detailed information on actions to INFO')

    @staticmethod
    def x_days_ago(days):
        date = datetime.datetime.now() - datetime.timedelta(days=days)
        return date.isoformat()

    @staticmethod
    def print_key(key):
        if isinstance(key, prefix.Prefix):
            return "DIRECTORY: %s" % key.name

        if key.is_latest:
            latest = "*"
        else:
            latest = " "

        if isinstance(key, deletemarker.DeleteMarker):
            dm = "DEL"
        else:
            dm = "   "
        return '%s: %s %s(%s) %s' %  (key.name, dm, latest,key.last_modified, key.version_id)

    @staticmethod
    def delete_objects(bucket, delete_list, excludes_list=[], dryrun=True, verbose=False):
        if verbose:
            for x in delete_list:
                if isinstance(x, key.Key) or isinstance(x, deletemarker.DeleteMarker):
                    logging.info("deleting: %s %s" % (x.name,x.version_id))
                else:
                    logging.info("deleting: %s" % (x))

        if not dryrun:
            bucket.delete_keys(delete_list)
        else:
            logging.info( "DRYRUN: delete_keys for %s keys" % len(delete_list) )

    def handle(self, *args, **options):
        conn = S3Connection(settings.AWS_ACCESS_KEY_ID,
                            settings.AWS_SECRET_ACCESS_KEY)
        self.bucket = conn.get_bucket(settings.S3_BUCKET, validate=False)
        self.now_mark = self.x_days_ago(int(options['markdays']))
        self.now_delete = self.x_days_ago(int(options['deletedays']))

        self.handle_bucket_retry(*args, **options)

    # wrap the handle_bucket() method in a try/except to allow us to
    # retry if there's an httplib error.
    def handle_bucket_retry(self, *args, **options):
        retries = 3
        while retries > 0:
            try:
                self.handle_bucket(*args, **options)
                retries = 0
            except httplib.BadStatusLine as e:
                logging.error("httplib error handle_bucket():  %s" % e)
                retries -= 1
                sleep(30)

    def handle_bucket(self, *args, **options):
        logging.info( "--> %s" % options['prefix'])

        bucket_keys = self.bucket.list_versions(options['prefix'], delimiter='/')

        objs = {}
        delete_list = []
        subdirs = []

        if options['verbose']:
            logging.info( "Delete day: %s" % self.now_delete)
            logging.info( "Mark day: %s" % self.now_mark)

        for key in bucket_keys:
            if options['verbose']:
                logging.info("%s - %s" %(self.print_key(key), type(key)))

            # if it's a subdir, then we need to descend into in a separate
            # call
            if isinstance(key, prefix.Prefix):
                subdirs.append(key.name)
                continue

            if key.name not in objs:
                objs[key.name] = {'last':None, 'delete':None}

            # flatten everything by filtering out everything except the
            # latest versions of the key and/or deletemarker
            if isinstance(key, deletemarker.DeleteMarker):
                if objs[key.name]['delete'] is None:
                    objs[key.name]['delete'] = key
                elif key.last_modified > objs[key.name]['delete'].last_modified:
                    delete_list.append(objs[key.name]['delete'])
                    objs[key.name]['delete'] = key
                else:
                    delete_list.append(key)
            else:
                if objs[key.name]['last'] is None:
                    objs[key.name]['last'] = key
                elif key.last_modified > objs[key.name]['last'].last_modified:
                    delete_list.append(objs[key.name]['last'])
                    objs[key.name]['last'] = key
                else:
                    delete_list.append(key)

                # if the new 'last' is newer than an existing deletemarker,
                # delete the deletemarker
                if objs[key.name]['delete'] and \
                   objs[key.name]['last'].last_modified >= objs[key.name]['delete'].last_modified:
                    delete_list.append(objs[key.name]['delete'])
                    objs[key.name]['delete'] = None

            # purge as we go
            if len(delete_list) > 1000:
                while delete_list:
                    self.delete_objects(self.bucket, delete_list[0:1000], settings.S3_PURGE_EXCLUDES, options['dryrun'], options['verbose'])
                    delete_list = delete_list[1000:]

        if options['verbose']:
            logging.info("done with flatten")

        # search through everything w/ a delete marker to delete
        for candidate in [ x for x in objs if objs[x]['delete']]:
            # if in exclude we ignore it even if it has a delete marker
            if any(fnmatch(candidate, p) for p in settings.S3_PURGE_EXCLUDES):
                if options['verbose']:
                    logging.info("excluded: %s" % candidate)
                continue
            else:
                if objs[candidate]['last'] is None:
                    # no point in keeping around a delete marker that points to nothing
                    delete_list.append(objs[candidate]['delete'])
                else:
                    # check last_modified on the last real file, not delete marker
                    if objs[candidate]['last'].last_modified < self.now_delete:
                        delete_list.append(objs[candidate]['delete'])
                        delete_list.append(objs[candidate]['last'])

        if options['verbose']:
            logging.info("done with now_delete")

        # search through everything w/o a delete marker to possibly mark
        for candidate in [ x for x in objs if not objs[x]['delete']]:
            if any(fnmatch(candidate, p) for p in settings.S3_PURGE_EXCLUDES):
                if options['verbose']:
                    logging.info("excluded: %s" % candidate)
                continue
            else:
                if objs[candidate]['last'].last_modified < self.now_mark:
                    if not options['dryrun']:
                        # by appending only the name rather than the key
                        # object, S3 should insert a delete marker
                        delete_list.append(objs[candidate]['last'].name)
                    else:
                        logging.info("DRYRUN: setting deletemarker on %s - %s" % (objs[candidate]['last'].name, objs[candidate]['last'].version_id))

        if options['verbose']:
            logging.info("done with now_mark")


        while len(delete_list) > 1000:
            self.delete_objects(self.bucket, delete_list[0:1000], settings.S3_PURGE_EXCLUDES, options['dryrun'], options['verbose'])
            delete_list = delete_list[1000:]
        self.delete_objects(self.bucket, delete_list[0:1000], settings.S3_PURGE_EXCLUDES, options['dryrun'], options['verbose'])
        if options['verbose']:
            logging.info("done with cleanup.")

        # clean up mem and descend to any child directories
        del objs
        for s in subdirs:
            new_opts = options
            new_opts['prefix'] = s
            self.handle_bucket_retry(*args, **new_opts)