Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

Is it possible to automatically delete objects older than 10 minutes in AWS S3?

We want to delete objects from S3, 10 minutes after they are created. Is it possible currently?

like image 266
Onkaar Singh Avatar asked Dec 01 '16 07:12

Onkaar Singh


People also ask

Which feature of S3 can automatically Delete objects?

You can set up a lifecycle rule to automatically delete objects such as log files. For more information, see Setting lifecycle configuration on a bucket.

How long do S3 buckets take to delete?

When you delete a bucket, there may be a delay of up to one hour before the bucket name is available for reuse in a new region or by a new bucket owner. If you re-create the bucket in the same region or with the same bucket owner, there is no delay.


1 Answers

I have a working solution that was built serverless with the help of AWS's Simple Queue Service and AWS Lambda. This works for all objects created in an s3 bucket.

Overview

When any object is created in your s3 bucket, the bucket will send an event with object details to an SQS queue configured with a 10 minute delivery delay. The SQS queue is also configured to trigger a Lambda function. The Lambda function reads the object details from the event sent and deletes the object from the s3 bucket. All three components involved (s3, SQS and Lambda) are low cost, loosely coupled, serverless and scale automatically to very large workloads.

Steps Involved

  1. Setup your Lambda Function First. In my solution, I used Python 3.7. The code for the function is:

    import json
    import boto3
    
    def lambda_handler(event, context):
    
    for record in event['Records']:
        v = json.loads(record['body'])
        for rec in v["Records"]:
    
            bucketName = rec["s3"]["bucket"]["name"]
            objectKey = rec["s3"]["object"]["key"]
            #print("bucket is " + bucketName + " and object is " + objectKey )
    
            sss = boto3.resource("s3")
            obj = sss.Object(bucketName, objectKey)
            obj.delete()
    
    return {
        'statusCode': 200,
        'body': json.dumps('Delete Completed.')
    }
    

This code and a sample message file were uploaded to a github repo.

  1. Create a vanilla SQS queue. Then configure the SQS queue to have a 10 minute delivery Delay. This setting can be found under Queue Actions -> Configure Queue -> 4 setting down

enter image description here

  1. Configure the SQS Queue to trigger the Lambda Function you created in Step 1. To do this use Queue Actions -> Configure Trigger for Lambda Function. The setup screen is self explanatory. If you don't see your Lambda function from step 1, redo it correctly and make sure you are using the same Region.

  2. Setup your S3 Bucket so that it fires an event to the SQS Queue you created in step 2. This is found on the main bucket screen, click Properties tab and select Events. Click the plus sign to add an event and fill out the following form:

enter image description here

Important points to select are to select All Object create events and to select the queue you created in Step 2 for the last pull down on this screen.

  1. Last step - Add an execute policy to your Lambda Function that allows it to only delete from the specific S3 bucket. You can do this via the Lambda function console. Scroll down the Lambda function screen of your console and configure it under Execution Role.

This works for files I've copied into a single s3 bucket. This solution could support many S3 buckets to 1 queue and 1 lambda.

like image 86
Taterhead Avatar answered Oct 18 '22 17:10

Taterhead