1
0
mirror of https://github.com/bregman-arie/devops-exercises.git synced 2024-11-19 20:31:47 +02:00

chore: added aws s3 event triggering script (#10255)

This commit is contained in:
Adarsh 2024-06-12 14:47:49 +05:30 committed by GitHub
parent df28b9bdb9
commit e58a9ccd73
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
5 changed files with 162 additions and 0 deletions

View File

@ -0,0 +1 @@
[](./sample.png)

View File

@ -0,0 +1,122 @@
#!/bin/bash
# always put up the detail of scripts . version, author, what it does, what event triggers and all ..
###
# Author: Adarsh Rawat
# Version: 1.0.0
# Objective: Automate Notification for a object uploaded or created in s3 bucket.
###
# debug what is happening
set -x
# all these cmds are aws cli commands | abhishek veermalla day 4-5 devops
# store aws account id in a variable
aws_account_id=$(aws sts get-caller-identity --query 'Account' --output text)
# print the account id from the variable
echo "aws account id: $aws_account_id"
# set aws region, bucket name and other variables
aws_region="us-east-1"
aws_bucket="s3-lambda-event-trigger-bucket"
aws_lambda="s3-lambda-function-1"
aws_role="s3-lambda-sns"
email_address="adarshrawat8304@gmail.com"
# create iam role for the project
role_response=$(aws iam create-role --role-name s3-lambda-sns --assume-role-policy-document '{
"Version": "2012-10-17",
"Statement": [{
"Action": "sts:AssumeRole",
"Effect": "Allow",
"Principal": {
"Service": [
"lambda.amazonaws.com",
"s3.amazonaws.com",
"sns.amazonaws.com"
]
}
}]
}')
# jq is json parser here parse the role we created
# extract the role arn from json resposne and store in variable
role_arn=$(echo "$role_response" | jq -r '.Role.Arn')
# print the role arn
echo "Role ARN: $role_arn"
# attach permissions to the role
aws iam attach-role-policy --role-name $aws_role --policy-arn arn:aws:iam::aws:policy/AWSLambda_FullAccess
aws iam attach-role-policy --role-name $aws_role --policy-arn arn:aws:iam::aws:policy/AmazonSNSFullAccess
# create s3 bucket and get the output in a variable
bucket_output=$(aws s3api create-bucket --bucket "$aws_bucket" --region "$aws_region")
# print the output from the variable
echo "bucket output: $bucket_output"
# upload a file to the bucket
aws s3 cp ./sample.png s3://"$aws_bucket"/sample.png
# create a zip file to upload lambda function
zip -r s3-lambda.zip ./s3-lambda
sleep 5
# create a lambda function
aws lambda create-function \
--region $aws_region \
--function $aws_lambda \
--runtime "python3.8" \
--handler "s3-lambda/s3-lambda.lambda_handler" \
--memory-size 128 \
--timeout 30 \
--role "arn:aws:iam::$aws_account_id:role/$aws_role" \
--zip-file "fileb://./s3-lambda.zip"
# add permissions to s3 bucket to invoke lambda
LambdaFunctionArn="arn:aws:lambda:us-east-1:$aws_account_id:function:s3-lambda"
aws s3api put-bucket-notification-configuration \
--region "$aws_region" \
--bucket "$aws_bucket" \
--notification-configuration '{
"LambdaFunctionConfigurations": [{
"LambdaFunctionArn": "'"$LambdaFunctionArn"'",
"Events": ["s3:ObjectCreated:*"]
}]
}'
aws s3api put-bucket-notification-configuration \
--region "$aws_region" \
--bucket "$aws_bucket" \
--notification-configuration '{
"LambdaFunctionConfigurations": [{
"LambdaFunctionArn": "'"$LambdaFunctionArn"'",
"Events": ["s3:ObjectCreated:*"]
}]
}'
# create an sns topic and save the topic arn to a variable
topic_arn=$(aws sns create-topic --name s3-lambda-sns --output json | jq -r '.TopicArn')
# print the topic arn
echo "SNS Topic ARN: $topic_arn"
# Trigger SNS topic using lambda function
# Add sns topic using lambda function
aws sns subscribe \
--topic-arn "$topic_arn" \
--protocol email \
--notification-endpoint "$email_address"
# publish sns
aws sns publish \
--topic-arn "$topic_arn" \
--subject "A new object created in s3 bucket" \
--message "Hey, a new data object just got delievered into the s3 bucket $aws_bucket"

View File

@ -0,0 +1 @@
boto3==1.17.95

View File

@ -0,0 +1,38 @@
import boto3
import json
def lambda_handler(event, context):
# i want to know that event thing
print(event)
# extract relevant information from the s3 event trigger
bucket_name=event['Records'][0]['s3']['bucket']['name']
object_key=event['Records'][0]['s3']['object']['key']
# perform desired operations with the upload file
print(f"File '{object_key}' was uploaded to bucket '{bucket_name}'")
# example: send a notification via sns
sns_client=boto3.client('sns')
topic_arn='arn:aws:sns:us-east-1:<account-id>:s3-lambda-sns'
sns_client.publish(
TopicArn=topic_arn,
Subject='s3 object created !!',
Message=f"File '{object_key}' was uploaded to bucket '{bucket_name}"
)
# Example: Trigger another Lambda function
# lambda_client = boto3.client('lambda')
# target_function_name = 'my-another-lambda-function'
# lambda_client.invoke(
# FunctionName=target_function_name,
# InvocationType='Event',
# Payload=json.dumps({'bucket_name': bucket_name, 'object_key': object_key})
# )
# in case of queuing and other objective similar to the netflix flow of triggering
return {
'statusCode': 200,
'body': json.dumps("Lambda function executed successfully !!")
}

Binary file not shown.

After

Width:  |  Height:  |  Size: 79 KiB