mirror of
https://github.com/simple-login/app.git
synced 2024-11-18 01:40:38 +01:00
0da1811311
* Cleanup tasks * Update * Added tests * Create cron job * Delete old data cron * Fix import * import fix * Added delete + script to disable pgp for proton mboxes
115 lines
3.2 KiB
Python
115 lines
3.2 KiB
Python
import os
|
|
from io import BytesIO
|
|
from typing import Optional
|
|
|
|
import boto3
|
|
import requests
|
|
|
|
from app import config
|
|
from app.log import LOG
|
|
|
|
_s3_client = None
|
|
|
|
|
|
def _get_s3client():
|
|
global _s3_client
|
|
if _s3_client is None:
|
|
args = {
|
|
"aws_access_key_id": config.AWS_ACCESS_KEY_ID,
|
|
"aws_secret_access_key": config.AWS_SECRET_ACCESS_KEY,
|
|
"region_name": config.AWS_REGION,
|
|
}
|
|
if config.AWS_ENDPOINT_URL:
|
|
args["endpoint_url"] = config.AWS_ENDPOINT_URL
|
|
_s3_client = boto3.client("s3", **args)
|
|
return _s3_client
|
|
|
|
|
|
def upload_from_bytesio(key: str, bs: BytesIO, content_type="application/octet-stream"):
|
|
bs.seek(0)
|
|
|
|
if config.LOCAL_FILE_UPLOAD:
|
|
file_path = os.path.join(config.UPLOAD_DIR, key)
|
|
file_dir = os.path.dirname(file_path)
|
|
os.makedirs(file_dir, exist_ok=True)
|
|
with open(file_path, "wb") as f:
|
|
f.write(bs.read())
|
|
|
|
else:
|
|
_get_s3client().put_object(
|
|
Bucket=config.BUCKET,
|
|
Key=key,
|
|
Body=bs,
|
|
ContentType=content_type,
|
|
)
|
|
|
|
|
|
def upload_email_from_bytesio(path: str, bs: BytesIO, filename):
|
|
bs.seek(0)
|
|
|
|
if config.LOCAL_FILE_UPLOAD:
|
|
file_path = os.path.join(config.UPLOAD_DIR, path)
|
|
file_dir = os.path.dirname(file_path)
|
|
os.makedirs(file_dir, exist_ok=True)
|
|
with open(file_path, "wb") as f:
|
|
f.write(bs.read())
|
|
|
|
else:
|
|
_get_s3client().put_object(
|
|
Bucket=config.BUCKET,
|
|
Key=path,
|
|
Body=bs,
|
|
# Support saving a remote file using Http header
|
|
# Also supports Safari. More info at
|
|
# https://github.com/eligrey/FileSaver.js/wiki/Saving-a-remote-file#using-http-header
|
|
ContentDisposition=f'attachment; filename="{filename}.eml";',
|
|
)
|
|
|
|
|
|
def download_email(path: str) -> Optional[str]:
|
|
if config.LOCAL_FILE_UPLOAD:
|
|
file_path = os.path.join(config.UPLOAD_DIR, path)
|
|
with open(file_path, "rb") as f:
|
|
return f.read()
|
|
resp = _get_s3client().get_object(
|
|
Bucket=config.BUCKET,
|
|
Key=path,
|
|
)
|
|
if not resp or "Body" not in resp:
|
|
return None
|
|
return resp["Body"].read
|
|
|
|
|
|
def upload_from_url(url: str, upload_path):
|
|
r = requests.get(url)
|
|
upload_from_bytesio(upload_path, BytesIO(r.content))
|
|
|
|
|
|
def get_url(key: str, expires_in=3600) -> str:
|
|
if config.LOCAL_FILE_UPLOAD:
|
|
return config.URL + "/static/upload/" + key
|
|
else:
|
|
return _get_s3client().generate_presigned_url(
|
|
ExpiresIn=expires_in,
|
|
ClientMethod="get_object",
|
|
Params={"Bucket": config.BUCKET, "Key": key},
|
|
)
|
|
|
|
|
|
def delete(path: str):
|
|
if config.LOCAL_FILE_UPLOAD:
|
|
file_path = os.path.join(config.UPLOAD_DIR, path)
|
|
os.remove(file_path)
|
|
else:
|
|
_get_s3client().delete_object(Bucket=config.BUCKET, Key=path)
|
|
|
|
|
|
def create_bucket_if_not_exists():
|
|
s3client = _get_s3client()
|
|
buckets = s3client.list_buckets()
|
|
for bucket in buckets["Buckets"]:
|
|
if bucket["Name"] == config.BUCKET:
|
|
LOG.i("Bucket already exists")
|
|
return
|
|
s3client.create_bucket(Bucket=config.BUCKET)
|
|
LOG.i(f"Bucket {config.BUCKET} created")
|