adding initial boilerplates
This commit is contained in:
2
README.rst
Normal file
2
README.rst
Normal file
@@ -0,0 +1,2 @@
|
||||
Boilerplates
|
||||
============
|
||||
257
python/aws/s3/downloadFromS3.py
Normal file
257
python/aws/s3/downloadFromS3.py
Normal file
@@ -0,0 +1,257 @@
|
||||
import boto3
|
||||
import botocore
|
||||
import os
|
||||
from datetime import date, timedelta
|
||||
import sys
|
||||
import logging
|
||||
import argparse
|
||||
import glob
|
||||
|
||||
# Set Global Variables
|
||||
log_location = 'pull.log'
|
||||
|
||||
remote_folder = ['bot_predictions/']
|
||||
remote_file_prefix = ['blocking_suggestions_']
|
||||
|
||||
append_date = ['True']
|
||||
date_format = ['%Y-%m-%d']
|
||||
|
||||
bucket = ['td-ingest-storage-williamhill']
|
||||
access_key = ['AKIAYJXVWMRHQ2OGNHLA']
|
||||
secret_key = ['0/4wxdBmpiU3gK1QHLk4me0zj2RHuNAcSOfgJm1B']
|
||||
|
||||
|
||||
class downloadFiles(object):
|
||||
"""docstring for downloadFiles"""
|
||||
today = date.today()
|
||||
yesterday = date.today() - timedelta(1)
|
||||
|
||||
def __init__(self,
|
||||
client,
|
||||
resource,
|
||||
bucket,
|
||||
remote_folder,
|
||||
remote_file_prefix,
|
||||
local_path,
|
||||
append_date=False,
|
||||
date_format=''):
|
||||
super(downloadFiles, self).__init__()
|
||||
self.client = client
|
||||
self.resource = resource
|
||||
self.bucket = bucket
|
||||
self.append_date = append_date
|
||||
self.date_format = date_format
|
||||
self.remote_folder = self._folder_fixer(remote_folder)
|
||||
self.dest = f'{self.remote_folder!s}{remote_file_prefix!s}'
|
||||
self.local_path = local_path
|
||||
self.remote_list, self.local_list, self.local_file_list = \
|
||||
(list() for _ in range(3))
|
||||
|
||||
@staticmethod
|
||||
def generate_date(date_format, relative_day='today'):
|
||||
if relative_day == 'today':
|
||||
date = downloadFiles.today.strftime(date_format)
|
||||
elif relative_day == 'yesterday':
|
||||
date = downloadFiles.yesterday.strftime(date_format)
|
||||
return date
|
||||
|
||||
@staticmethod
|
||||
def _folder_fixer(folder):
|
||||
try:
|
||||
if folder[-1] != '/':
|
||||
folder = f'{folder}/'
|
||||
except IndexError:
|
||||
folder = ''
|
||||
return folder
|
||||
|
||||
def get_path(self):
|
||||
if self.local_path:
|
||||
self.local_path = self._folder_fixer(self.local_path)
|
||||
logger.info(f'path entered is {self.local_path}')
|
||||
return self
|
||||
else:
|
||||
self.local_path = os.getcwd()
|
||||
self.local_path = self._folder_fixer(self.local_path)
|
||||
self.local_path = f'{self.local_path}blocking_suggestions/'
|
||||
logger.info(f'no path entered, using current directory '
|
||||
f'{self.local_path}')
|
||||
return self
|
||||
|
||||
def get_files(self):
|
||||
if self.append_date:
|
||||
date_today = self.generate_date(self.date_format)
|
||||
date_yesterday = self.generate_date(self.date_format, 'yesterday')
|
||||
else:
|
||||
date_today = ''
|
||||
date_yesterday = ''
|
||||
self.dest_list = []
|
||||
self.dest_list.append(f'{self.dest!s}{date_today!s}')
|
||||
self.dest_list.append(f'{self.dest!s}{date_yesterday!s}')
|
||||
for dest in self.dest_list:
|
||||
paginator = self.client.get_paginator('list_objects')
|
||||
iterator = paginator.paginate(Bucket=self.bucket, Prefix=dest)
|
||||
self.filtered = iterator.search('Contents[*].Key')
|
||||
for i in self.filtered:
|
||||
try:
|
||||
self.remote_list.append(i)
|
||||
self.local_list.append(
|
||||
f'{self.local_path}{i[len(self.remote_folder):]}'
|
||||
)
|
||||
self.local_file_list.append(
|
||||
f'{i[len(self.remote_folder):]}'
|
||||
)
|
||||
except TypeError:
|
||||
logger.info('no files available to download -- exiting')
|
||||
raise SystemExit
|
||||
logger.debug(f'remote files are {self.remote_list}')
|
||||
logger.debug(f'saving files locally to {self.local_list}')
|
||||
return self
|
||||
|
||||
def get_history(self):
|
||||
self.history_file = f'{self.local_path}.history.txt'
|
||||
try:
|
||||
logger.info('opening history file')
|
||||
open(self.history_file, 'a').close()
|
||||
pass
|
||||
except FileNotFoundError:
|
||||
logger.critical('history file cannot be found or created'
|
||||
' - check permissions of the folder.')
|
||||
raise
|
||||
self.history_list = \
|
||||
[line.rstrip('\n') for line in open(self.history_file)]
|
||||
return self
|
||||
|
||||
def remove_files(self):
|
||||
logger.info('attempting to clear current files')
|
||||
current_files = glob.glob(f'{self.local_path}[!history.txt]*')
|
||||
if current_files:
|
||||
for i in current_files:
|
||||
try:
|
||||
os.remove(i)
|
||||
logger.info(f'removed {i}')
|
||||
except OSError:
|
||||
logger.exception('Error:')
|
||||
else:
|
||||
logger.info('no files to remove')
|
||||
return self
|
||||
|
||||
def download_files(self):
|
||||
for remote_file, local_file_with_path, local_file in zip(
|
||||
self.remote_list, self.local_list, self.local_file_list):
|
||||
if local_file not in self.history_list:
|
||||
with open(local_file_with_path, 'wb'), \
|
||||
open(self.history_file, 'a') as hist:
|
||||
try:
|
||||
self.resource.Bucket(self.bucket).download_file(
|
||||
remote_file, local_file_with_path)
|
||||
hist.write(f'\n{local_file}')
|
||||
logger.info(f'downloaded {local_file}')
|
||||
except botocore.exceptions.ClientError as e:
|
||||
if e.response['Error']['Code'] == '404':
|
||||
print(f'The object {remote_file} does not exist.')
|
||||
else:
|
||||
raise
|
||||
if local_file in self.history_list:
|
||||
logger.debug(f'{local_file} already downloaded - skipping')
|
||||
return self
|
||||
|
||||
|
||||
def _call():
|
||||
global args, debug
|
||||
parser = argparse.ArgumentParser(description="""
|
||||
downloads any new files for the current day from an S3 bucket. \
|
||||
uses a local history file to track what has been \
|
||||
previously downloaded in the download path.
|
||||
""")
|
||||
parser.add_argument('--path', type=str,
|
||||
help='enter pull path to download to. if left \
|
||||
blank will use the same location as the script.',
|
||||
default='')
|
||||
parser.add_argument('--debug', action='store_true', default=False,
|
||||
help='Use this to log DEBUG information.')
|
||||
|
||||
args = parser.parse_args()
|
||||
debug = vars(args)['debug']
|
||||
|
||||
if debug:
|
||||
logger.setLevel(logging.DEBUG)
|
||||
else:
|
||||
logger.setLevel(logging.INFO)
|
||||
|
||||
main(_clients=_clients,
|
||||
_resources=_resources,
|
||||
_buckets=_buckets,
|
||||
remote_folder=remote_folder,
|
||||
remote_file_prefix=remote_file_prefix,
|
||||
append_date=append_date,
|
||||
date_format=date_format,
|
||||
**vars(args))
|
||||
|
||||
|
||||
def main(*args,
|
||||
_clients={'client0': ''},
|
||||
_resources={'resource0': ''},
|
||||
_buckets={'bucket0': ''},
|
||||
remote_folder=[''],
|
||||
remote_file_prefix=[''],
|
||||
append_date=['True'],
|
||||
date_format=[''],
|
||||
path='',
|
||||
**kwargs):
|
||||
logger.info('========= SCRIPT STARTED =========')
|
||||
instance = downloadFiles(client=_clients['client0'],
|
||||
resource=_resources['resource0'],
|
||||
bucket=_buckets['bucket0'],
|
||||
remote_folder=remote_folder[0],
|
||||
remote_file_prefix=remote_file_prefix[0],
|
||||
local_path=path,
|
||||
append_date=append_date[0],
|
||||
date_format=date_format[0])
|
||||
instance.get_path().get_files().get_history().remove_files()\
|
||||
.download_files()
|
||||
logger.info('========= SCRIPT FINISHED =========')
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
args, debug = '', ''
|
||||
|
||||
# define logging
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
c_handler = logging.StreamHandler(sys.stdout)
|
||||
f_handler = logging.FileHandler(log_location)
|
||||
|
||||
c_format = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')
|
||||
f_format = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')
|
||||
c_handler.setFormatter(c_format)
|
||||
f_handler.setFormatter(f_format)
|
||||
|
||||
logger.addHandler(c_handler)
|
||||
logger.addHandler(f_handler)
|
||||
|
||||
_clients = {}
|
||||
_resources = {}
|
||||
_buckets = {}
|
||||
|
||||
for i in range(0, len(bucket)):
|
||||
_clients[f'client{i}'] =\
|
||||
boto3.client('s3',
|
||||
aws_access_key_id=f'{access_key[i]}',
|
||||
aws_secret_access_key=f'{secret_key[i]}')
|
||||
_resources[f'resource{i}'] =\
|
||||
boto3.resource('s3',
|
||||
aws_access_key_id=f'{access_key[i]}',
|
||||
aws_secret_access_key=f'{secret_key[i]}')
|
||||
_buckets[f'bucket{i}'] = f'{bucket[i]}'
|
||||
|
||||
try:
|
||||
_length = len(remote_folder)
|
||||
if _length == 0:
|
||||
remote_folder = ['']
|
||||
elif remote_folder[0] == 'root':
|
||||
remote_folder = ['']
|
||||
else:
|
||||
pass
|
||||
except NameError:
|
||||
remote_folder = ['']
|
||||
_call()
|
||||
156
python/aws/s3/getS3Info.py
Normal file
156
python/aws/s3/getS3Info.py
Normal file
@@ -0,0 +1,156 @@
|
||||
import boto3
|
||||
import logging
|
||||
import os
|
||||
import datetime
|
||||
from splunk_hec_handler import SplunkHecHandler
|
||||
import warnings
|
||||
import customLogLevel
|
||||
from socket import gaierror
|
||||
|
||||
|
||||
class getFileDetails(object):
|
||||
"""Query an S3 bucket for information about files stored
|
||||
|
||||
Required arguments:
|
||||
bucket: a string with the name of the bucket
|
||||
remoteFiles: a string containing the path and filename of the files
|
||||
client: a boto3 s3 client
|
||||
resource: a boto3 s3 resource
|
||||
|
||||
Optional arguments:
|
||||
logger: a splunk HEC handler for the logging module
|
||||
"""
|
||||
|
||||
def __init__(self,
|
||||
bucket: str,
|
||||
remoteFiles: str,
|
||||
client: boto3.client,
|
||||
resource: boto3.resource,
|
||||
logger: logging.getLogger = None
|
||||
):
|
||||
super().__init__()
|
||||
self.bucket = bucket
|
||||
self.remoteFiles = remoteFiles
|
||||
self.client = client
|
||||
self.resource = resource
|
||||
self.logger = logger
|
||||
|
||||
@staticmethod
|
||||
def formatFolder(remoteFolder: str):
|
||||
try:
|
||||
if remoteFolder[-1] != '/':
|
||||
remoteFolder = f'{remoteFolder}/'
|
||||
except IndexError:
|
||||
remoteFolder = ''
|
||||
return remoteFolder
|
||||
|
||||
@staticmethod
|
||||
def generateDate(date_format='%Y-%m-%d',
|
||||
time: datetime.datetime = None
|
||||
):
|
||||
"""
|
||||
Generates a human readable time string for a
|
||||
datetime.datetime object.
|
||||
By default will use today with %Y-%m-%d format """
|
||||
if time is None:
|
||||
time = datetime.date.today()
|
||||
date = time.strftime(date_format)
|
||||
return date
|
||||
|
||||
@staticmethod
|
||||
def getEpochTime(time: datetime.datetime):
|
||||
epoch = datetime.datetime.timestamp(time)
|
||||
return epoch
|
||||
|
||||
def getS3Files(self):
|
||||
self.paginator = self.client.get_paginator('list_objects')
|
||||
self.iterator = self.paginator.paginate(Bucket=self.bucket,
|
||||
Prefix=self.remoteFiles)
|
||||
self.filtered = self.iterator.search('Contents[*]')
|
||||
self.fileDict = dict()
|
||||
counter = 0
|
||||
|
||||
for i in self.filtered:
|
||||
now = self.getEpochTime(datetime.datetime.now())
|
||||
i.update({'_time': now})
|
||||
i['LastModified'] = self.getEpochTime(i['LastModified'])
|
||||
self.fileDict[counter] = i
|
||||
counter += 1
|
||||
return self
|
||||
|
||||
def sendToSplunk(self):
|
||||
for i in self.fileDict.values():
|
||||
try:
|
||||
self.logger.splunk(i)
|
||||
except AttributeError:
|
||||
raise Exception("No logger level exists for the custom Splunk"
|
||||
" level. Try creating a customLogLevel for"
|
||||
" splunk and try again.")
|
||||
return self
|
||||
|
||||
|
||||
def _call():
|
||||
# Define env variables for AWS boto3
|
||||
os.environ['AWS_PROFILE'] = 'netacea'
|
||||
os.environ['AWS_DEFAULT_REGION'] = 'eu-west-1'
|
||||
|
||||
# Define bucket and file names
|
||||
bucket = 'td-ingest-storage-williamhill'
|
||||
remoteFolder = 'bot_predictions/'
|
||||
remoteFilePrefix = 'blocking_suggestions_'
|
||||
append_date = True
|
||||
date_format = '%Y-%m-%d'
|
||||
|
||||
# Define splunk hec handler for logging
|
||||
try:
|
||||
splunk_handler = SplunkHecHandler('sc1uxpremn81.prod.williamhill.plc',
|
||||
'ea641e31-870e-4f5f-965f'
|
||||
'-57e0c9a2aa3d',
|
||||
port=8088, proto='https',
|
||||
ssl_verify=False,
|
||||
sourcetype='httpevent')
|
||||
except gaierror:
|
||||
raise SystemExit
|
||||
|
||||
logger = logging.getLogger('SplunkHecHandlerExample')
|
||||
logger.setLevel(logging.DEBUG)
|
||||
logger.addHandler(splunk_handler)
|
||||
customLogLevel.addLoggingLevel('splunk', 15, logging)
|
||||
|
||||
main(bucket,
|
||||
remoteFolder,
|
||||
remoteFilePrefix,
|
||||
append_date,
|
||||
date_format,
|
||||
logger)
|
||||
|
||||
|
||||
def main(bucket: str,
|
||||
remoteFolder: str,
|
||||
remoteFilePrefix: str,
|
||||
append_date: bool,
|
||||
date_format: str,
|
||||
logger: logging.getLogger):
|
||||
|
||||
if append_date:
|
||||
remoteFileDate = getFileDetails.generateDate(date_format)
|
||||
else:
|
||||
remoteFileDate = ''
|
||||
|
||||
remoteFolder = getFileDetails.formatFolder(remoteFolder)
|
||||
remoteFiles = f'{remoteFolder}{remoteFilePrefix}{remoteFileDate}'
|
||||
|
||||
client = boto3.client('s3')
|
||||
resource = boto3.resource('s3')
|
||||
|
||||
instance = getFileDetails(bucket,
|
||||
remoteFiles,
|
||||
client,
|
||||
resource,
|
||||
logger)
|
||||
instance.getS3Files().sendToSplunk()
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
warnings.filterwarnings("ignore")
|
||||
_call()
|
||||
141
python/aws/sqs/receiveFromSQS.py
Normal file
141
python/aws/sqs/receiveFromSQS.py
Normal file
@@ -0,0 +1,141 @@
|
||||
import boto3
|
||||
import base64
|
||||
import math
|
||||
import json
|
||||
from datetime import datetime
|
||||
|
||||
|
||||
class receiveFromSQS(object):
|
||||
"""docstring for receiveFromSQS"""
|
||||
|
||||
def __init__(self, session, queueURL):
|
||||
super(receiveFromSQS, self).__init__()
|
||||
self.session = session
|
||||
self.sqs = session.client('sqs')
|
||||
self.queueURL = queueURL
|
||||
self.messages = []
|
||||
|
||||
@classmethod
|
||||
def createSession(cls, profileName, queueURL):
|
||||
session = boto3.Session(profile_name=profileName)
|
||||
return cls(session, queueURL)
|
||||
|
||||
def getQueueLength(self):
|
||||
attributeNames = ['ApproximateNumberOfMessages']
|
||||
self.queueAttributes = self.sqs.get_queue_attributes(
|
||||
QueueUrl=self.queueURL, AttributeNames=attributeNames
|
||||
)
|
||||
self.queueLength = int(
|
||||
self.queueAttributes['Attributes']['ApproximateNumberOfMessages']
|
||||
)
|
||||
return self.queueLength
|
||||
|
||||
def _receiveSQSMessage(
|
||||
self, totalNumberOfMessages, maxNumberOfMessages=10
|
||||
):
|
||||
self.resp = []
|
||||
self.loops = int(
|
||||
math.ceil(totalNumberOfMessages / maxNumberOfMessages)
|
||||
)
|
||||
loopTrack = 0
|
||||
if totalNumberOfMessages <= 10:
|
||||
maxNumberOfMessages = totalNumberOfMessages
|
||||
else:
|
||||
maxNumberOfMessagesFinal = 10 - (
|
||||
(self.loops * maxNumberOfMessages) - totalNumberOfMessages
|
||||
)
|
||||
print(maxNumberOfMessagesFinal)
|
||||
if self.loops == 0:
|
||||
raise RuntimeError('No messages in the queue')
|
||||
for i in range(0, self.loops):
|
||||
if loopTrack == self.loops - 1 and totalNumberOfMessages > 10:
|
||||
maxNumberOfMessages = maxNumberOfMessagesFinal
|
||||
self.resp.append(
|
||||
self.sqs.receive_message(
|
||||
QueueUrl=self.queueURL,
|
||||
MaxNumberOfMessages=maxNumberOfMessages,
|
||||
)
|
||||
)
|
||||
try:
|
||||
entries = [
|
||||
{
|
||||
'Id': msg['MessageId'],
|
||||
'ReceiptHandle': msg['ReceiptHandle'],
|
||||
}
|
||||
for msg in self.resp[i]['Messages']
|
||||
]
|
||||
self._deleteSQSMessages(entries)
|
||||
loopTrack += 1
|
||||
except KeyError:
|
||||
print("No messages in the queue")
|
||||
return self
|
||||
|
||||
def _extractMessageFromSQS(self, totalNumberOfMessages):
|
||||
self.extractedMessages = []
|
||||
self.receiptHandles = []
|
||||
try:
|
||||
for i in range(0, self.loops):
|
||||
_loops = len(self.resp[i]['Messages'])
|
||||
for j in range(0, _loops):
|
||||
if 'Messages' in self.resp[i]:
|
||||
self.extractedMessages.append(
|
||||
self.resp[i]['Messages'][j]['Body']
|
||||
)
|
||||
else:
|
||||
print('No messages in the queue')
|
||||
except KeyError:
|
||||
print('No messages in the queue key')
|
||||
return self
|
||||
|
||||
def _deleteSQSMessages(self, entries):
|
||||
self.respDelete = self.sqs.delete_message_batch(
|
||||
QueueUrl=self.queueURL, Entries=entries
|
||||
)
|
||||
if len(self.respDelete['Successful']) != len(entries):
|
||||
raise RuntimeError(
|
||||
f'Failed to delete messages: entries={entries!r}'
|
||||
f' resp={self.respDelete!r}'
|
||||
)
|
||||
|
||||
def _decodeMessages(self):
|
||||
if len(self.extractedMessages) == 0:
|
||||
print('No messages to process')
|
||||
else:
|
||||
for message in self.extractedMessages:
|
||||
decoded = base64.b64decode(message).decode()
|
||||
self.messages.append(decoded)
|
||||
return self
|
||||
|
||||
def receiveAllMessages(self, b64=True, _totalNumberOfMessages=None):
|
||||
if _totalNumberOfMessages is None:
|
||||
totalNumberOfMessages = self.getQueueLength()
|
||||
else:
|
||||
totalNumberOfMessages = _totalNumberOfMessages
|
||||
self._receiveSQSMessage(totalNumberOfMessages)
|
||||
self._extractMessageFromSQS(
|
||||
totalNumberOfMessages=totalNumberOfMessages
|
||||
)
|
||||
if b64:
|
||||
self._decodeMessages()
|
||||
else:
|
||||
self.messages = self.extractedMessages
|
||||
return self
|
||||
|
||||
def receiveNMessages(self, numberOfMessages, b64=True):
|
||||
self.receiveAllMessages(
|
||||
b64=b64, _totalNumberOfMessages=numberOfMessages
|
||||
)
|
||||
return self
|
||||
|
||||
# def generateOutput(self, outputType='json'):
|
||||
# if outputType == 'json':
|
||||
# self.output = json.dumps(self.messages)
|
||||
# return self.output
|
||||
|
||||
def savetoDisk(self, path):
|
||||
self.timeNow = datetime.now().strftime('%d-%m-%Y_%H:%M:%S')
|
||||
if len(self.messages) > 0:
|
||||
with open(f'{path}/{self.timeNow}.json', 'w+') as outputFile:
|
||||
json.dump(self.messages, outputFile)
|
||||
else:
|
||||
print('No messages to save')
|
||||
63
python/aws/sqs/sendToSQS.py
Normal file
63
python/aws/sqs/sendToSQS.py
Normal file
@@ -0,0 +1,63 @@
|
||||
import boto3
|
||||
import base64
|
||||
import hashlib
|
||||
from datetime import datetime
|
||||
import os
|
||||
import sys
|
||||
|
||||
sys.path.append(os.getcwd())
|
||||
from pullTrafficInfo import getTrafficInfo
|
||||
|
||||
|
||||
class sendToSQS(object):
|
||||
"""docstring for sendToSQS"""
|
||||
|
||||
def __init__(self, session, queueURL):
|
||||
super(sendToSQS, self).__init__()
|
||||
self.session = session
|
||||
self.sqs = session.client('sqs')
|
||||
self.queueURL = queueURL
|
||||
|
||||
@classmethod
|
||||
def createSession(cls, profileName, queueURL):
|
||||
session = boto3.Session(profile_name=profileName)
|
||||
return cls(session, queueURL)
|
||||
|
||||
def sendMessage(self, message, messageGroupId, b64=True, dedup=False):
|
||||
currentTime = datetime.now().strftime('%H:%M:%S.%f')
|
||||
if b64:
|
||||
message = (base64.b64encode(message.encode())).decode()
|
||||
if not dedup:
|
||||
dedupId = hashlib.md5((message + currentTime).encode()).hexdigest()
|
||||
msg = self.sqs.send_message(
|
||||
QueueUrl=self.queueURL,
|
||||
MessageBody=message,
|
||||
MessageGroupId=messageGroupId,
|
||||
MessageDeduplicationId=dedupId,
|
||||
)
|
||||
else:
|
||||
msg = self.sqs.send_message(
|
||||
QueueUrl=self.queueURL,
|
||||
MessageBody=message,
|
||||
MessageGroupId=messageGroupId,
|
||||
)
|
||||
if msg is not None:
|
||||
print(msg['MessageId'])
|
||||
|
||||
|
||||
# inst = sendToSQS.createSession(
|
||||
# profileName='plex-aws',
|
||||
# queueURL='https://sqs.eu-west-1.amazonaws.com'
|
||||
# '/745437999005/slack-bot.fifo',
|
||||
# )
|
||||
|
||||
# instM = (
|
||||
# getTrafficInfo.getTrafficURL('M62')
|
||||
# .findIncidents()
|
||||
# .getIncidentInformation()
|
||||
# .generateOutput()
|
||||
# )
|
||||
|
||||
# for _ in range(0, 5):
|
||||
# for item in instM.output:
|
||||
# inst.sendMessage(message=item, messageGroupId='slack-bot-M62')
|
||||
0
python/binaries/README.md
Normal file
0
python/binaries/README.md
Normal file
4
python/binaries/install.sh
Executable file
4
python/binaries/install.sh
Executable file
@@ -0,0 +1,4 @@
|
||||
curl -LOemoji https://github.com/dtomlinson91/emoji-cmd/releases/download/1.0/emoji
|
||||
chmod +x emoji
|
||||
mv ./emoji ~/.local/bin
|
||||
|
||||
15
python/cli-click/multipleSubCommands-use.py
Normal file
15
python/cli-click/multipleSubCommands-use.py
Normal file
@@ -0,0 +1,15 @@
|
||||
import click
|
||||
from complex.cli import pass_environment
|
||||
|
||||
|
||||
@click.command('init', short_help='Initializes a repo.')
|
||||
@click.argument('path', required=False, type=click.Path(resolve_path=True))
|
||||
@pass_environment
|
||||
def cli(ctx, path):
|
||||
"""Initializes a repository."""
|
||||
print(f'{ctx=}')
|
||||
print(f'{dir(ctx)=}')
|
||||
if path is None:
|
||||
path = ctx.home
|
||||
ctx.log('Initialized the repository in %s',
|
||||
click.format_filename(path))
|
||||
64
python/cli-click/multipleSubCommands.py
Normal file
64
python/cli-click/multipleSubCommands.py
Normal file
@@ -0,0 +1,64 @@
|
||||
import os
|
||||
import sys
|
||||
import click
|
||||
|
||||
CONTEXT_SETTINGS = dict(auto_envvar_prefix='COMPLEX')
|
||||
|
||||
|
||||
class Environment(object):
|
||||
|
||||
def __init__(self):
|
||||
self.verbose = False
|
||||
self.home = os.getcwd()
|
||||
|
||||
def log(self, msg, *args):
|
||||
"""Logs a message to stderr."""
|
||||
if args:
|
||||
msg %= args
|
||||
click.echo(msg, file=sys.stderr)
|
||||
|
||||
def vlog(self, msg, *args):
|
||||
"""Logs a message to stderr only if verbose is enabled."""
|
||||
if self.verbose:
|
||||
self.log(msg, *args)
|
||||
|
||||
|
||||
pass_environment = click.make_pass_decorator(Environment, ensure=True)
|
||||
cmd_folder = os.path.abspath(os.path.join(os.path.dirname(__file__),
|
||||
'commands'))
|
||||
|
||||
|
||||
class ComplexCLI(click.MultiCommand):
|
||||
|
||||
def list_commands(self, ctx):
|
||||
rv = []
|
||||
for filename in os.listdir(cmd_folder):
|
||||
if filename.endswith('.py') and \
|
||||
filename.startswith('cmd_'):
|
||||
rv.append(filename[4:-3])
|
||||
rv.sort()
|
||||
return rv
|
||||
|
||||
def get_command(self, ctx, name):
|
||||
try:
|
||||
if sys.version_info[0] == 2:
|
||||
name = name.encode('ascii', 'replace')
|
||||
mod = __import__('complex.commands.cmd_' + name,
|
||||
None, None, ['cli'])
|
||||
except ImportError:
|
||||
return
|
||||
return mod.cli
|
||||
|
||||
|
||||
@click.command(cls=ComplexCLI, context_settings=CONTEXT_SETTINGS)
|
||||
@click.option('--home', type=click.Path(exists=True, file_okay=False,
|
||||
resolve_path=True),
|
||||
help='Changes the folder to operate on.')
|
||||
@click.option('-v', '--verbose', is_flag=True,
|
||||
help='Enables verbose mode.')
|
||||
@pass_environment
|
||||
def cli(ctx, verbose, home):
|
||||
"""A complex command line interface."""
|
||||
ctx.verbose = verbose
|
||||
if home is not None:
|
||||
ctx.home = home
|
||||
87
python/cli-click/settingsOptionsClass.py
Normal file
87
python/cli-click/settingsOptionsClass.py
Normal file
@@ -0,0 +1,87 @@
|
||||
import click
|
||||
import sys
|
||||
import time
|
||||
|
||||
|
||||
def build_settings_option_class(settings_instance):
|
||||
def set_default(default_name):
|
||||
class Cls(click.Option):
|
||||
def __init__(self, *args, **kwargs):
|
||||
kwargs['default'] = getattr(settings_instance, default_name)
|
||||
super(Cls, self).__init__(*args, **kwargs)
|
||||
|
||||
def handle_parse_result(self, ctx, opts, args):
|
||||
obj = ctx.find_object(type(settings_instance))
|
||||
if obj is None:
|
||||
ctx.obj = settings_instance
|
||||
|
||||
return super(Cls, self).handle_parse_result(ctx, opts, args)
|
||||
|
||||
return Cls
|
||||
|
||||
return set_default
|
||||
|
||||
|
||||
class Settings(object):
|
||||
def __init__(self):
|
||||
self.instance_disk_size = 100
|
||||
self.instance_disk_type = 'pd-ssd'
|
||||
|
||||
|
||||
# import pudb; pudb.set_trace()
|
||||
settings = Settings()
|
||||
settings_option_cls = build_settings_option_class(settings)
|
||||
pass_settings = click.make_pass_decorator(Settings)
|
||||
|
||||
|
||||
@click.command()
|
||||
@click.help_option('-h', '--help')
|
||||
@click.option(
|
||||
'-s',
|
||||
'--disk-size',
|
||||
cls=settings_option_cls('instance_disk_size'),
|
||||
help="Disk size",
|
||||
show_default=True,
|
||||
type=int,
|
||||
)
|
||||
@click.option(
|
||||
'-t',
|
||||
'--disk-type',
|
||||
cls=settings_option_cls('instance_disk_type'),
|
||||
help="Disk type",
|
||||
show_default=True,
|
||||
type=click.Choice(['pd-standard', 'pd-ssd']),
|
||||
)
|
||||
@pass_settings
|
||||
def create(settings_test, disk_size, disk_type):
|
||||
print(f'{settings_test.instance_disk_type=}')
|
||||
print(f'{dir(settings_test)=}')
|
||||
print(disk_size)
|
||||
print(disk_type)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
commands = (
|
||||
'-t pd-standard -s 200',
|
||||
'-t pd-standard',
|
||||
'-s 200',
|
||||
'',
|
||||
'--help',
|
||||
)
|
||||
|
||||
time.sleep(1)
|
||||
print('Click Version: {}'.format(click.__version__))
|
||||
print('Python Version: {}'.format(sys.version))
|
||||
for cmd in commands:
|
||||
try:
|
||||
time.sleep(0.1)
|
||||
print('-----------')
|
||||
print('> ' + cmd)
|
||||
time.sleep(0.1)
|
||||
create(cmd.split())
|
||||
|
||||
except BaseException as exc:
|
||||
if str(exc) != '0' and not isinstance(
|
||||
exc, (click.ClickException, SystemExit)
|
||||
):
|
||||
raise
|
||||
87
python/documentation/_static/custom.css
Normal file
87
python/documentation/_static/custom.css
Normal file
@@ -0,0 +1,87 @@
|
||||
@import url("css/theme.css");
|
||||
|
||||
.modname {
|
||||
font-size: 0.8em;
|
||||
opacity: 0.4;
|
||||
}
|
||||
.modname::before {
|
||||
content: '- ';
|
||||
}
|
||||
.title {
|
||||
font-weight: bold;
|
||||
font-size: 1.2em;
|
||||
background-color: #eee;
|
||||
display: block;
|
||||
padding: 1px 5px;
|
||||
border-left: 2px solid #ddd;
|
||||
}
|
||||
|
||||
/*colour of the text in the toc*/
|
||||
.wy-menu-vertical header, .wy-menu-vertical p.caption{
|
||||
color: #b26d46;;
|
||||
}
|
||||
|
||||
/*colour of the top left header*/
|
||||
.wy-side-nav-search{
|
||||
background-color: #31465a;
|
||||
}
|
||||
|
||||
/*colours of the class definitions*/
|
||||
.rst-content dl:not(.docutils) dt{
|
||||
background: #e2d7d1;
|
||||
color: #0b2852;
|
||||
border-top: solid 3px #31465a;
|
||||
}
|
||||
|
||||
/*colour of the link in the class defintions*/
|
||||
.rst-content .viewcode-link, .rst-content .viewcode-back{
|
||||
color: #4b674a;
|
||||
}
|
||||
|
||||
/*colour of the function definitions*/
|
||||
.rst-content dl:not(.docutils) dl dt{
|
||||
border-left: solid 3px #31465a;
|
||||
background: #e2d7d1;
|
||||
color: #0b2852;
|
||||
}
|
||||
|
||||
/*colour of the link in the function definitions*/
|
||||
.rst-content .viewcode-link, .rst-content .viewcode-back{
|
||||
color: #4b674a;
|
||||
}
|
||||
|
||||
/*edit the width of the body*/
|
||||
.wy-nav-content{
|
||||
max-width: 1200px;
|
||||
}
|
||||
|
||||
/*code example blocks*/
|
||||
.rst-content pre.literal-block, .rst-content div[class^="highlight"] pre, .rst-content .linenodiv pre{
|
||||
background: #b4bec8;
|
||||
color: black;
|
||||
/*border-style: solid;*/
|
||||
/*border-width: thin;*/
|
||||
}
|
||||
|
||||
/*colour of inline code blocks using ``*/
|
||||
.rst-content tt.literal, .rst-content tt.literal, .rst-content code.literal{
|
||||
color: #b26d46;
|
||||
}
|
||||
|
||||
/* Change code blocks font and bump up font size slightly (normally 12px)*/
|
||||
.rst-content pre.literal-block, .rst-content div[class^="highlight"] pre, .rst-content .linenodiv pre {
|
||||
font-family: 'Inconsolata', monospace !important;
|
||||
font-size: 14px !important;
|
||||
white-space: pre-wrap;
|
||||
}
|
||||
|
||||
/* Change code descriptions and literal blocks (inline code via ``) to match the normal font size being used in the sphinx_RTD_theme text (normally 14px)*/
|
||||
.rst-content dl:not(.docutils) tt.descname,.rst-content dl:not(.docutils) tt.descclassname,.rst-content dl:not(.docutils) tt.descname,.rst-content dl:not(.docutils) code.descname,.rst-content dl:not(.docutils) tt.descclassname,.rst-content dl:not(.docutils) code.descclassname, code.docutils {
|
||||
font-family: 'Inconsolata', monospace !important;
|
||||
font-size: 14px !important;
|
||||
}
|
||||
|
||||
/*variables text*/
|
||||
dl.class > dd > table.docutils.field-list tbody tr.field-odd.field th.field-name::before{
|
||||
content: '(Class Attributes) ';
|
||||
}
|
||||
17
python/documentation/changelog.rst
Normal file
17
python/documentation/changelog.rst
Normal file
@@ -0,0 +1,17 @@
|
||||
Version history
|
||||
================
|
||||
|
||||
0.3.0 (29/11/19)
|
||||
----------------
|
||||
|
||||
- Adding badges to documentation
|
||||
|
||||
0.2.0 (28/11/19)
|
||||
----------------
|
||||
|
||||
- Updating documentation
|
||||
|
||||
0.1.0 (28/11/19)
|
||||
----------------
|
||||
|
||||
- Initial release.
|
||||
95
python/documentation/conf.py
Normal file
95
python/documentation/conf.py
Normal file
@@ -0,0 +1,95 @@
|
||||
# Configuration file for the Sphinx documentation builder.
|
||||
#
|
||||
# This file only contains a selection of the most common options. For a full
|
||||
# list see the documentation:
|
||||
# https://www.sphinx-doc.org/en/master/usage/configuration.html
|
||||
|
||||
# -- Path setup --------------------------------------------------------------
|
||||
|
||||
# If extensions (or modules to document with autodoc) are in another directory,
|
||||
# add these directories to sys.path here. If the directory is relative to the
|
||||
# documentation root, use os.path.abspath to make it absolute, like shown here.
|
||||
#
|
||||
# import os
|
||||
# import sys
|
||||
# sys.path.insert(0, os.path.abspath('.'))
|
||||
import plex_posters
|
||||
from plex_posters.__version__ import __version__ as version
|
||||
import sphinx_rtd_theme
|
||||
|
||||
# -- Project information -----------------------------------------------------
|
||||
|
||||
project = 'plex_posters'
|
||||
copyright = '2019, Daniel Tomlinson'
|
||||
author = 'Daniel Tomlinson'
|
||||
|
||||
# The full version, including alpha/beta/rc tags
|
||||
release = version
|
||||
|
||||
|
||||
# -- General configuration ---------------------------------------------------
|
||||
|
||||
# Add any Sphinx extension module names here, as strings. They can be
|
||||
# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
|
||||
# ones.
|
||||
extensions = [
|
||||
'sphinx.ext.autodoc',
|
||||
'sphinx.ext.viewcode',
|
||||
'sphinx.ext.napoleon',
|
||||
'sphinx.ext.todo',
|
||||
]
|
||||
|
||||
# -- Napoleon Settings -----------------------------------------------------
|
||||
napoleon_google_docstring = False
|
||||
napoleon_numpy_docstring = True
|
||||
napoleon_include_init_with_doc = True
|
||||
napoleon_include_private_with_doc = False
|
||||
napoleon_include_special_with_doc = False
|
||||
napoleon_use_admonition_for_examples = False
|
||||
napoleon_use_admonition_for_notes = False
|
||||
napoleon_use_admonition_for_references = False
|
||||
napoleon_use_ivar = True
|
||||
napoleon_use_param = True
|
||||
napoleon_use_rtype = True
|
||||
napoleon_use_keyword = True
|
||||
autodoc_member_order = 'bysource'
|
||||
|
||||
# Add any paths that contain templates here, relative to this directory.
|
||||
templates_path = ['_templates']
|
||||
|
||||
# The master toctree document.
|
||||
master_doc = 'index'
|
||||
|
||||
# List of patterns, relative to source directory, that match files and
|
||||
# directories to ignore when looking for source files.
|
||||
# This pattern also affects html_static_path and html_extra_path.
|
||||
exclude_patterns = []
|
||||
|
||||
|
||||
# -- Options for HTML output -------------------------------------------------
|
||||
|
||||
# The theme to use for HTML and HTML Help pages. See the documentation for
|
||||
# a list of builtin themes.
|
||||
#
|
||||
html_theme = "sphinx_rtd_theme"
|
||||
|
||||
# Add any paths that contain custom static files (such as style sheets) here,
|
||||
# relative to this directory. They are copied after the builtin static files,
|
||||
# so a file named "default.css" will overwrite the builtin "default.css".
|
||||
|
||||
html_theme = "sphinx_rtd_theme"
|
||||
html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
|
||||
html_static_path = ['_static']
|
||||
html_context = {'css_files': ['_static/custom.css']}
|
||||
html_theme_options = {
|
||||
'collapse_navigation': True,
|
||||
'display_version': True,
|
||||
'prev_next_buttons_location': 'both',
|
||||
#'navigation_depth': 3,
|
||||
}
|
||||
|
||||
# The name of the Pygments (syntax highlighting) style to use.
|
||||
pygments_style = 'sphinx'
|
||||
|
||||
# Enable todo
|
||||
todo_include_todos = True
|
||||
4
python/documentation/configuration.rst
Normal file
4
python/documentation/configuration.rst
Normal file
@@ -0,0 +1,4 @@
|
||||
Configuration
|
||||
=============
|
||||
|
||||
.. todo:: fill in configuration options
|
||||
5
python/documentation/global.rst
Normal file
5
python/documentation/global.rst
Normal file
@@ -0,0 +1,5 @@
|
||||
.. role:: modname
|
||||
:class: modname
|
||||
|
||||
.. role:: title
|
||||
:class: title
|
||||
3
python/documentation/index.rst
Normal file
3
python/documentation/index.rst
Normal file
@@ -0,0 +1,3 @@
|
||||
Table of Contents
|
||||
=================
|
||||
.. include:: toc.rst
|
||||
1
python/documentation/introduction.rst
Normal file
1
python/documentation/introduction.rst
Normal file
@@ -0,0 +1 @@
|
||||
.. include:: ../../README.rst
|
||||
7
python/documentation/modules/modules.rst
Normal file
7
python/documentation/modules/modules.rst
Normal file
@@ -0,0 +1,7 @@
|
||||
plex_posters
|
||||
============
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 4
|
||||
|
||||
plex_posters
|
||||
@@ -0,0 +1,9 @@
|
||||
.. include:: ../global.rst
|
||||
|
||||
Config :modname:`plex_posters.config.config`
|
||||
---------------------------------------------
|
||||
|
||||
.. automodule:: plex_posters.config.config
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
16
python/documentation/modules/plex_posters.config.rst
Normal file
16
python/documentation/modules/plex_posters.config.rst
Normal file
@@ -0,0 +1,16 @@
|
||||
.. include:: ../global.rst
|
||||
|
||||
Config :modname:`plex_posters.config`
|
||||
-------------------------------------
|
||||
|
||||
.. automodule:: plex_posters.config
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
|
||||
Submodules
|
||||
----------
|
||||
|
||||
.. toctree::
|
||||
|
||||
plex_posters.config.config
|
||||
9
python/documentation/modules/plex_posters.library.rst
Normal file
9
python/documentation/modules/plex_posters.library.rst
Normal file
@@ -0,0 +1,9 @@
|
||||
.. include:: ../global.rst
|
||||
|
||||
Library :modname:`plex_posters.library`
|
||||
----------------------------------------
|
||||
|
||||
.. automodule:: plex_posters.library
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
17
python/documentation/modules/plex_posters.rst
Normal file
17
python/documentation/modules/plex_posters.rst
Normal file
@@ -0,0 +1,17 @@
|
||||
.. include:: ../global.rst
|
||||
|
||||
plex_posters :modname:`plex_posters`
|
||||
------------------------------------
|
||||
|
||||
.. automodule:: plex_posters
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
|
||||
Subpackages
|
||||
-----------
|
||||
|
||||
.. toctree::
|
||||
|
||||
plex_posters.config
|
||||
plex_posters.library
|
||||
24
python/documentation/toc.rst
Normal file
24
python/documentation/toc.rst
Normal file
@@ -0,0 +1,24 @@
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:caption: Overview
|
||||
:titlesonly:
|
||||
|
||||
introduction
|
||||
configuration
|
||||
changelog
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 4
|
||||
:caption: Modules
|
||||
:titlesonly:
|
||||
|
||||
modules/plex_posters.rst
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 4
|
||||
:caption: Submodules
|
||||
:titlesonly:
|
||||
|
||||
modules/plex_posters.config.config.rst
|
||||
modules/plex_posters.library.rst
|
||||
|
||||
5
python/libraries/README.md
Normal file
5
python/libraries/README.md
Normal file
@@ -0,0 +1,5 @@
|
||||
Config will create a config class to store variables from a `config.toml` or from environment variables.
|
||||
|
||||
Library contains utility functions:
|
||||
- @export exports modules in an `__init__.py` by adding to `__all__`
|
||||
- set_config() allows variables to be set on the config class
|
||||
48
python/logging/customLogLevel.py
Normal file
48
python/logging/customLogLevel.py
Normal file
@@ -0,0 +1,48 @@
|
||||
omprehensively adds a new logging level to the `logging` module and the
|
||||
currently configured logging class.
|
||||
|
||||
`levelName` becomes an attribute of the `logging` module with the value
|
||||
`levelNum`. `methodName` becomes a convenience method for both `logging`
|
||||
itself and the class returned by `logging.getLoggerClass()` (usually just
|
||||
`logging.Logger`). If `methodName` is not specified, `levelName.lower()` is
|
||||
used.
|
||||
|
||||
To avoid accidental clobberings of existing attributes, this method will
|
||||
raise an `AttributeError` if the level name is already an attribute of the
|
||||
`logging` module or if the method name is already present
|
||||
|
||||
Example
|
||||
-------
|
||||
>>> addLoggingLevel('TRACE', logging.DEBUG - 5)
|
||||
>>> logging.getLogger(__name__).setLevel("TRACE")
|
||||
>>> logging.getLogger(__name__).trace('that worked')
|
||||
>>> logging.trace('so did this')
|
||||
>>> logging.TRACE
|
||||
5
|
||||
|
||||
"""
|
||||
if not methodName:
|
||||
methodName = levelName.lower()
|
||||
|
||||
if hasattr(logging, levelName):
|
||||
raise AttributeError(
|
||||
'{} already defined in logging module'.format(levelName))
|
||||
if hasattr(logging, methodName):
|
||||
raise AttributeError(
|
||||
'{} already defined in logging module'.format(methodName))
|
||||
if hasattr(logging.getLoggerClass(), methodName):
|
||||
raise AttributeError(
|
||||
'{} already defined in logger class'.format(methodName))
|
||||
|
||||
def logForLevel(self, message, *args, **kwargs):
|
||||
if self.isEnabledFor(levelNum):
|
||||
self._log(levelNum, message, args, **kwargs)
|
||||
|
||||
def logToRoot(message, *args, **kwargs):
|
||||
logging.log(levelNum, message, *args, **kwargs)
|
||||
|
||||
logging.addLevelName(levelNum, levelName)
|
||||
setattr(logging, levelName, levelNum)
|
||||
setattr(logging.getLoggerClass(), methodName, logForLevel)
|
||||
setattr(logging, methodName, logToRoot)
|
||||
|
||||
Reference in New Issue
Block a user