aa
This commit is contained in:
21
CsvLoader.py
21
CsvLoader.py
@@ -7,6 +7,7 @@ import logging
|
|||||||
import csv
|
import csv
|
||||||
import re
|
import re
|
||||||
import mariadb
|
import mariadb
|
||||||
|
import shutil
|
||||||
|
|
||||||
from asebat.timefmt import timestamp_fmt as ts
|
from asebat.timefmt import timestamp_fmt as ts
|
||||||
from asebat.timefmt import date_refmt as df
|
from asebat.timefmt import date_refmt as df
|
||||||
@@ -100,7 +101,7 @@ def callback_ase(ch, method, properties, body, config): #body è di tipo byte
|
|||||||
sql.write_db()
|
sql.write_db()
|
||||||
stmlst.clear()
|
stmlst.clear()
|
||||||
except:
|
except:
|
||||||
print("errore nell'inseriento")
|
print("errore nell'inserimento")
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
if len(stmlst) > 0:
|
if len(stmlst) > 0:
|
||||||
@@ -109,8 +110,24 @@ def callback_ase(ch, method, properties, body, config): #body è di tipo byte
|
|||||||
sql.write_db()
|
sql.write_db()
|
||||||
ch.basic_ack(delivery_tag=method.delivery_tag)
|
ch.basic_ack(delivery_tag=method.delivery_tag)
|
||||||
except:
|
except:
|
||||||
print("errore nell'inseriento")
|
print("errore nell'inserimento")
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
newFilename = msg[6].replace("received", "loaded")
|
||||||
|
newPath, filenameExt = os.path.split(newFilename)
|
||||||
|
try:
|
||||||
|
os.makedirs(newPath)
|
||||||
|
logging.info("PID {:>5} >> path {} created.".format(
|
||||||
|
os.getpid(), newPath))
|
||||||
|
except FileExistsError:
|
||||||
|
logging.info("PID {:>5} >> path {} already exists.".format(
|
||||||
|
os.getpid(), newPath))
|
||||||
|
try:
|
||||||
|
shutil.move(msg[6], newFilename)
|
||||||
|
logging.info("PID {:>5} >> {} moved into {}.".format(
|
||||||
|
os.getpid(), filenameExt, newFilename))
|
||||||
|
except OSError:
|
||||||
|
logging.error("PID {:>5} >> Error to move {} into {}.".format(
|
||||||
|
os.getpid(), filenameExt, newFilename))
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|||||||
@@ -19,13 +19,18 @@ from pyftpdlib.handlers import FTPHandler
|
|||||||
from pyftpdlib.servers import FTPServer
|
from pyftpdlib.servers import FTPServer
|
||||||
from pyftpdlib.authorizers import UnixAuthorizer
|
from pyftpdlib.authorizers import UnixAuthorizer
|
||||||
from pyftpdlib.filesystems import UnixFilesystem
|
from pyftpdlib.filesystems import UnixFilesystem
|
||||||
|
|
||||||
|
|
||||||
def send_mail(sev, msg, cfg):
|
def send_mail(sev, msg, cfg):
|
||||||
msg = MIMEText(cfg.message + "\n" + msg)
|
msg = MIMEText(cfg.message + "\n" + msg)
|
||||||
msg["Subject"] = cfg.subject + " " + sev
|
msg["Subject"] = cfg.subject + " " + sev
|
||||||
msg["From"] = cfg.sender
|
msg["From"] = cfg.sender
|
||||||
msg["To"] = cfg.receivers
|
msg["To"] = cfg.receivers
|
||||||
conn = SMTP(host=cfg.smtphost, port=cfg.smtpport, local_hostname=None, timeout=5, source_address=None )
|
conn = SMTP(host=cfg.smtphost,
|
||||||
|
port=cfg.smtpport,
|
||||||
|
local_hostname=None,
|
||||||
|
timeout=5,
|
||||||
|
source_address=None)
|
||||||
conn.set_debuglevel(cfg.debuglevel)
|
conn.set_debuglevel(cfg.debuglevel)
|
||||||
try:
|
try:
|
||||||
conn.login(cfg.sender, cfg.password)
|
conn.login(cfg.sender, cfg.password)
|
||||||
@@ -43,8 +48,8 @@ def send_mail(sev, msg, cfg):
|
|||||||
class mq():
|
class mq():
|
||||||
def __init__(self, cfg):
|
def __init__(self, cfg):
|
||||||
parameters = pika.URLParameters('amqp://' + cfg.mquser + ':' +
|
parameters = pika.URLParameters('amqp://' + cfg.mquser + ':' +
|
||||||
cfg.mqpass + '@' + cfg.mqhost +
|
cfg.mqpass + '@' + cfg.mqhost + ':' +
|
||||||
':' + cfg.mqport +'/%2F')
|
cfg.mqport + '/%2F')
|
||||||
connection = pika.BlockingConnection(parameters)
|
connection = pika.BlockingConnection(parameters)
|
||||||
self.channel = connection.channel()
|
self.channel = connection.channel()
|
||||||
self.channel.queue_declare(queue=cfg.csv_queue, durable=True)
|
self.channel.queue_declare(queue=cfg.csv_queue, durable=True)
|
||||||
@@ -61,123 +66,139 @@ class mq():
|
|||||||
logging.info("PID {:>5} >> write message {} in queue".format(
|
logging.info("PID {:>5} >> write message {} in queue".format(
|
||||||
os.getpid(), msg))
|
os.getpid(), msg))
|
||||||
except:
|
except:
|
||||||
logging.error("PID {:>5} >> error write message {} in queue".format(
|
logging.error(
|
||||||
os.getpid(), msg))
|
"PID {:>5} >> error write message {} in queue".format(
|
||||||
def close(self):
|
os.getpid(), msg))
|
||||||
self.connection.close()
|
|
||||||
|
def close(self):
|
||||||
|
self.channel.close()
|
||||||
|
|
||||||
|
|
||||||
class ASEHandler(FTPHandler):
|
class ASEHandler(FTPHandler):
|
||||||
def on_file_received(self, file):
|
def on_file_received(self, file):
|
||||||
cfg = self.cfg
|
unitType = ''
|
||||||
path, filenameExt = os.path.split(file)
|
unitName = ''
|
||||||
filename, fileExtension = os.path.splitext(filenameExt)
|
toolName = ''
|
||||||
|
toolType = ''
|
||||||
|
fileDate = ''
|
||||||
|
fileTime = ''
|
||||||
|
queue = ''
|
||||||
|
if not os.stat(file).st_size:
|
||||||
|
os.remove(file)
|
||||||
|
logging.info("PID {:>5} >> file {} was empty: removed.".format(
|
||||||
|
os.getpid(), file))
|
||||||
|
else:
|
||||||
|
cfg = self.cfg
|
||||||
|
path, filenameExt = os.path.split(file)
|
||||||
|
filename, fileExtension = os.path.splitext(filenameExt)
|
||||||
|
|
||||||
if (m := re.match(
|
if (m := re.match(
|
||||||
r"^(G\d\d\d)_(ID\d\d\d\d)_(DT\d\d\d\d)_(\d\d)(\d\d)(\d\d\d\d)(\d\d)(\d\d)(\d\d)$",
|
r"^(G\d\d\d)_(ID\d\d\d\d)_(DT\d\d\d\d)_(\d\d)(\d\d)(\d\d\d\d)(\d\d)(\d\d)(\d\d)$",
|
||||||
filename,
|
filename,
|
||||||
re.I,
|
re.I,
|
||||||
)):
|
)):
|
||||||
unitType = m.group(1).upper()
|
unitType = m.group(1).upper()
|
||||||
unitName = m.group(2).upper()
|
unitName = m.group(2).upper()
|
||||||
toolName = m.group(3).upper()
|
toolName = m.group(3).upper()
|
||||||
toolType = "N/A"
|
toolType = "N/A"
|
||||||
fileDate = m.group(6) + "/" + m.group(5) + "/" + m.group(4)
|
fileDate = m.group(6) + "/" + m.group(5) + "/" + m.group(4)
|
||||||
fileTime = m.group(7) + ":" + m.group(8) + ":" + m.group(9)
|
fileTime = m.group(7) + ":" + m.group(8) + ":" + m.group(9)
|
||||||
elif re.match(r"^(\d\d_\d\d\d\d|)(DT\d\d\d\d|LOC\d\d\d\d|GD\d\d\d\d)$",
|
elif re.match(
|
||||||
filename, re.I):
|
r"^(\d\d_\d\d\d\d|)(DT\d\d\d\d|LOC\d\d\d\d|GD\d\d\d\d)$",
|
||||||
with open(file, "r") as fileCsv:
|
filename, re.I):
|
||||||
try:
|
with open(file, "r") as fileCsv:
|
||||||
for i, line in enumerate(fileCsv.readlines(4096), 1):
|
try:
|
||||||
if (m1 := re.match(
|
for i, line in enumerate(fileCsv.readlines(4096), 1):
|
||||||
r"^(File Creation Date:\s)?(\d*\/\d*\/\d*)\s(\d*:\d*:\d*)\;*\n?$",
|
if (m1 := re.match(
|
||||||
line,
|
r"^(File Creation Date:\s)?(\d*\/\d*\/\d*)\s(\d*:\d*:\d*)\;*\n?$",
|
||||||
re.I,
|
line,
|
||||||
|
re.I,
|
||||||
)):
|
)):
|
||||||
fileDate = m1.group(2)
|
fileDate = m1.group(2)
|
||||||
fileTime = m1.group(3)
|
fileTime = m1.group(3)
|
||||||
|
|
||||||
elif (m2 := re.match(
|
elif (m2 := re.match(
|
||||||
r"^(\w+\d+)\s(\w+\d+)\;*\n?$",
|
r"^(\w+\d+)\s(\w+\d+)\;*\n?$",
|
||||||
line,
|
line,
|
||||||
re.I,
|
re.I,
|
||||||
)):
|
)):
|
||||||
unitType = m2.group(1).upper()
|
unitType = m2.group(1).upper()
|
||||||
unitName = m2.group(2).upper()
|
unitName = m2.group(2).upper()
|
||||||
|
|
||||||
elif (m3 := re.match(
|
elif (m3 := re.match(
|
||||||
r"^SD path: a:\/\w+\/(\w+)(?:\.\w+)?\/*(\w*)(?:\.\w+)?\;*\n?$",
|
r"^SD path: a:\/\w+\/(\w+)(?:\.\w+)?\/*(\w*)(?:\.\w+)?\;*\n?$",
|
||||||
line,
|
line, re.I)):
|
||||||
re.I
|
if m3.group(2):
|
||||||
)):
|
toolType = m3.group(1).upper()
|
||||||
if m3.group(2):
|
toolName = m3.group(2).upper()
|
||||||
toolType = m3.group(1).upper()
|
else:
|
||||||
toolName = m3.group(2).upper()
|
toolType = "".join(
|
||||||
else:
|
re.findall("^[a-zA-Z]+",
|
||||||
toolType = "".join(re.findall("^[a-zA-Z]+", m3.group(1))).upper()
|
m3.group(1))).upper()
|
||||||
toolName = m3.group(1).upper()
|
toolName = m3.group(1).upper()
|
||||||
break
|
break
|
||||||
except:
|
except:
|
||||||
logging.error("PID {:>5} >> Error: {}.".format(
|
logging.error("PID {:>5} >> Error: {}.".format(
|
||||||
os.getpid(),
|
os.getpid(),
|
||||||
sys.exc_info()[1]))
|
sys.exc_info()[1]))
|
||||||
fileCsv.close
|
fileCsv.close
|
||||||
|
|
||||||
logging.info("PID {:>5} >> {} - {} - {} - {} - {} {}.".format(
|
|
||||||
os.getpid(),
|
|
||||||
unitType,
|
|
||||||
unitName,
|
|
||||||
toolName,
|
|
||||||
toolType,
|
|
||||||
df.dateFmt(fileDate),
|
|
||||||
fileTime,
|
|
||||||
))
|
|
||||||
newPath = cfg.csvfs + self.username + "/received/" + unitName.upper() + "/"
|
|
||||||
newFilename = (newPath + filename + "_" +
|
|
||||||
str(ts.timestamp("tms") + fileExtension))
|
|
||||||
fileRenamed = (file + "_" + str(ts.timestamp("tms")))
|
|
||||||
os.rename(file, fileRenamed)
|
|
||||||
try:
|
|
||||||
os.makedirs(newPath)
|
|
||||||
logging.info("PID {:>5} >> path {} created.".format(
|
|
||||||
os.getpid(), newPath))
|
|
||||||
except FileExistsError:
|
|
||||||
logging.info("PID {:>5} >> path {} already exists.".format(
|
|
||||||
os.getpid(), newPath))
|
|
||||||
try:
|
|
||||||
shutil.move(fileRenamed, newFilename)
|
|
||||||
logging.info("PID {:>5} >> {} moved into {}.".format(
|
|
||||||
os.getpid(), filenameExt, newFilename))
|
|
||||||
except OSError:
|
|
||||||
logging.error("PID {:>5} >> Error to move {} into {}.".format(
|
|
||||||
os.getpid(), filenameExt, newFilename))
|
|
||||||
send_mail("Error",
|
|
||||||
"OS error move " + filenameExt + " to " + newFilename, cfg)
|
|
||||||
|
|
||||||
|
|
||||||
mq_message = "{};{};{};{};{};{};{}".format(
|
|
||||||
unitType,
|
|
||||||
unitName,
|
|
||||||
toolName,
|
|
||||||
toolType,
|
|
||||||
df.dateFmt(fileDate),
|
|
||||||
fileTime,
|
|
||||||
newFilename,
|
|
||||||
)
|
|
||||||
try:
|
|
||||||
queue = mq(cfg)
|
|
||||||
queue.write(mq_message, cfg)
|
|
||||||
logging.info("PID {:>5} >> queue message: {}.".format(
|
|
||||||
os.getpid(), mq_message))
|
|
||||||
except:
|
|
||||||
logging.error("PID {:>5} >> Error to put message in queue: {}.".format(
|
|
||||||
os.getpid(), mq_message))
|
|
||||||
send_mail("Error",
|
|
||||||
"Error to put message " + mq_message + " in queue.", cfg)
|
|
||||||
finally:
|
|
||||||
queue.close()
|
|
||||||
|
|
||||||
|
logging.info("PID {:>5} >> {} - {} - {} - {} - {} {}.".format(
|
||||||
|
os.getpid(),
|
||||||
|
unitType,
|
||||||
|
unitName,
|
||||||
|
toolName,
|
||||||
|
toolType,
|
||||||
|
df.dateFmt(fileDate),
|
||||||
|
fileTime,
|
||||||
|
))
|
||||||
|
newPath = cfg.csvfs + self.username + "/received/" + unitName.upper(
|
||||||
|
) + "/"
|
||||||
|
newFilename = (newPath + filename + "_" +
|
||||||
|
str(ts.timestamp("tms") + fileExtension))
|
||||||
|
fileRenamed = (file + "_" + str(ts.timestamp("tms")))
|
||||||
|
os.rename(file, fileRenamed)
|
||||||
|
try:
|
||||||
|
os.makedirs(newPath)
|
||||||
|
logging.info("PID {:>5} >> path {} created.".format(
|
||||||
|
os.getpid(), newPath))
|
||||||
|
except FileExistsError:
|
||||||
|
logging.info("PID {:>5} >> path {} already exists.".format(
|
||||||
|
os.getpid(), newPath))
|
||||||
|
try:
|
||||||
|
shutil.move(fileRenamed, newFilename)
|
||||||
|
logging.info("PID {:>5} >> {} moved into {}.".format(
|
||||||
|
os.getpid(), filenameExt, newFilename))
|
||||||
|
except OSError:
|
||||||
|
logging.error("PID {:>5} >> Error to move {} into {}.".format(
|
||||||
|
os.getpid(), filenameExt, newFilename))
|
||||||
|
send_mail(
|
||||||
|
"Error",
|
||||||
|
"OS error move " + filenameExt + " to " + newFilename, cfg)
|
||||||
|
|
||||||
|
mq_message = "{};{};{};{};{};{};{}".format(
|
||||||
|
unitType,
|
||||||
|
unitName,
|
||||||
|
toolName,
|
||||||
|
toolType,
|
||||||
|
df.dateFmt(fileDate),
|
||||||
|
fileTime,
|
||||||
|
newFilename,
|
||||||
|
)
|
||||||
|
try:
|
||||||
|
queue = mq(cfg)
|
||||||
|
queue.write(mq_message, cfg)
|
||||||
|
logging.info("PID {:>5} >> queue message: {}.".format(
|
||||||
|
os.getpid(), mq_message))
|
||||||
|
except:
|
||||||
|
logging.error(
|
||||||
|
"PID {:>5} >> Error to put message in queue: {}.".format(
|
||||||
|
os.getpid(), mq_message))
|
||||||
|
send_mail("Error",
|
||||||
|
"Error to put message " + mq_message + " in queue.",
|
||||||
|
cfg)
|
||||||
|
finally:
|
||||||
|
queue.close()
|
||||||
|
|
||||||
def on_incomplete_file_received(self, file):
|
def on_incomplete_file_received(self, file):
|
||||||
# remove partially uploaded files
|
# remove partially uploaded files
|
||||||
|
|||||||
0
checkDBsync.py
Normal file
0
checkDBsync.py
Normal file
Reference in New Issue
Block a user