add send ftp
This commit is contained in:
@@ -1,3 +1,4 @@
|
||||
CSV_RECEIVED = 0
|
||||
DATA_LOADED = 1
|
||||
DATA_ELABORATED = 2
|
||||
DATA_ELABORATED = 2
|
||||
DATA_SENT = 3
|
||||
60
utils/database/elab_query.py
Normal file
60
utils/database/elab_query.py
Normal file
@@ -0,0 +1,60 @@
|
||||
import csv
|
||||
from io import StringIO
|
||||
import logging
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
async def get_data_as_csv(cfg: dict, id_recv: int, unit: str, tool: str, matlab_timestamp: float, pool: object) -> str:
|
||||
"""
|
||||
Retrieves elaborated data from the database and formats it as a CSV string.
|
||||
|
||||
The query selects data from the `ElabDataView` based on `UnitName`, `ToolNameID`,
|
||||
and a `updated_at` timestamp, then orders it. The first row of the CSV will be
|
||||
the column headers.
|
||||
|
||||
Args:
|
||||
cfg (dict): Configuration dictionary (not directly used in the query but passed for consistency).
|
||||
id (int): The ID of the record being processed (used for logging).
|
||||
pool (object): The database connection pool.
|
||||
unit (str): The name of the unit to filter the data.
|
||||
tool (str): The ID of the tool to filter the data.
|
||||
matlab_timestamp (float): A timestamp used to filter data updated after this time.
|
||||
|
||||
Returns:
|
||||
str: A string containing the elaborated data in CSV format.
|
||||
"""
|
||||
query = """
|
||||
select * from (
|
||||
select 'ToolNameID', 'EventDate', 'EventTime', 'NodeNum', 'NodeType', 'NodeDepth',
|
||||
'XShift', 'YShift', 'ZShift' , 'X', 'Y', 'Z', 'HShift', 'HShiftDir', 'HShift_local',
|
||||
'speed', 'speed_local', 'acceleration', 'acceleration_local', 'T_node', 'water_level', 'pressure', 'load_value', 'AlfaX', 'AlfaY', 'CalcErr'
|
||||
union all
|
||||
select ToolNameID, EventDate, EventTime, NodeNum, NodeType, NodeDepth,
|
||||
XShift, YShift, ZShift , X, Y, Z, HShift, HShiftDir, HShift_local,
|
||||
speed, speed_local, acceleration, acceleration_local, T_node, water_level, pressure, load_value, AlfaX, AlfaY, calcerr
|
||||
from ElabDataView
|
||||
where UnitName = %s and ToolNameID = %s and updated_at > %s
|
||||
order by ToolNameID DESC, concat(EventDate, EventTime), convert(`NodeNum`, UNSIGNED INTEGER) DESC
|
||||
) resulting_set
|
||||
"""
|
||||
async with pool.acquire() as conn:
|
||||
async with conn.cursor() as cur:
|
||||
try:
|
||||
await cur.execute(query, (unit, tool, matlab_timestamp))
|
||||
results = await cur.fetchall()
|
||||
logger.info(f"id {id_recv} - {unit} - {tool}: estratti i dati per invio CSV")
|
||||
logger.info(f"Numero di righe estratte: {len(results)}")
|
||||
|
||||
# Creare CSV in memoria
|
||||
output = StringIO()
|
||||
writer = csv.writer(output, delimiter=";", lineterminator="\n", quoting=csv.QUOTE_MINIMAL)
|
||||
for row in results:
|
||||
writer.writerow(row)
|
||||
csv_data = output.getvalue()
|
||||
output.close()
|
||||
|
||||
return csv_data
|
||||
|
||||
except Exception as e:
|
||||
logging.error(f"id {id_recv} - {unit} - {tool} - errore nel query creazione csv: {e}")
|
||||
return None
|
||||
@@ -4,7 +4,7 @@ import asyncio
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
timestamp_cols = ["inserted_at", "loaded_at", "elaborated_at"]
|
||||
timestamp_cols = ["inserted_at", "loaded_at", "elaborated_at", "sent_at"]
|
||||
|
||||
|
||||
async def load_data(cfg: object, matrice_valori: list, pool: object) -> bool:
|
||||
@@ -109,7 +109,7 @@ async def update_status(cfg: object, id: int, status: int, pool: object) -> None
|
||||
f"update {cfg.dbrectable} set status = {status}, {timestamp_cols[status]} = now() where id = {id}"
|
||||
)
|
||||
await conn.commit()
|
||||
logging.info("Status updated.")
|
||||
logging.info(f"Status updated id {id}.")
|
||||
except Exception as e:
|
||||
await conn.rollback()
|
||||
logging.error(f"Error: {e}")
|
||||
|
||||
Reference in New Issue
Block a user