reorg ini e log
This commit is contained in:
@@ -9,7 +9,7 @@ import asyncio
|
|||||||
import subprocess
|
import subprocess
|
||||||
|
|
||||||
# Import custom modules for configuration and database connection
|
# Import custom modules for configuration and database connection
|
||||||
from utils.config import loader as setting
|
from utils.config import loader_ftp_csv as setting
|
||||||
from utils.database.connection import connetti_db
|
from utils.database.connection import connetti_db
|
||||||
from utils.database.loader_action import DATA_LOADED, get_matlab_cmd
|
from utils.database.loader_action import DATA_LOADED, get_matlab_cmd
|
||||||
|
|
||||||
|
|||||||
15
env/db.ini
vendored
Normal file
15
env/db.ini
vendored
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
# to generete adminuser password hash:
|
||||||
|
# python3 -c 'from hashlib import sha256;print(sha256("????password???".encode("UTF-8")).hexdigest())'
|
||||||
|
|
||||||
|
[db]
|
||||||
|
hostname = 10.211.114.173
|
||||||
|
port = 3306
|
||||||
|
user = root
|
||||||
|
password = batt1l0
|
||||||
|
dbName = ase_lar
|
||||||
|
|
||||||
|
[tables]
|
||||||
|
userTableName = virtusers
|
||||||
|
recTableName = received
|
||||||
|
rawTableName = RAWDATACOR
|
||||||
|
nodesTableName = nodes
|
||||||
0
env/elab.ini
vendored
Normal file
0
env/elab.ini
vendored
Normal file
17
ftp_csv_receiver.ini → env/ftp.ini
vendored
17
ftp_csv_receiver.ini → env/ftp.ini
vendored
@@ -3,7 +3,6 @@
|
|||||||
|
|
||||||
[ftpserver]
|
[ftpserver]
|
||||||
firstPort = 40000
|
firstPort = 40000
|
||||||
logFilename = ./ftppylog.log
|
|
||||||
proxyAddr = 0.0.0.0
|
proxyAddr = 0.0.0.0
|
||||||
portRangeWidth = 500
|
portRangeWidth = 500
|
||||||
virtpath = /home/alex/aseftp/
|
virtpath = /home/alex/aseftp/
|
||||||
@@ -17,20 +16,8 @@
|
|||||||
[csvfs]
|
[csvfs]
|
||||||
path = /home/alex/aseftp/csvfs/
|
path = /home/alex/aseftp/csvfs/
|
||||||
|
|
||||||
[csvelab]
|
[logging]
|
||||||
logFilename = csvElab.log
|
logFilename = ./ftp_csv_rec.log
|
||||||
max_threads = 10
|
|
||||||
|
|
||||||
[db]
|
|
||||||
hostname = 10.211.114.173
|
|
||||||
port = 3306
|
|
||||||
user = root
|
|
||||||
password = batt1l0
|
|
||||||
dbName = ase_lar
|
|
||||||
userTableName = virtusers
|
|
||||||
recTableName = received
|
|
||||||
rawTableName = RAWDATACOR
|
|
||||||
nodesTableName = nodes
|
|
||||||
|
|
||||||
[unit]
|
[unit]
|
||||||
Types = G801|G201|G301|G802|D2W|GFLOW|CR1000X|TLP|GS1|HORTUS
|
Types = G801|G201|G301|G802|D2W|GFLOW|CR1000X|TLP|GS1|HORTUS
|
||||||
5
env/load.ini
vendored
Normal file
5
env/load.ini
vendored
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
[logging]
|
||||||
|
logFilename = ./load_raw_data.log
|
||||||
|
|
||||||
|
[threads]
|
||||||
|
max_num = 10
|
||||||
@@ -6,7 +6,7 @@ import logging
|
|||||||
from hashlib import sha256
|
from hashlib import sha256
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
from utils.config import loader
|
from utils.config import loader_ftp_csv
|
||||||
from utils.database.connection import connetti_db
|
from utils.database.connection import connetti_db
|
||||||
from utils.ftp import user_admin, file_management
|
from utils.ftp import user_admin, file_management
|
||||||
|
|
||||||
@@ -114,7 +114,7 @@ class ASEHandler(FTPHandler):
|
|||||||
def main():
|
def main():
|
||||||
"""Main function to start the FTP server."""
|
"""Main function to start the FTP server."""
|
||||||
# Load the configuration settings
|
# Load the configuration settings
|
||||||
cfg = loader.Config()
|
cfg = loader_ftp_csv.Config()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Initialize the authorizer and handler
|
# Initialize the authorizer and handler
|
||||||
|
|||||||
@@ -5,21 +5,21 @@ import mysql.connector
|
|||||||
import logging
|
import logging
|
||||||
import importlib
|
import importlib
|
||||||
import asyncio
|
import asyncio
|
||||||
|
import os
|
||||||
|
|
||||||
# Import custom modules for configuration and database connection
|
# Import custom modules for configuration and database connection
|
||||||
from utils.config import loader as setting
|
from utils.config import loader_load_data as setting
|
||||||
from utils.database.connection import connetti_db
|
from utils.database.connection import connetti_db
|
||||||
from utils.database.loader_action import CSV_RECEIVED
|
from utils.database.loader_action import CSV_RECEIVED
|
||||||
|
|
||||||
# Initialize the logger for this module
|
# Initialize the logger for this module
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger()
|
||||||
|
|
||||||
# Delay tra un processamento CSV e il successivo (in secondi)
|
# Delay tra un processamento CSV e il successivo (in secondi)
|
||||||
CSV_PROCESSING_DELAY = 0.1
|
CSV_PROCESSING_DELAY = 0.1
|
||||||
# Tempo di attesa se non ci sono record da elaborare
|
# Tempo di attesa se non ci sono record da elaborare
|
||||||
NO_RECORD_SLEEP = 20
|
NO_RECORD_SLEEP = 20
|
||||||
|
|
||||||
|
|
||||||
async def worker(worker_id: int, queue: asyncio.Queue, cfg: object) -> None:
|
async def worker(worker_id: int, queue: asyncio.Queue, cfg: object) -> None:
|
||||||
"""
|
"""
|
||||||
Worker asyncrono che preleva lavori dalla coda e li esegue.
|
Worker asyncrono che preleva lavori dalla coda e li esegue.
|
||||||
@@ -29,6 +29,7 @@ async def worker(worker_id: int, queue: asyncio.Queue, cfg: object) -> None:
|
|||||||
queue (asyncio.Queue): Coda da cui prendere i lavori.
|
queue (asyncio.Queue): Coda da cui prendere i lavori.
|
||||||
cfg (object): Configurazione caricata.
|
cfg (object): Configurazione caricata.
|
||||||
"""
|
"""
|
||||||
|
debug_mode = (logging.getLogger().getEffectiveLevel() == logging.DEBUG)
|
||||||
logger.info(f"Worker {worker_id} - Avviato")
|
logger.info(f"Worker {worker_id} - Avviato")
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
@@ -48,12 +49,12 @@ async def worker(worker_id: int, queue: asyncio.Queue, cfg: object) -> None:
|
|||||||
await asyncio.sleep(CSV_PROCESSING_DELAY)
|
await asyncio.sleep(CSV_PROCESSING_DELAY)
|
||||||
else:
|
else:
|
||||||
logger.debug(f"Worker {worker_id} - Elaborazione completata correttamente")
|
logger.debug(f"Worker {worker_id} - Elaborazione completata correttamente")
|
||||||
await asyncio.sleep(CSV_PROCESSING_DELAY)
|
await asyncio.sleep(CSV_PROCESSING_DELAY*worker_id)
|
||||||
|
|
||||||
# Segnala che il lavoro è completato
|
# Segnala che il lavoro è completato
|
||||||
queue.task_done()
|
queue.task_done()
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Worker {worker_id} - Errore durante l'esecuzione: {e}", exc_info=True)
|
logger.error(f"Worker {worker_id} - Errore durante l'esecuzione: {e}", exc_info=debug_mode)
|
||||||
queue.task_done()
|
queue.task_done()
|
||||||
|
|
||||||
|
|
||||||
@@ -67,6 +68,8 @@ async def load_csv(cfg: object) -> tuple:
|
|||||||
Returns:
|
Returns:
|
||||||
bool: True se è stato trovato ed elaborato un record, False altrimenti.
|
bool: True se è stato trovato ed elaborato un record, False altrimenti.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
debug_mode = (logging.getLogger().getEffectiveLevel() == logging.DEBUG)
|
||||||
logger.debug("Inizio ricerca nuovo CSV da elaborare")
|
logger.debug("Inizio ricerca nuovo CSV da elaborare")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
@@ -107,14 +110,14 @@ async def load_csv(cfg: object) -> tuple:
|
|||||||
logger.info(f"Elaborazione completata per ID={id}")
|
logger.info(f"Elaborazione completata per ID={id}")
|
||||||
return True, True
|
return True, True
|
||||||
except (ImportError, AttributeError) as e:
|
except (ImportError, AttributeError) as e:
|
||||||
logger.error(f"Errore nel caricamento del modulo o della funzione: {e}", exc_info=True)
|
logger.error(f"Errore nel caricamento del modulo o della funzione: {e}", exc_info=debug_mode)
|
||||||
return True, False
|
return True, False
|
||||||
else:
|
else:
|
||||||
logger.debug("Nessun record disponibile per l'elaborazione")
|
logger.debug("Nessun record disponibile per l'elaborazione")
|
||||||
return False, False
|
return False, False
|
||||||
|
|
||||||
except mysql.connector.Error as e:
|
except mysql.connector.Error as e:
|
||||||
logger.error(f"Errore di database: {e}", exc_info=True)
|
logger.error(f"Errore di database: {e}", exc_info=debug_mode)
|
||||||
return False, False
|
return False, False
|
||||||
|
|
||||||
|
|
||||||
@@ -127,15 +130,18 @@ async def main():
|
|||||||
|
|
||||||
try:
|
try:
|
||||||
# Configura il logging globale
|
# Configura il logging globale
|
||||||
|
log_level = os.getenv("LOG_LEVEL", "INFO").upper()
|
||||||
|
debug_mode = (logging.getLogger().getEffectiveLevel() == logging.DEBUG)
|
||||||
|
|
||||||
logging.basicConfig(
|
logging.basicConfig(
|
||||||
format="%(asctime)s - PID: %(process)d.%(name)s.%(levelname)s: %(message)s ",
|
format="%(asctime)s - PID: %(process)d.%(name)s.%(levelname)s: %(message)s ",
|
||||||
filename=cfg.logfilename,
|
filename=cfg.logfilename,
|
||||||
level=logging.INFO,
|
level=log_level,
|
||||||
)
|
)
|
||||||
logger.info("Logging configurato correttamente")
|
logger.info("Logging configurato correttamente")
|
||||||
|
|
||||||
# Crea una coda di lavoro illimitata
|
# Crea una coda di lavoro illimitata
|
||||||
queue = asyncio.Queue(maxsize=cfg.queue_maxsize or 10)
|
queue = asyncio.Queue(maxsize=cfg.max_threads * 2 or 20)
|
||||||
logger.debug("Coda di lavoro creata")
|
logger.debug("Coda di lavoro creata")
|
||||||
|
|
||||||
# Numero massimo di worker concorrenti
|
# Numero massimo di worker concorrenti
|
||||||
@@ -169,12 +175,12 @@ async def main():
|
|||||||
for task in workers:
|
for task in workers:
|
||||||
task.cancel()
|
task.cancel()
|
||||||
|
|
||||||
await asyncio.gather(*workers, return_exceptions=True)
|
await asyncio.gather(*workers, return_exceptions=debug_mode)
|
||||||
|
|
||||||
logger.info("Info: Tutti i task terminati. Uscita.")
|
logger.info("Info: Tutti i task terminati. Uscita.")
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Errore principale: {e}", exc_info=True)
|
logger.error(f"Errore principale: {e}", exc_info=debug_mode)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
|||||||
@@ -5,11 +5,12 @@ from configparser import ConfigParser
|
|||||||
|
|
||||||
class Config:
|
class Config:
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
|
|
||||||
c = ConfigParser()
|
c = ConfigParser()
|
||||||
c.read(["/etc/aseftp/ftp_csv_receiver.ini", "./ftp_csv_receiver.ini"])
|
c.read(["env/ftp.ini", "env/db.ini"])
|
||||||
|
|
||||||
# FTP setting
|
# FTP setting
|
||||||
self.firstport = c.getint("ftpserver", "firstPort")
|
self.firstport = c.getint("ftpserver", "firstPort")
|
||||||
self.logfilename = c.get("ftpserver", "logFilename")
|
|
||||||
self.proxyaddr = c.get("ftpserver", "proxyAddr")
|
self.proxyaddr = c.get("ftpserver", "proxyAddr")
|
||||||
self.portrangewidth = c.getint("ftpserver", "portRangeWidth")
|
self.portrangewidth = c.getint("ftpserver", "portRangeWidth")
|
||||||
self.virtpath = c.get("ftpserver", "virtpath")
|
self.virtpath = c.get("ftpserver", "virtpath")
|
||||||
@@ -22,9 +23,8 @@ class Config:
|
|||||||
# CSV FILE setting
|
# CSV FILE setting
|
||||||
self.csvfs = c.get("csvfs", "path")
|
self.csvfs = c.get("csvfs", "path")
|
||||||
|
|
||||||
# LOADER setting
|
# LOG setting
|
||||||
self.elablog = c.get("csvelab", "logFilename")
|
self.logfilename = c.get("logging", "logFilename")
|
||||||
self.max_threads = c.getint("csvelab", "max_threads")
|
|
||||||
|
|
||||||
# DB setting
|
# DB setting
|
||||||
self.dbhost = c.get("db", "hostname")
|
self.dbhost = c.get("db", "hostname")
|
||||||
@@ -32,12 +32,13 @@ class Config:
|
|||||||
self.dbuser = c.get("db", "user")
|
self.dbuser = c.get("db", "user")
|
||||||
self.dbpass = c.get("db", "password")
|
self.dbpass = c.get("db", "password")
|
||||||
self.dbname = c.get("db", "dbName")
|
self.dbname = c.get("db", "dbName")
|
||||||
#self.dbschema = c.get("db", "dbSchema")
|
|
||||||
self.dbusertable = c.get("db", "userTableName")
|
# Tables
|
||||||
self.dbrectable = c.get("db", "recTableName")
|
self.dbusertable = c.get("tables", "userTableName")
|
||||||
self.dbrawdata = c.get("db", "rawTableName")
|
self.dbrectable = c.get("tables", "recTableName")
|
||||||
self.dbrawdata = c.get("db", "rawTableName")
|
self.dbrawdata = c.get("tables", "rawTableName")
|
||||||
self.dbnodes = c.get("db", "nodesTableName")
|
self.dbrawdata = c.get("tables", "rawTableName")
|
||||||
|
self.dbnodes = c.get("tables", "nodesTableName")
|
||||||
|
|
||||||
# unit setting
|
# unit setting
|
||||||
self.units_name = [part for part in c.get("unit", "Names").split('|')]
|
self.units_name = [part for part in c.get("unit", "Names").split('|')]
|
||||||
31
utils/config/loader_load_data.py
Normal file
31
utils/config/loader_load_data.py
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
"""set configurations
|
||||||
|
|
||||||
|
"""
|
||||||
|
from configparser import ConfigParser
|
||||||
|
|
||||||
|
class Config:
|
||||||
|
def __init__(self):
|
||||||
|
|
||||||
|
c = ConfigParser()
|
||||||
|
c.read(["env/load.ini", "env/db.ini"])
|
||||||
|
|
||||||
|
# LOG setting
|
||||||
|
self.logfilename = c.get("logging", "logFilename")
|
||||||
|
|
||||||
|
# Worker setting
|
||||||
|
self.max_threads = c.getint("threads", "max_num")
|
||||||
|
|
||||||
|
# DB setting
|
||||||
|
self.dbhost = c.get("db", "hostname")
|
||||||
|
self.dbport = c.getint("db", "port")
|
||||||
|
self.dbuser = c.get("db", "user")
|
||||||
|
self.dbpass = c.get("db", "password")
|
||||||
|
self.dbname = c.get("db", "dbName")
|
||||||
|
|
||||||
|
# Tables
|
||||||
|
self.dbusertable = c.get("tables", "userTableName")
|
||||||
|
self.dbrectable = c.get("tables", "recTableName")
|
||||||
|
self.dbrawdata = c.get("tables", "rawTableName")
|
||||||
|
self.dbrawdata = c.get("tables", "rawTableName")
|
||||||
|
self.dbnodes = c.get("tables", "nodesTableName")
|
||||||
|
|
||||||
30
utils/config/loader_matlab_elab.py
Normal file
30
utils/config/loader_matlab_elab.py
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
"""set configurations
|
||||||
|
|
||||||
|
"""
|
||||||
|
from configparser import ConfigParser
|
||||||
|
|
||||||
|
class Config:
|
||||||
|
def __init__(self):
|
||||||
|
|
||||||
|
c = ConfigParser()
|
||||||
|
c.read(["env/elab.ini", "env/db.ini"])
|
||||||
|
|
||||||
|
# LOG setting
|
||||||
|
self.logfilename = c.get("logging", "logFilename")
|
||||||
|
|
||||||
|
# Worker setting
|
||||||
|
self.max_threads = c.getint("threads", "max_num")
|
||||||
|
|
||||||
|
# DB setting
|
||||||
|
self.dbhost = c.get("db", "hostname")
|
||||||
|
self.dbport = c.getint("db", "port")
|
||||||
|
self.dbuser = c.get("db", "user")
|
||||||
|
self.dbpass = c.get("db", "password")
|
||||||
|
self.dbname = c.get("db", "dbName")
|
||||||
|
|
||||||
|
# Tables
|
||||||
|
self.dbusertable = c.get("tables", "userTableName")
|
||||||
|
self.dbrectable = c.get("tables", "recTableName")
|
||||||
|
self.dbrawdata = c.get("tables", "rawTableName")
|
||||||
|
self.dbrawdata = c.get("tables", "rawTableName")
|
||||||
|
self.dbnodes = c.get("tables", "nodesTableName")
|
||||||
1
utils/csv/__init__.py
Normal file
1
utils/csv/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""Parser delle centraline"""
|
||||||
@@ -5,7 +5,7 @@ import mysql.connector
|
|||||||
|
|
||||||
from utils.database.connection import connetti_db
|
from utils.database.connection import connetti_db
|
||||||
|
|
||||||
from utils.config.parser import extract_value
|
from utils.csv.parser import extract_value
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
from .g801_mums import main_loader as g801_mums_main_loader
|
from .g801_mums import main_loader as g801_mums_main_loader
|
||||||
|
|
||||||
def main_loader(cfg: object, id: int) -> None:
|
async def main_loader(cfg: object, id: int) -> None:
|
||||||
return g801_mums_main_loader(cfg, id)
|
return g801_mums_main_loader(cfg, id)
|
||||||
@@ -1,11 +1,11 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
from utils.database.loader_action import load_data
|
from utils.database.loader_action import load_data
|
||||||
from utils.parsers.data_preparation import make_loc_matrix
|
from utils.csv.data_preparation import make_loc_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
def main_loader(cfg, id):
|
async def main_loader(cfg, id):
|
||||||
matrice_valori = make_loc_matrix(cfg, id)
|
matrice_valori = make_loc_matrix(cfg, id)
|
||||||
load_data(cfg, matrice_valori)
|
load_data(cfg, matrice_valori)
|
||||||
|
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
from .g801_mums import main_loader as g801_mums_main_loader
|
from .g801_mums import main_loader as g801_mums_main_loader
|
||||||
|
|
||||||
def main_loader(cfg: object, id: int) -> None:
|
async def main_loader(cfg: object, id: int) -> None:
|
||||||
return g801_mums_main_loader(cfg, id)
|
return g801_mums_main_loader(cfg, id)
|
||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
from .g801_loc import main_loader as g801_loc_main_loader
|
from .g801_loc import main_loader as g801_loc_main_loader
|
||||||
|
|
||||||
def main_loader(cfg: object, id: int) -> None:
|
async def main_loader(cfg: object, id: int) -> None:
|
||||||
return g801_loc_main_loader(cfg, id)
|
return g801_loc_main_loader(cfg, id)
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
from .g801_mums import main_loader as g801_mums_main_loader
|
from .g801_mums import main_loader as g801_mums_main_loader
|
||||||
|
|
||||||
def main_loader(cfg: object, id: int) -> None:
|
async def main_loader(cfg: object, id: int) -> None:
|
||||||
return g801_mums_main_loader(cfg, id)
|
return g801_mums_main_loader(cfg, id)
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
from .g801_mums import main_loader as g801_mums_main_loader
|
from .g801_mums import main_loader as g801_mums_main_loader
|
||||||
|
|
||||||
def main_loader(cfg: object, id: int) -> None:
|
async def main_loader(cfg: object, id: int) -> None:
|
||||||
return g801_mums_main_loader(cfg, id)
|
return g801_mums_main_loader(cfg, id)
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
from .g801_mux import main_loader as g801_mux_main_loader
|
from .g801_mux import main_loader as g801_mux_main_loader
|
||||||
|
|
||||||
def main_loader(cfg: object, id: int) -> None:
|
async def main_loader(cfg: object, id: int) -> None:
|
||||||
return g801_mux_main_loader(cfg, id)
|
return g801_mux_main_loader(cfg, id)
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
from .tlp_tlp import main_loader as tlp_tlp_main_loader
|
from .tlp_tlp import main_loader as tlp_tlp_main_loader
|
||||||
|
|
||||||
def main_loader(cfg: object, id: int) -> None:
|
async def main_loader(cfg: object, id: int) -> None:
|
||||||
return tlp_tlp_main_loader(cfg, id)
|
return tlp_tlp_main_loader(cfg, id)
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
from .cr1000x_cr1000x import main_loader as cr1000x_cr1000x_main_loader
|
from .cr1000x_cr1000x import main_loader as cr1000x_cr1000x_main_loader
|
||||||
|
|
||||||
def main_loader(cfg: object, id: int) -> None:
|
async def main_loader(cfg: object, id: int) -> None:
|
||||||
return cr1000x_cr1000x_main_loader(cfg, id)
|
return cr1000x_cr1000x_main_loader(cfg, id)
|
||||||
@@ -1,7 +1,7 @@
|
|||||||
#!.venv/bin/python
|
#!.venv/bin/python
|
||||||
# Import necessary modules
|
# Import necessary modules
|
||||||
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
from utils.database.loader_action import load_data, update_status, DATA_LOADED
|
||||||
from utils.parsers.data_preparation import make_matrix
|
from utils.csv.data_preparation import make_matrix
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
Reference in New Issue
Block a user