You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
270 lines
7.8 KiB
Python
270 lines
7.8 KiB
Python
import os
|
|
import sys
|
|
|
|
|
|
if sys.version_info[0] > 2 :
|
|
from transport.common import Reader, Writer #, factory
|
|
else:
|
|
from common import Reader,Writer
|
|
# import nujson as json
|
|
import json
|
|
# from threading import Lock
|
|
import sqlite3
|
|
import pandas as pd
|
|
from multiprocessing import Lock
|
|
from transport.common import Reader, Writer, IEncoder
|
|
import sqlalchemy
|
|
from sqlalchemy import create_engine
|
|
class DiskReader(Reader) :
|
|
"""
|
|
This class is designed to read data from disk (location on hard drive)
|
|
@pre : isready() == True
|
|
"""
|
|
|
|
def __init__(self,**params):
|
|
"""
|
|
|
|
@param path absolute path of the file to be read
|
|
"""
|
|
|
|
Reader.__init__(self)
|
|
self.path = params['path'] if 'path' in params else None
|
|
self.delimiter = params['delimiter'] if 'delimiter' in params else ','
|
|
|
|
def isready(self):
|
|
return os.path.exists(self.path)
|
|
def meta(self,**_args):
|
|
return []
|
|
def read(self,**args):
|
|
_path = self.path if 'path' not in args else args['path']
|
|
_delimiter = self.delimiter if 'delimiter' not in args else args['delimiter']
|
|
return pd.read_csv(_path,delimiter=self.delimiter)
|
|
def stream(self,**args):
|
|
"""
|
|
This function reads the rows from a designated location on disk
|
|
@param size number of rows to be read, -1 suggests all rows
|
|
"""
|
|
|
|
size = -1 if 'size' not in args else int(args['size'])
|
|
f = open(self.path,'rU')
|
|
i = 1
|
|
for row in f:
|
|
|
|
i += 1
|
|
if size == i:
|
|
break
|
|
if self.delimiter :
|
|
yield row.split(self.delimiter)
|
|
yield row
|
|
f.close()
|
|
class DiskWriter(Writer):
|
|
|
|
"""
|
|
This function writes output to disk in a designated location. The function will write a text to a text file
|
|
- If a delimiter is provided it will use that to generate a xchar-delimited file
|
|
- If not then the object will be dumped as is
|
|
"""
|
|
THREAD_LOCK = Lock()
|
|
def __init__(self,**params):
|
|
super().__init__()
|
|
self._path = params['path']
|
|
self._delimiter = params['delimiter'] if 'delimiter' in params else None
|
|
self._mode = 'w' if 'mode' not in params else params['mode']
|
|
# def meta(self):
|
|
# return self.cache['meta']
|
|
# def isready(self):
|
|
# """
|
|
# This function determines if the class is ready for execution or not
|
|
# i.e it determines if the preconditions of met prior execution
|
|
# """
|
|
# return True
|
|
# # p = self.path is not None and os.path.exists(self.path)
|
|
# # q = self.name is not None
|
|
# # return p and q
|
|
# def format (self,row):
|
|
# self.cache['meta']['cols'] += len(row) if isinstance(row,list) else len(row.keys())
|
|
# self.cache['meta']['rows'] += 1
|
|
# return (self.delimiter.join(row) if self.delimiter else json.dumps(row))+"\n"
|
|
def write(self,info,**_args):
|
|
"""
|
|
This function writes a record to a designated file
|
|
@param label <passed|broken|fixed|stats>
|
|
@param row row to be written
|
|
"""
|
|
try:
|
|
|
|
|
|
DiskWriter.THREAD_LOCK.acquire()
|
|
|
|
_delim = self._delimiter if 'delimiter' not in _args else _args['delimiter']
|
|
_path = self._path if 'path' not in _args else _args['path']
|
|
_mode = self._mode if 'mode' not in _args else _args['mode']
|
|
info.to_csv(_path,index=False,sep=_delim)
|
|
pass
|
|
except Exception as e:
|
|
#
|
|
# Not sure what should be done here ...
|
|
pass
|
|
finally:
|
|
DiskWriter.THREAD_LOCK.release()
|
|
class SQLite :
|
|
def __init__(self,**_args) :
|
|
self.path = _args['database'] if 'database' in _args else _args['path']
|
|
self.conn = sqlite3.connect(self.path,isolation_level="IMMEDIATE")
|
|
self.conn.row_factory = sqlite3.Row
|
|
self.fields = _args['fields'] if 'fields' in _args else []
|
|
def has (self,**_args):
|
|
found = False
|
|
try:
|
|
if 'table' in _args :
|
|
table = _args['table']
|
|
sql = "SELECT * FROM :table limit 1".replace(":table",table)
|
|
_df = pd.read_sql(sql,self.conn)
|
|
found = _df.columns.size > 0
|
|
except Exception as e:
|
|
pass
|
|
return found
|
|
def close(self):
|
|
try:
|
|
self.conn.close()
|
|
except Exception as e :
|
|
print(e)
|
|
def apply(self,sql):
|
|
try:
|
|
if not sql.lower().startswith('select'):
|
|
cursor = self.conn.cursor()
|
|
cursor.execute(sql)
|
|
cursor.close()
|
|
self.conn.commit()
|
|
else:
|
|
return pd.read_sql(sql,self.conn)
|
|
except Exception as e:
|
|
print (e)
|
|
class SQLiteReader (SQLite,DiskReader):
|
|
def __init__(self,**args):
|
|
super().__init__(**args)
|
|
# DiskReader.__init__(self,**args)
|
|
# self.path = args['database'] if 'database' in args else args['path']
|
|
# self.conn = sqlite3.connect(self.path,isolation_level=None)
|
|
# self.conn.row_factory = sqlite3.Row
|
|
self.table = args['table'] if 'table' in args else None
|
|
def read(self,**args):
|
|
if 'sql' in args :
|
|
sql = args['sql']
|
|
elif 'filter' in args :
|
|
sql = "SELECT :fields FROM ",self.table, "WHERE (:filter)".replace(":filter",args['filter'])
|
|
sql = sql.replace(":fields",args['fields']) if 'fields' in args else sql.replace(":fields","*")
|
|
else:
|
|
sql = ' '.join(['SELECT * FROM ',self.table])
|
|
if 'limit' in args :
|
|
sql = sql + " LIMIT "+args['limit']
|
|
return pd.read_sql(sql,self.conn)
|
|
def close(self):
|
|
try:
|
|
self.conn.close()
|
|
except Exception as e :
|
|
pass
|
|
|
|
class SQLiteWriter(SQLite,DiskWriter) :
|
|
connection = None
|
|
LOCK = Lock()
|
|
def __init__(self,**args):
|
|
"""
|
|
:path
|
|
:fields json|csv
|
|
"""
|
|
# DiskWriter.__init__(self,**args)
|
|
super().__init__(**args)
|
|
self.table = args['table'] if 'table' in args else None
|
|
path = self.path
|
|
self._engine = create_engine(f'sqlite:///{path}')
|
|
|
|
# self.conn = sqlite3.connect(self.path,isolation_level="IMMEDIATE")
|
|
# self.conn.row_factory = sqlite3.Row
|
|
# self.fields = args['fields'] if 'fields' in args else []
|
|
|
|
if self.fields and not self.isready() and self.table:
|
|
self.init(self.fields)
|
|
SQLiteWriter.connection = self.conn
|
|
def init(self,fields):
|
|
self.fields = fields;
|
|
sql = " ".join(["CREATE TABLE IF NOT EXISTS ",self.table," (", ",".join(self.fields),")"])
|
|
|
|
cursor = self.conn.cursor()
|
|
cursor.execute(sql)
|
|
cursor.close()
|
|
self.conn.commit()
|
|
def isready(self):
|
|
try:
|
|
sql = "SELECT count(*) FROM sqlite_master where name=':table'"
|
|
sql = sql.replace(":table",self.table)
|
|
cursor = self.conn.cursor()
|
|
|
|
r = cursor.execute(sql)
|
|
r = r.fetchall()
|
|
cursor.close()
|
|
|
|
return r[0][0] != 0
|
|
except Exception as e:
|
|
pass
|
|
return 0
|
|
#
|
|
# If the table doesn't exist we should create it
|
|
#
|
|
# def write(self,_data,**_args):
|
|
# SQLiteWriter.LOCK.acquire()
|
|
# try:
|
|
# if type(_data) == dict :
|
|
# _data = [_data]
|
|
# _table = self.table if 'table' not in _args else _args['table']
|
|
# _df = pd.DataFrame(_data)
|
|
# _df.to_sql(_table,self._engine.connect(),if_exists='append',index=False)
|
|
# except Exception as e:
|
|
# print (e)
|
|
# SQLiteWriter.LOCK.release()
|
|
def write(self,info,**_args):
|
|
"""
|
|
"""
|
|
|
|
#if not self.fields :
|
|
# #if type(info) == pd.DataFrame :
|
|
# # _columns = list(info.columns)
|
|
# #self.init(list(info.keys()))
|
|
|
|
if type(info) == dict :
|
|
info = [info]
|
|
elif type(info) == pd.DataFrame :
|
|
info = info.fillna('')
|
|
info = info.to_dict(orient='records')
|
|
|
|
if not self.fields :
|
|
_rec = info[0]
|
|
self.init(list(_rec.keys()))
|
|
|
|
SQLiteWriter.LOCK.acquire()
|
|
try:
|
|
|
|
cursor = self.conn.cursor()
|
|
sql = " " .join(["INSERT INTO ",self.table,"(", ",".join(self.fields) ,")", "values(:values)"])
|
|
for row in info :
|
|
values = [ str(row[field]) if type(row[field]) not in [list,dict] else json.dumps(row[field],cls=IEncoder) for field in self.fields]
|
|
values = ["".join(["'",value,"'"]) for value in values]
|
|
|
|
# stream =["".join(["",value,""]) if type(value) == str else value for value in row.values()]
|
|
# stream = json.dumps(stream,cls=IEncoder)
|
|
# stream = stream.replace("[","").replace("]","")
|
|
|
|
# print (sql.replace(":values",stream))
|
|
# self.conn.execute(sql.replace(":values",stream) )
|
|
self.conn.execute(sql.replace(":values", ",".join(values)) )
|
|
# cursor.commit()
|
|
|
|
self.conn.commit()
|
|
# print (sql)
|
|
except Exception as e :
|
|
print ()
|
|
|
|
print (e)
|
|
pass
|
|
SQLiteWriter.LOCK.release()
|