mirror of
https://github.com/PiBrewing/craftbeerpi4.git
synced 2024-11-10 01:17:42 +01:00
Merge pull request #110 from prash3r/hookable-log-data-revisited
Hookable log data revisited
This commit is contained in:
commit
e01850f2dc
11 changed files with 182 additions and 73 deletions
1
.gitignore
vendored
1
.gitignore
vendored
|
@ -20,3 +20,4 @@ logs/
|
||||||
.devcontainer/cbpi-dev-config/*
|
.devcontainer/cbpi-dev-config/*
|
||||||
cbpi4-*
|
cbpi4-*
|
||||||
temp*
|
temp*
|
||||||
|
*.patch
|
6
.vscode/launch.json
vendored
6
.vscode/launch.json
vendored
|
@ -10,7 +10,11 @@
|
||||||
"type": "python",
|
"type": "python",
|
||||||
"request": "launch",
|
"request": "launch",
|
||||||
"module": "run",
|
"module": "run",
|
||||||
"args": ["--config-folder-path=./.devcontainer/cbpi-dev-config", "start"],
|
"args": [
|
||||||
|
"--config-folder-path=./.devcontainer/cbpi-dev-config",
|
||||||
|
"--debug-log-level=20",
|
||||||
|
"start"
|
||||||
|
],
|
||||||
"preLaunchTask": "copy default cbpi config files if dev config files dont exist"
|
"preLaunchTask": "copy default cbpi config files if dev config files dont exist"
|
||||||
},
|
},
|
||||||
|
|
||||||
|
|
|
@ -13,6 +13,7 @@ from cbpi.api import *
|
||||||
from cbpi.api.config import ConfigType
|
from cbpi.api.config import ConfigType
|
||||||
from cbpi.api.base import CBPiBase
|
from cbpi.api.base import CBPiBase
|
||||||
import asyncio
|
import asyncio
|
||||||
|
import shortuuid
|
||||||
|
|
||||||
|
|
||||||
class LogController:
|
class LogController:
|
||||||
|
@ -28,66 +29,34 @@ class LogController:
|
||||||
self.datalogger = {}
|
self.datalogger = {}
|
||||||
self.logsFolderPath = self.cbpi.config_folder.logsFolderPath
|
self.logsFolderPath = self.cbpi.config_folder.logsFolderPath
|
||||||
self.logger.info("Log folder path : " + self.logsFolderPath)
|
self.logger.info("Log folder path : " + self.logsFolderPath)
|
||||||
|
self.sensor_data_listeners = {}
|
||||||
|
|
||||||
def log_data(self, name: str, value: str) -> None:
|
def add_sensor_data_listener(self, method):
|
||||||
self.logfiles = self.cbpi.config.get("CSVLOGFILES", "Yes")
|
listener_id = shortuuid.uuid()
|
||||||
self.influxdb = self.cbpi.config.get("INFLUXDB", "No")
|
self.sensor_data_listeners[listener_id] = method
|
||||||
if self.logfiles == "Yes":
|
return listener_id
|
||||||
if name not in self.datalogger:
|
|
||||||
max_bytes = int(self.cbpi.config.get("SENSOR_LOG_MAX_BYTES", 100000))
|
|
||||||
backup_count = int(self.cbpi.config.get("SENSOR_LOG_BACKUP_COUNT", 3))
|
|
||||||
|
|
||||||
data_logger = logging.getLogger('cbpi.sensor.%s' % name)
|
def remove_sensor_data_listener(self, listener_id):
|
||||||
data_logger.propagate = False
|
|
||||||
data_logger.setLevel(logging.DEBUG)
|
|
||||||
handler = RotatingFileHandler(os.path.join(self.logsFolderPath, f"sensor_{name}.log"), maxBytes=max_bytes, backupCount=backup_count)
|
|
||||||
data_logger.addHandler(handler)
|
|
||||||
self.datalogger[name] = data_logger
|
|
||||||
|
|
||||||
formatted_time = strftime("%Y-%m-%d %H:%M:%S", localtime())
|
|
||||||
self.datalogger[name].info("%s,%s" % (formatted_time, str(value)))
|
|
||||||
|
|
||||||
if self.influxdb == "Yes":
|
|
||||||
## Write to influxdb in an asyncio task
|
|
||||||
self._task = asyncio.create_task(self.log_influx(name,value))
|
|
||||||
|
|
||||||
async def log_influx(self, name:str, value:str):
|
|
||||||
self.influxdbcloud = self.cbpi.config.get("INFLUXDBCLOUD", "No")
|
|
||||||
self.influxdbaddr = self.cbpi.config.get("INFLUXDBADDR", None)
|
|
||||||
self.influxdbname = self.cbpi.config.get("INFLUXDBNAME", None)
|
|
||||||
self.influxdbuser = self.cbpi.config.get("INFLUXDBUSER", None)
|
|
||||||
self.influxdbpwd = self.cbpi.config.get("INFLUXDBPWD", None)
|
|
||||||
self.influxdbmeasurement = self.cbpi.config.get("INFLUXDBMEASUREMENT", "measurement")
|
|
||||||
id = name
|
|
||||||
timeout = Timeout(connect=5.0, read=None)
|
|
||||||
try:
|
try:
|
||||||
sensor=self.cbpi.sensor.find_by_id(name)
|
del self.sensor_data_listener[listener_id]
|
||||||
|
except:
|
||||||
|
self.logger.error("Failed to remove listener {}".format(listener_id))
|
||||||
|
|
||||||
|
async def _call_sensor_data_listeners(self, sensor_id, value, formatted_time, name):
|
||||||
|
for listener_id, method in self.sensor_data_listeners.items():
|
||||||
|
asyncio.create_task(method(self.cbpi, sensor_id, value, formatted_time, name))
|
||||||
|
|
||||||
|
def log_data(self, id: str, value: str) -> None:
|
||||||
|
# all plugin targets:
|
||||||
|
if self.sensor_data_listeners: # true if there are listners
|
||||||
|
try:
|
||||||
|
sensor=self.cbpi.sensor.find_by_id(id)
|
||||||
if sensor is not None:
|
if sensor is not None:
|
||||||
itemname=sensor.name.replace(" ", "_")
|
name = sensor.name.replace(" ", "_")
|
||||||
out=str(self.influxdbmeasurement)+",source=" + itemname + ",itemID=" + str(id) + " value="+str(value)
|
formatted_time = strftime("%Y-%m-%d %H:%M:%S", localtime())
|
||||||
|
asyncio.create_task(self._call_sensor_data_listeners(id, value, formatted_time, name))
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.error("InfluxDB ID Error: {}".format(e))
|
logging.error("sensor logging listener exception: {}".format(e))
|
||||||
|
|
||||||
if self.influxdbcloud == "Yes":
|
|
||||||
self.influxdburl=self.influxdbaddr + "/api/v2/write?org=" + self.influxdbuser + "&bucket=" + self.influxdbname + "&precision=s"
|
|
||||||
try:
|
|
||||||
header = {'User-Agent': name, 'Authorization': "Token {}".format(self.influxdbpwd)}
|
|
||||||
http = PoolManager(timeout=timeout)
|
|
||||||
req = http.request('POST',self.influxdburl, body=out.encode(), headers = header)
|
|
||||||
except Exception as e:
|
|
||||||
logging.error("InfluxDB cloud write Error: {}".format(e))
|
|
||||||
|
|
||||||
else:
|
|
||||||
self.base64string = base64.b64encode(('%s:%s' % (self.influxdbuser,self.influxdbpwd)).encode())
|
|
||||||
self.influxdburl= self.influxdbaddr + '/write?db=' + self.influxdbname
|
|
||||||
try:
|
|
||||||
header = {'User-Agent': name, 'Content-Type': 'application/x-www-form-urlencoded','Authorization': 'Basic %s' % self.base64string.decode('utf-8')}
|
|
||||||
http = PoolManager(timeout=timeout)
|
|
||||||
req = http.request('POST',self.influxdburl, body=out.encode(), headers = header)
|
|
||||||
except Exception as e:
|
|
||||||
logging.error("InfluxDB write Error: {}".format(e))
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
async def get_data(self, names, sample_rate='60s'):
|
async def get_data(self, names, sample_rate='60s'):
|
||||||
logging.info("Start Log for {}".format(names))
|
logging.info("Start Log for {}".format(names))
|
||||||
|
@ -183,6 +152,8 @@ class LogController:
|
||||||
def clear_log(self, name:str ) -> str:
|
def clear_log(self, name:str ) -> str:
|
||||||
all_filenames = glob.glob(os.path.join(self.logsFolderPath, f"sensor_{name}.log*"))
|
all_filenames = glob.glob(os.path.join(self.logsFolderPath, f"sensor_{name}.log*"))
|
||||||
|
|
||||||
|
logging.info(f'Deleting logfiles for sensor {name}.')
|
||||||
|
|
||||||
if name in self.datalogger:
|
if name in self.datalogger:
|
||||||
self.datalogger[name].removeHandler(self.datalogger[name].handlers[0])
|
self.datalogger[name].removeHandler(self.datalogger[name].handlers[0])
|
||||||
del self.datalogger[name]
|
del self.datalogger[name]
|
||||||
|
|
|
@ -278,7 +278,7 @@ class ConfigUpdate(CBPiExtension):
|
||||||
if logfiles is None:
|
if logfiles is None:
|
||||||
logger.info("INIT CSV logfiles")
|
logger.info("INIT CSV logfiles")
|
||||||
try:
|
try:
|
||||||
await self.cbpi.config.add("CSVLOGFILES", "Yes", type=ConfigType.SELECT, description="Write sensor data to csv logfiles",
|
await self.cbpi.config.add("CSVLOGFILES", "Yes", type=ConfigType.SELECT, description="Write sensor data to csv logfiles (enabling requires restart)",
|
||||||
source="craftbeerpi",
|
source="craftbeerpi",
|
||||||
options= [{"label": "Yes", "value": "Yes"},
|
options= [{"label": "Yes", "value": "Yes"},
|
||||||
{"label": "No", "value": "No"}])
|
{"label": "No", "value": "No"}])
|
||||||
|
@ -289,7 +289,7 @@ class ConfigUpdate(CBPiExtension):
|
||||||
if influxdb is None:
|
if influxdb is None:
|
||||||
logger.info("INIT Influxdb")
|
logger.info("INIT Influxdb")
|
||||||
try:
|
try:
|
||||||
await self.cbpi.config.add("INFLUXDB", "No", type=ConfigType.SELECT, description="Write sensor data to influxdb",
|
await self.cbpi.config.add("INFLUXDB", "No", type=ConfigType.SELECT, description="Write sensor data to influxdb (enabling requires restart)",
|
||||||
source="craftbeerpi",
|
source="craftbeerpi",
|
||||||
options= [{"label": "Yes", "value": "Yes"},
|
options= [{"label": "Yes", "value": "Yes"},
|
||||||
{"label": "No", "value": "No"}])
|
{"label": "No", "value": "No"}])
|
||||||
|
|
51
cbpi/extension/SensorLogTarget_CSV/__init__.py
Normal file
51
cbpi/extension/SensorLogTarget_CSV/__init__.py
Normal file
|
@ -0,0 +1,51 @@
|
||||||
|
|
||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
import os
|
||||||
|
from logging.handlers import RotatingFileHandler
|
||||||
|
import logging
|
||||||
|
from unittest.mock import MagicMock, patch
|
||||||
|
import asyncio
|
||||||
|
import random
|
||||||
|
from cbpi.api import *
|
||||||
|
from cbpi.api.config import ConfigType
|
||||||
|
import urllib3
|
||||||
|
import base64
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
class SensorLogTargetCSV(CBPiExtension):
|
||||||
|
|
||||||
|
def __init__(self, cbpi): # called from cbpi on start
|
||||||
|
self.cbpi = cbpi
|
||||||
|
self.logfiles = self.cbpi.config.get("CSVLOGFILES", "Yes")
|
||||||
|
if self.logfiles == "No":
|
||||||
|
return # never run()
|
||||||
|
self._task = asyncio.create_task(self.run()) # one time run() only
|
||||||
|
|
||||||
|
|
||||||
|
async def run(self): # called by __init__ once on start if CSV is enabled
|
||||||
|
self.listener_ID = self.cbpi.log.add_sensor_data_listener(self.log_data_to_CSV)
|
||||||
|
logger.info("CSV sensor log target listener ID: {}".format(self.listener_ID))
|
||||||
|
|
||||||
|
async def log_data_to_CSV(self, cbpi, id:str, value:str, formatted_time, name): # called by log_data() hook from the log file controller
|
||||||
|
self.logfiles = self.cbpi.config.get("CSVLOGFILES", "Yes")
|
||||||
|
if self.logfiles == "No":
|
||||||
|
# We intentionally do not unsubscribe the listener here because then we had no way of resubscribing him without a restart of cbpi
|
||||||
|
# as long as cbpi was STARTED with CSVLOGFILES set to Yes this function is still subscribed, so changes can be made on the fly.
|
||||||
|
# but after initially enabling this logging target a restart is required.
|
||||||
|
return
|
||||||
|
if id not in self.cbpi.log.datalogger:
|
||||||
|
max_bytes = int(self.cbpi.config.get("SENSOR_LOG_MAX_BYTES", 100000))
|
||||||
|
backup_count = int(self.cbpi.config.get("SENSOR_LOG_BACKUP_COUNT", 3))
|
||||||
|
|
||||||
|
data_logger = logging.getLogger('cbpi.sensor.%s' % id)
|
||||||
|
data_logger.propagate = False
|
||||||
|
data_logger.setLevel(logging.DEBUG)
|
||||||
|
handler = RotatingFileHandler(os.path.join(self.cbpi.log.logsFolderPath, f"sensor_{id}.log"), maxBytes=max_bytes, backupCount=backup_count)
|
||||||
|
data_logger.addHandler(handler)
|
||||||
|
self.cbpi.log.datalogger[id] = data_logger
|
||||||
|
|
||||||
|
self.cbpi.log.datalogger[id].info("%s,%s" % (formatted_time, str(value)))
|
||||||
|
|
||||||
|
def setup(cbpi):
|
||||||
|
cbpi.plugin.register("SensorLogTargetCSV", SensorLogTargetCSV)
|
3
cbpi/extension/SensorLogTarget_CSV/config.yaml
Normal file
3
cbpi/extension/SensorLogTarget_CSV/config.yaml
Normal file
|
@ -0,0 +1,3 @@
|
||||||
|
name: SensorLogTargetCSV
|
||||||
|
version: 4
|
||||||
|
active: true
|
76
cbpi/extension/SensorLogTarget_InfluxDB/__init__.py
Normal file
76
cbpi/extension/SensorLogTarget_InfluxDB/__init__.py
Normal file
|
@ -0,0 +1,76 @@
|
||||||
|
|
||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
import os
|
||||||
|
from urllib3 import Timeout, PoolManager
|
||||||
|
import logging
|
||||||
|
from unittest.mock import MagicMock, patch
|
||||||
|
import asyncio
|
||||||
|
import random
|
||||||
|
from cbpi.api import *
|
||||||
|
from cbpi.api.config import ConfigType
|
||||||
|
import urllib3
|
||||||
|
import base64
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
class SensorLogTargetInfluxDB(CBPiExtension):
|
||||||
|
|
||||||
|
def __init__(self, cbpi): # called from cbpi on start
|
||||||
|
self.cbpi = cbpi
|
||||||
|
self.influxdb = self.cbpi.config.get("INFLUXDB", "No")
|
||||||
|
if self.influxdb == "No":
|
||||||
|
return # never run()
|
||||||
|
self._task = asyncio.create_task(self.run()) # one time run() only
|
||||||
|
|
||||||
|
|
||||||
|
async def run(self): # called by __init__ once on start if influx is enabled
|
||||||
|
self.listener_ID = self.cbpi.log.add_sensor_data_listener(self.log_data_to_InfluxDB)
|
||||||
|
logger.info("InfluxDB sensor log target listener ID: {}".format(self.listener_ID))
|
||||||
|
|
||||||
|
async def log_data_to_InfluxDB(self, cbpi, id:str, value:str, timestamp, name): # called by log_data() hook from the log file controller
|
||||||
|
self.influxdb = self.cbpi.config.get("INFLUXDB", "No")
|
||||||
|
if self.influxdb == "No":
|
||||||
|
# We intentionally do not unsubscribe the listener here because then we had no way of resubscribing him without a restart of cbpi
|
||||||
|
# as long as cbpi was STARTED with INFLUXDB set to Yes this function is still subscribed, so changes can be made on the fly.
|
||||||
|
# but after initially enabling this logging target a restart is required.
|
||||||
|
return
|
||||||
|
self.influxdbcloud = self.cbpi.config.get("INFLUXDBCLOUD", "No")
|
||||||
|
self.influxdbaddr = self.cbpi.config.get("INFLUXDBADDR", None)
|
||||||
|
self.influxdbname = self.cbpi.config.get("INFLUXDBNAME", None)
|
||||||
|
self.influxdbuser = self.cbpi.config.get("INFLUXDBUSER", None)
|
||||||
|
self.influxdbpwd = self.cbpi.config.get("INFLUXDBPWD", None)
|
||||||
|
self.influxdbmeasurement = self.cbpi.config.get("INFLUXDBMEASUREMENT", "measurement")
|
||||||
|
timeout = Timeout(connect=5.0, read=None)
|
||||||
|
try:
|
||||||
|
sensor=self.cbpi.sensor.find_by_id(id)
|
||||||
|
if sensor is not None:
|
||||||
|
itemname=sensor.name.replace(" ", "_")
|
||||||
|
out=str(self.influxdbmeasurement)+",source=" + itemname + ",itemID=" + str(id) + " value="+str(value)
|
||||||
|
except Exception as e:
|
||||||
|
logging.error("InfluxDB ID Error: {}".format(e))
|
||||||
|
|
||||||
|
if self.influxdbcloud == "Yes":
|
||||||
|
self.influxdburl=self.influxdbaddr + "/api/v2/write?org=" + self.influxdbuser + "&bucket=" + self.influxdbname + "&precision=s"
|
||||||
|
try:
|
||||||
|
header = {'User-Agent': id, 'Authorization': "Token {}".format(self.influxdbpwd)}
|
||||||
|
http = PoolManager(timeout=timeout)
|
||||||
|
req = http.request('POST',self.influxdburl, body=out.encode(), headers = header)
|
||||||
|
if req.status != 204:
|
||||||
|
raise Exception(f'InfluxDB Status code {req.status}')
|
||||||
|
except Exception as e:
|
||||||
|
logging.error("InfluxDB cloud write Error: {}".format(e))
|
||||||
|
|
||||||
|
else:
|
||||||
|
self.base64string = base64.b64encode(('%s:%s' % (self.influxdbuser,self.influxdbpwd)).encode())
|
||||||
|
self.influxdburl= self.influxdbaddr + '/write?db=' + self.influxdbname
|
||||||
|
try:
|
||||||
|
header = {'User-Agent': id, 'Content-Type': 'application/x-www-form-urlencoded','Authorization': 'Basic %s' % self.base64string.decode('utf-8')}
|
||||||
|
http = PoolManager(timeout=timeout)
|
||||||
|
req = http.request('POST',self.influxdburl, body=out.encode(), headers = header)
|
||||||
|
if req.status != 204:
|
||||||
|
raise Exception(f'InfluxDB Status code {req.status}')
|
||||||
|
except Exception as e:
|
||||||
|
logging.error("InfluxDB write Error: {}".format(e))
|
||||||
|
|
||||||
|
def setup(cbpi):
|
||||||
|
cbpi.plugin.register("SensorLogTargetInfluxDB", SensorLogTargetInfluxDB)
|
3
cbpi/extension/SensorLogTarget_InfluxDB/config.yaml
Normal file
3
cbpi/extension/SensorLogTarget_InfluxDB/config.yaml
Normal file
|
@ -0,0 +1,3 @@
|
||||||
|
name: SensorLogTargetInfluxDB
|
||||||
|
version: 4
|
||||||
|
active: true
|
|
@ -80,7 +80,7 @@
|
||||||
"options": null,
|
"options": null,
|
||||||
"source": "hidden",
|
"source": "hidden",
|
||||||
"type": "string",
|
"type": "string",
|
||||||
"value": "4.1.8.a11"
|
"value": "4.1.10.a1"
|
||||||
},
|
},
|
||||||
"CSVLOGFILES": {
|
"CSVLOGFILES": {
|
||||||
"description": "Write sensor data to csv logfiles",
|
"description": "Write sensor data to csv logfiles",
|
||||||
|
@ -117,7 +117,7 @@
|
||||||
"value": "No"
|
"value": "No"
|
||||||
},
|
},
|
||||||
"INFLUXDBADDR": {
|
"INFLUXDBADDR": {
|
||||||
"description": "IP Address of your influxdb server (If INFLUXDBCLOUD set to Yes use URL Address of your influxdb cloud server)",
|
"description": "URL Address of your influxdb server incl. http:// and port, e.g. http://localhost:8086 (If INFLUXDBCLOUD set to Yes use URL Address of your influxdb cloud server)",
|
||||||
"name": "INFLUXDBADDR",
|
"name": "INFLUXDBADDR",
|
||||||
"options": null,
|
"options": null,
|
||||||
"source": "craftbeerpi",
|
"source": "craftbeerpi",
|
||||||
|
@ -157,14 +157,6 @@
|
||||||
"type": "string",
|
"type": "string",
|
||||||
"value": "cbpi4"
|
"value": "cbpi4"
|
||||||
},
|
},
|
||||||
"INFLUXDBPORT": {
|
|
||||||
"description": "Port of your influxdb server",
|
|
||||||
"name": "INFLUXDBPORT",
|
|
||||||
"options": null,
|
|
||||||
"source": "craftbeerpi",
|
|
||||||
"type": "string",
|
|
||||||
"value": "8086"
|
|
||||||
},
|
|
||||||
"INFLUXDBPWD": {
|
"INFLUXDBPWD": {
|
||||||
"description": "Password for your influxdb database (only if required)(If INFLUXDBCLOUD set to Yes use token of your influxdb cloud database)",
|
"description": "Password for your influxdb database (only if required)(If INFLUXDBCLOUD set to Yes use token of your influxdb cloud database)",
|
||||||
"name": "INFLUXDBPWD",
|
"name": "INFLUXDBPWD",
|
||||||
|
@ -174,7 +166,7 @@
|
||||||
"value": " "
|
"value": " "
|
||||||
},
|
},
|
||||||
"INFLUXDBUSER": {
|
"INFLUXDBUSER": {
|
||||||
"description": "User name for your influxdb database (only if required)(If INFLUXDBCLOUD set to Yes use organisation of your influxdb cloud database)",
|
"description": "User Name for your influxdb database (only if required)(If INFLUXDBCLOUD set to Yes use organisation of your influxdb cloud database)",
|
||||||
"name": "INFLUXDBUSER",
|
"name": "INFLUXDBUSER",
|
||||||
"options": null,
|
"options": null,
|
||||||
"source": "craftbeerpi",
|
"source": "craftbeerpi",
|
||||||
|
|
|
@ -1,3 +1,11 @@
|
||||||
{
|
{
|
||||||
"data": []
|
"data": [
|
||||||
|
{
|
||||||
|
"id": "unconfigured_test_sensor_ID",
|
||||||
|
"name": "unconfigured_mqtt_sensor",
|
||||||
|
"props": {},
|
||||||
|
"state": false,
|
||||||
|
"type": "MQTTSensor"
|
||||||
|
}
|
||||||
|
]
|
||||||
}
|
}
|
|
@ -10,10 +10,10 @@ class LoggerTestCase(CraftBeerPiTestCase):
|
||||||
async def test_log_data(self):
|
async def test_log_data(self):
|
||||||
|
|
||||||
os.makedirs(os.path.join(".", "tests", "logs"), exist_ok=True)
|
os.makedirs(os.path.join(".", "tests", "logs"), exist_ok=True)
|
||||||
log_name = "test"
|
log_name = "unconfigured_test_sensor_ID"
|
||||||
#clear all logs
|
#clear all logs
|
||||||
self.cbpi.log.clear_log(log_name)
|
self.cbpi.log.clear_log(log_name)
|
||||||
assert len(glob.glob(os.path.join(".", "tests", "logs", f"sensor_{log_name}.log*"))) == 0
|
assert len(glob.glob(os.path.join(self.cbpi.log.logsFolderPath, f"sensor_{log_name}.log*"))) == 0
|
||||||
|
|
||||||
# write log entries
|
# write log entries
|
||||||
for i in range(5):
|
for i in range(5):
|
||||||
|
|
Loading…
Reference in a new issue