Difference between revisions of "Data Loggers"

From HSYCO
Jump to navigation Jump to search
 
(27 intermediate revisions by 2 users not shown)
Line 19: Line 19:
 
The first step to create a counter data logger is to define it in the HSYCO configuration.
 
The first step to create a counter data logger is to define it in the HSYCO configuration.
  
Refer to the [[Settings#Data_Loggers|Data Loggers]] section in the configuration chapter for a complete description of all the available parameters and options.
+
Refer to the [[Settings#Data_Loggers|Data Loggers section]] in the configuration chapter for a complete description of all the available parameters and options.
  
 
In the following examples we have defined a counter data logger named “energy” with default parameters.
 
In the following examples we have defined a counter data logger named “energy” with default parameters.
Line 25: Line 25:
 
Now, we need to provide the data logger with the incremental value we want to process.
 
Now, we need to provide the data logger with the incremental value we want to process.
  
Assuming we have an I/O server named “elektro” which provides the energy consumption value, we write the following lines in events.txt:
+
Assuming we have an I/O datapoint named “elektro.kwh” which provides the energy consumption value, we write the following lines in EVENTS:
  
  IO elektro : DATALOGGER energy = IO elektro, DATALOGGER energy = FILE LOG energy.csv
+
  IO elektro.kwh : DATALOGGER energy = IO elektro.kwh, DATALOGGER energy = FILE LOG energy.csv
  
Whenever the I/O server will provide a new value, the data logger will be updated and a log entry will be added to the file “energy.csv”.
+
Whenever the I/O datapoint will provide a new value, the data logger will be updated and a log entry will be added to the file “energy.csv”.
  
Finally, in order to display the data logger’s charts, we should use the (datalogger) object, for example:
+
Finally, in order to display the data logger’s charts, we should use the [[Datalogger]] object and set its ID to "energy"
 
 
(datalogger energy; x80y60; 820; 500; kWh;)
 
  
 
For a complete reference of all the available actions on data loggers refer to the relative section in the Events programming chapter.
 
For a complete reference of all the available actions on data loggers refer to the relative section in the Events programming chapter.
Line 39: Line 37:
 
The same actions can be performed using JavaScript or Java code.
 
The same actions can be performed using JavaScript or Java code.
  
We could also create a data logger to monitor the expenses relative to energy consumption, initializing a new counter data logger and declaring the time slot rates in the file dataloggers.ini.
+
We could also create a data logger to monitor the expenses relative to energy consumption, initializing a new counter data logger and declaring the time slot rates in the file dataloggers.ini located in HSYCO's root folder.
  
In hsyco.ini we declare our new data logger:
+
We add a new data logger with ID "cost" in Settings, then, we set the "elektro.kwh" I/O datapoint to update the data logger. In EVENTS we add this line:
DataLoggers = cost
 
DataLoggers.cost.Type = counter
 
  
Then, we set the elektro I/O server to update the data logger. In events.txt we add this line:
+
  IO elektro.kwh OR TIME : DATALOGGER cost = IO elektro.kwh
  IO elektro OR TIME : DATALOGGER cost = IO elektro
 
  
 
Note that this time we are updating the data logger not only when a new value is available, but also every minute. This is necessary to have an accurate calculation of the expenses.
 
Note that this time we are updating the data logger not only when a new value is available, but also every minute. This is necessary to have an accurate calculation of the expenses.
Line 71: Line 66:
 
Refer to the section about the dataloggers.ini file for a complete reference on how to write time slots rules.
 
Refer to the section about the dataloggers.ini file for a complete reference on how to write time slots rules.
  
At this point we should add the (datalogger) object to the Web interface to display the charts and totals. The user interface also allows you to filter data by rate.
+
At this point we should add the Datalogger object to the user interface to display the charts. The user interface also allows you to filter data by rate.
 +
 
 +
Further, we can have a detailed report of all the processed data by specifying the path of a log file using the "Rates Log File" option in the data logger's settings:
  
Further, we can have a detailed report of all the processed data by specifying the path of a log file in hsyco.ini:
+
[[File:Data_Loggers_Rates_Log_File.png|border]]
DataLoggers = cost
 
DataLoggers.cost.Type = counter
 
DataLoggers.cost.RatesLogFile = counterRates/log.csv
 
  
 
This will create a CSV file listing a row for each update of the data logger divided in the following columns:
 
This will create a CSV file listing a row for each update of the data logger divided in the following columns:
 +
 
  data_logger_name, update_time, provided_value, calculated_cost, slot_id, applied_rate
 
  data_logger_name, update_time, provided_value, calculated_cost, slot_id, applied_rate
  
Line 85: Line 80:
 
In the following examples we define a range data logger, named “temperature”, with default parameters.
 
In the following examples we define a range data logger, named “temperature”, with default parameters.
  
Note that the average is calculated as the sum of the values divided by the number of records.
+
We set the data logger to follow the temperature measured by a probe identified by the I/O datapoint “termo.temp”.
 +
 
 +
We specify in events that every minute it must be updated, plus we log the measured value in the file “temp.csv”:
 +
 
 +
TIME : DATALOGGER temperature = IO termo.temp, DATALOGGER temperature = FILE LOG temp.csv
 +
 
 +
The created charts will automatically be scaled to range from the lowest measured value to the highest and charts belonging to the same period will be aligned.
 +
 
 +
{{note|The average is calculated as the sum of the values divided by the number of records. It is therefore advised to provide the data logger with periodical values so to have a correct calculation.}}
 +
 
 +
By default, the origin of the charts is set to correspond to the lowest measured value.
 +
 
 +
One might prefer to set the origin to a specified value (for instance zero) so to highlight the difference between values below and above it.
 +
 
 +
We can achieve this by setting the Origin option in the data logger settings:
 +
 
 +
[[File:Data_Loggers_Origin.png]]
 +
 
 +
Finally, in order to display the data logger’s charts, we should add the Datalogger object to our interface.
 +
 
 +
== CSV Import/Export ==
  
It is therefore advised to provide the data logger with periodical values so to have a correct calculation.
+
Using the HSYCO's development API, in Events, JavaScript and Java, you can export a full dump of all data underlying each data logger.
  
Now, we set the data logger to follow the temperature measured by a probe identified by the I/O server “temp”.
+
The CSV file uses \r\n as line separator and ';' as field separator.
  
We specify in events that every minute it must be updated, plus we log the measured value in the file “temp.csv”:
+
This is an example of the CSV file for a counter data logger.
  TIME : DATALOGGER temperature = IO temp, DATALOGGER temperature = FILE LOG temp.csv
+
 
 +
ts;delta;cost;locked
 +
20160129130000;2.0830569986031082;0.0;1
 +
20160129140000;3.464736775902562;0.0;1
 +
20160129150000;0.04146936782360733;0.0;1
 +
20160220120100;0.005090918751661816;0.0;0
 +
20160220120200;0.039745587753242655;0.0;0
 +
 
 +
The record format has the following four fields:
 +
* ts: the timestamp as YYYYMMDDhhmmss
 +
* delta: a double precision floating point of the accumulated values for the period between this record's timestamp and the next
 +
* cost: a double precision floating point of the cost value for the period
 +
* locked: a boolean, represented as 1 for locked records, and 0 for unlocked records. A record is marked as locked when its cost value is permanently set, and will not change if the rates rules are changed in the future.
 +
 
 +
This is an example of the CSV file for a range data logger.
 +
 
 +
ts;vmin;vmax;vavg;vcount
 +
20170114150100;3.607471109385042;4.694435324229342;4.241931611566256;15
 +
20170114150200;3.8066206378643987;5.446351986025894;4.575574507553171;16
 +
20170114150300;2.398800160049782;5.2533792997638775;3.6368449963207854;17
 +
20170114150400;1.5474068317929874;3.1531099637853406;2.290833667227986;10
 +
 
 +
The record format has the following four fields:
 +
* ts: the timestamp as YYYYMMDDhhmmss
 +
* vmin: a double precision floating point for the minimum value logged in the period
 +
* vmax: a double precision floating point for the maximum value logged in the period
 +
* vavg: a double precision floating point for the average value logged in the period
 +
* vcount: a positive integer counter of the number of values logged in the period; this is used to compute a weighted average when new values are passed to the data logger.
 +
 
 +
=== Export API ===
 +
 
 +
Use the following event action to create a CSV file for a data logger. The file, named &lt;data logger id&gt;.csv is written in the <path> directory:
 +
  DATALOGGER name = FILE CSVWRITE path
 +
 
 +
You can also use the equivalent JavaScript command API:
 +
dataLoggerSave("csvwrite", names, path, timestamp)
 +
 
 +
And the Java command API (setting action to "csvwrite"):
 +
static boolean dataLoggerSave(String action, String[] names, String path, boolean timestamp)
 +
 
 +
=== Import API ===
 +
 
 +
Raw data, using the same CSV format of the export APIs, can be imported to a data logger.
 +
 
 +
You can selectively update existing records, delete records, but also completely wipe existing data and replace them with the data from a the CSV file or a string.
 +
To delete records, leave just the timestamp value in the CSV file for the records you want to delete, for example:
 +
 
 +
ts;vmin;vmax;vavg;vcount
 +
20170114150200;3.8066206378643987;5.446351986025894;4.575574507553171;16
 +
20170114150300
 +
20170114150400
 +
This file will update record with update the values of one record, and delete two records.
 +
 
 +
Note that the first record in the CSV file is assumed to be the header, and not considered as valid data.
  
The created charts will automatically be scaled to range from the lowest measured value to the highest and charts belonging to the same period will be aligned. If for instance we only want to show the chart showing the lowest values, it might be better not to have this chart aligned to the others.
+
At the end of the import process, one or two files are generated, with the same name of the import file, with the "-ok" and "-err" suffixes to the data logger id, before the .csv extension.
 +
The ok file contains a copy of all records that have been successfully processed, while the err file will contain all records that were discarded due to format errors or other error conditions. If no errors are detected, only the ok file is created during the import process.
  
To do this, we can specify in hsyco.ini to scale the charts of the same period independently:
+
Use the following event action to upload a CSV file to insert, update or delete records of a data logger. The file, named &lt;data logger id&gt;.csv, is read from the &lt; path&gt; directory:
  DataLoggers.temperature.PeriodAlign = false
+
  DATALOGGER name = FILE CSVREAD path
  
Further, by default, the origin of the charts is set to correspond to the lowest measured value.
+
Use the following event action to upload a CSV file that will replace all records of a data logger:
 +
DATALOGGER name = FILE CSVREADOVER path
  
One might prefer to set the origin to a specified value (for instance zero) so to highlight the difference between values below and above it.
+
You can upload a CSV line directly from a string instead of from a file, just replace the file path with the complete CSV string:
 +
DATALOGGER name = FILE CSVREAD "20170114150200;3.8066206378643987;5.446351986025894;4.575574507553171;16"
 +
DATALOGGER name = FILE CSVREAD "20170114150300"
  
We can achieve this by setting the Origin option:
+
You can also use the equivalent JavaScript command API:
  DataLoggers.temperature.Origin = 0
+
  dataLoggerSave("csvread", names, path, timestamp)
 +
dataLoggerSave("csvreadover", names, path, timestamp)
  
Finally, in order to display the data logger’s charts, we should use the (datalogger) object, for example:
+
And the Java command API (setting action to "csvread" or "csvreadover"):
  (datalogger temperature; x80y60; 820; 500; C;)
+
  static boolean dataLoggerSave(String action, String[] names, String path, boolean timestamp)

Latest revision as of 10:51, 21 January 2020

UI Object datalogger.1.png
UI Object datalogger.2.png
UI Object datalogger.3.png

Data loggers allow to gather, process and visualize statistical data on the variations of a data set.

They can be used to automatically generate periodic charts of its trend and/or to log data as CSV files.

Data loggers collect data and group them in order to visualize the trends in charts for each hour (or group of hours) of the current and past day, each day of the current and past month, and each month of the current and past year. It is also possible to have an higher resolution of a single minute, but this implies a notably higher disk space usage.

Further, they allows for the logging of every processed value.

There are two types of data logger:

  • counter: suitable for the creation of statistics of an incremental value (e.g. energy consumption or production). It calculates the variation (delta) of the value for each time interval with respect to the previous one. Further, it is possible to specify time slots and the relative rates to calculate for instance the costs for energy consumption.
  • range: aimed at the computation of the maximum, minimum, and average of a fluctuating value (e.g. a measured temperature or humidity) for each time interval.

Once a data logger is defined in HSYCO's Settings, you can easily pass data to it using a simple action in EVENTS, a JavaScript function or Java method, and display statistical data using the Datalogger UI object.

Counter Data Loggers

The first step to create a counter data logger is to define it in the HSYCO configuration.

Refer to the Data Loggers section in the configuration chapter for a complete description of all the available parameters and options.

In the following examples we have defined a counter data logger named “energy” with default parameters.

Now, we need to provide the data logger with the incremental value we want to process.

Assuming we have an I/O datapoint named “elektro.kwh” which provides the energy consumption value, we write the following lines in EVENTS:

IO elektro.kwh : DATALOGGER energy = IO elektro.kwh, DATALOGGER energy = FILE LOG energy.csv

Whenever the I/O datapoint will provide a new value, the data logger will be updated and a log entry will be added to the file “energy.csv”.

Finally, in order to display the data logger’s charts, we should use the Datalogger object and set its ID to "energy"

For a complete reference of all the available actions on data loggers refer to the relative section in the Events programming chapter.

The same actions can be performed using JavaScript or Java code.

We could also create a data logger to monitor the expenses relative to energy consumption, initializing a new counter data logger and declaring the time slot rates in the file dataloggers.ini located in HSYCO's root folder.

We add a new data logger with ID "cost" in Settings, then, we set the "elektro.kwh" I/O datapoint to update the data logger. In EVENTS we add this line:

IO elektro.kwh OR TIME : DATALOGGER cost = IO elektro.kwh

Note that this time we are updating the data logger not only when a new value is available, but also every minute. This is necessary to have an accurate calculation of the expenses.

Now we need to declare our time slots in the file dataloggers.ini, for instance:

cost; 0; 00:00-11:59; *; *; 0.3
cost; 1; 12:00-16:59; *; *; 1.4
cost; 2; 17:00-23:59; *; *; 0.8

These rules state that every day from midnight to 11:59:59 we are in slot 0 and the rate to be applied is 0.3, from 12:00:00 until 16:59:59 we are in slot 1 and the applicable rate is 1.4, later on slot 2 applies with a rate of 0.8.

We can include some exceptions by adding some lines before these rules:

cost; 0; *; *; 2000/12/10-2000/12/31; 0.3
cost; 2; *; 67; *; 0.8
cost; 0; 00:00-11:59; *; *; 0.3
cost; 1; 12:00-16:59; *; *; 1.4
cost; 2; 17:00-23:59; *; *; 0.8

The previous rules will still apply, except on Saturdays (6) and Sundays (7) when during all day slot 2 will apply.

Further, during the period between December 10th and December 31st of 2000, slot 0 will always apply despite the day of the week or the time.

Refer to the section about the dataloggers.ini file for a complete reference on how to write time slots rules.

At this point we should add the Datalogger object to the user interface to display the charts. The user interface also allows you to filter data by rate.

Further, we can have a detailed report of all the processed data by specifying the path of a log file using the "Rates Log File" option in the data logger's settings:

Data Loggers Rates Log File.png

This will create a CSV file listing a row for each update of the data logger divided in the following columns:

data_logger_name, update_time, provided_value, calculated_cost, slot_id, applied_rate

Range Data Loggers

In the following examples we define a range data logger, named “temperature”, with default parameters.

We set the data logger to follow the temperature measured by a probe identified by the I/O datapoint “termo.temp”.

We specify in events that every minute it must be updated, plus we log the measured value in the file “temp.csv”:

TIME : DATALOGGER temperature = IO termo.temp, DATALOGGER temperature = FILE LOG temp.csv

The created charts will automatically be scaled to range from the lowest measured value to the highest and charts belonging to the same period will be aligned.

The average is calculated as the sum of the values divided by the number of records. It is therefore advised to provide the data logger with periodical values so to have a correct calculation.


By default, the origin of the charts is set to correspond to the lowest measured value.

One might prefer to set the origin to a specified value (for instance zero) so to highlight the difference between values below and above it.

We can achieve this by setting the Origin option in the data logger settings:

Data Loggers Origin.png

Finally, in order to display the data logger’s charts, we should add the Datalogger object to our interface.

CSV Import/Export

Using the HSYCO's development API, in Events, JavaScript and Java, you can export a full dump of all data underlying each data logger.

The CSV file uses \r\n as line separator and ';' as field separator.

This is an example of the CSV file for a counter data logger.

ts;delta;cost;locked
20160129130000;2.0830569986031082;0.0;1
20160129140000;3.464736775902562;0.0;1
20160129150000;0.04146936782360733;0.0;1
20160220120100;0.005090918751661816;0.0;0
20160220120200;0.039745587753242655;0.0;0

The record format has the following four fields:

  • ts: the timestamp as YYYYMMDDhhmmss
  • delta: a double precision floating point of the accumulated values for the period between this record's timestamp and the next
  • cost: a double precision floating point of the cost value for the period
  • locked: a boolean, represented as 1 for locked records, and 0 for unlocked records. A record is marked as locked when its cost value is permanently set, and will not change if the rates rules are changed in the future.

This is an example of the CSV file for a range data logger.

ts;vmin;vmax;vavg;vcount
20170114150100;3.607471109385042;4.694435324229342;4.241931611566256;15
20170114150200;3.8066206378643987;5.446351986025894;4.575574507553171;16
20170114150300;2.398800160049782;5.2533792997638775;3.6368449963207854;17
20170114150400;1.5474068317929874;3.1531099637853406;2.290833667227986;10

The record format has the following four fields:

  • ts: the timestamp as YYYYMMDDhhmmss
  • vmin: a double precision floating point for the minimum value logged in the period
  • vmax: a double precision floating point for the maximum value logged in the period
  • vavg: a double precision floating point for the average value logged in the period
  • vcount: a positive integer counter of the number of values logged in the period; this is used to compute a weighted average when new values are passed to the data logger.

Export API

Use the following event action to create a CSV file for a data logger. The file, named <data logger id>.csv is written in the <path> directory:

DATALOGGER name = FILE CSVWRITE path

You can also use the equivalent JavaScript command API:

dataLoggerSave("csvwrite", names, path, timestamp)

And the Java command API (setting action to "csvwrite"):

static boolean dataLoggerSave(String action, String[] names, String path, boolean timestamp)

Import API

Raw data, using the same CSV format of the export APIs, can be imported to a data logger.

You can selectively update existing records, delete records, but also completely wipe existing data and replace them with the data from a the CSV file or a string. To delete records, leave just the timestamp value in the CSV file for the records you want to delete, for example:

ts;vmin;vmax;vavg;vcount
20170114150200;3.8066206378643987;5.446351986025894;4.575574507553171;16
20170114150300
20170114150400

This file will update record with update the values of one record, and delete two records.

Note that the first record in the CSV file is assumed to be the header, and not considered as valid data.

At the end of the import process, one or two files are generated, with the same name of the import file, with the "-ok" and "-err" suffixes to the data logger id, before the .csv extension. The ok file contains a copy of all records that have been successfully processed, while the err file will contain all records that were discarded due to format errors or other error conditions. If no errors are detected, only the ok file is created during the import process.

Use the following event action to upload a CSV file to insert, update or delete records of a data logger. The file, named <data logger id>.csv, is read from the < path> directory:

DATALOGGER name = FILE CSVREAD path

Use the following event action to upload a CSV file that will replace all records of a data logger:

DATALOGGER name = FILE CSVREADOVER path

You can upload a CSV line directly from a string instead of from a file, just replace the file path with the complete CSV string:

DATALOGGER name = FILE CSVREAD "20170114150200;3.8066206378643987;5.446351986025894;4.575574507553171;16"
DATALOGGER name = FILE CSVREAD "20170114150300"

You can also use the equivalent JavaScript command API:

dataLoggerSave("csvread", names, path, timestamp)
dataLoggerSave("csvreadover", names, path, timestamp)

And the Java command API (setting action to "csvread" or "csvreadover"):

static boolean dataLoggerSave(String action, String[] names, String path, boolean timestamp)