From 23d260a351c0f0ceba9d6bf151c5d49bf074fd04 Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Wed, 17 Jan 2024 12:57:38 +0000 Subject: [PATCH 01/15] Prettified Code! --- documentation/FreeDATA-protocols.md | 108 +++++++++++++--------------- 1 file changed, 51 insertions(+), 57 deletions(-) diff --git a/documentation/FreeDATA-protocols.md b/documentation/FreeDATA-protocols.md index 68bf5ded..8cf833d8 100644 --- a/documentation/FreeDATA-protocols.md +++ b/documentation/FreeDATA-protocols.md @@ -1,6 +1,7 @@ # FreeDATA - Protocols ## ARQ Sessions + An ARQ Session represents a reliable data transmission session from a sending station (A) to a receiving station (B). It uses automatic repeat request on top of different codec2 modes according to the transmission channel conditions. So lets say A wants to send some data to B. A typical scenario would be like this: @@ -22,22 +23,19 @@ ISS->(1)IRS:BURST (ID, offset, payload),(ID, offset, payload),(ID, offset, paylo IRS->(1)ISS:DATA ACK NACK (ID, next_offset, speed level, frames, snr) ``` - ### Frame details - #### SESSION_OPEN_REQ ISS sends this first DATAC13 Mode (12 bytes) -|field|bytes| -|-|-| -|session id|1| -|origin|6| -|destination_crc|3| - +| field | bytes | +| --------------- | ----- | +| session id | 1 | +| origin | 6 | +| destination_crc | 3 | #### SESSION_OPEN_ACK @@ -45,14 +43,13 @@ Sent by the IRS in response to a SESSION_OPEN_REQ DATAC13 Mode (12 bytes) -|field|bytes| -|-|-| -|session id|1| -|origin|6| -|destination_crc|3| -|protocol version|1| -|snr|1| - +| field | bytes | +| ---------------- | ----- | +| session id | 1 | +| origin | 6 | +| destination_crc | 3 | +| protocol version | 1 | +| snr | 1 | #### SESSION_INFO @@ -60,13 +57,12 @@ ISS sends this in response to a SESSION_OPEN_ACK DATAC13 Mode (12 bytes) -|field|bytes| -|-|-| -|session id|1| -|total bytes|4| -|total crc|4| -|snr|1| - +| field | bytes | +| ----------- | ----- | +| session id | 1 | +| total bytes | 4 | +| total crc | 4 | +| snr | 1 | #### SESSION_INFO_ACK @@ -74,14 +70,13 @@ IRS sends this in response to a SESSION_INFO DATAC13 Mode (12 bytes) -|field|bytes| -|-|-| -|session id|1| -|total crc|4| -|snr|1| -|speed level|1| -|frames per burst|1| - +| field | bytes | +| ---------------- | ----- | +| session id | 1 | +| total crc | 4 | +| snr | 1 | +| speed level | 1 | +| frames per burst | 1 | #### Data Burst @@ -92,50 +87,49 @@ Mode according to handshake speed level Frames per burst according to handshake ##### Modulation + Each burst is composed of frames_per_burst frames: |preamble|f1|f2|f3|...|postamble| ##### Each data frame -|field|bytes| -|-|-| -|session id|1| -|offset|4| -|payload|(the remaining payload length)| - +| field | bytes | +| ---------- | ------------------------------ | +| session id | 1 | +| offset | 4 | +| payload | (the remaining payload length) | #### DATA_BURST_ACK Sent by the IRS following successful decoding of burst. -|field|bytes| -|-|-| -|session id|1| -|next offset|4| -|next speed level|1| -|next frames per burst|1| -|snr|1| - +| field | bytes | +| --------------------- | ----- | +| session id | 1 | +| next offset | 4 | +| next speed level | 1 | +| next frames per burst | 1 | +| snr | 1 | #### DATA_BURST_NACK Sent by the IRS following unsuccessful decoding of burst or timeout. -|field|bytes| -|-|-| -|session id|1| -|next offset|4| -|next speed level|1| -|next frames per burst|1| -|snr|1| +| field | bytes | +| --------------------- | ----- | +| session id | 1 | +| next offset | 4 | +| next speed level | 1 | +| next frames per burst | 1 | +| snr | 1 | #### DATA ACK NACK Sent by the IRS after receiving data with a state information. -| field |bytes| -|------------|-| -| session id |1| -| state |1| -| snr |1| \ No newline at end of file +| field | bytes | +| ---------- | ----- | +| session id | 1 | +| state | 1 | +| snr | 1 | From 2e2444eb471dff9f3ec2f993b1b513e49960e0c2 Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Wed, 17 Jan 2024 13:57:55 +0100 Subject: [PATCH 02/15] version update --- gui/package.json | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gui/package.json b/gui/package.json index 85f605b7..b3a89bf5 100644 --- a/gui/package.json +++ b/gui/package.json @@ -2,7 +2,7 @@ "name": "FreeDATA", "description": "FreeDATA", "private": true, - "version": "0.11.1-alpha.3", + "version": "0.12.0-alpha", "main": "dist-electron/main/index.js", "scripts": { "start": "vite", From f0a8b92d1b888041cb672f9c21726b08dd8af9ba Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Wed, 17 Jan 2024 19:44:39 +0100 Subject: [PATCH 03/15] fixed missing config for bundles --- gui/package.json | 2 +- gui/src/js/chatHandler.ts | 31 +++---------------------------- gui/src/js/freedata.ts | 25 +++++++++++++++++++++++++ gui/src/store/settingsStore.js | 16 +++++++++++++--- modem/server.py | 2 +- 5 files changed, 43 insertions(+), 33 deletions(-) diff --git a/gui/package.json b/gui/package.json index b3a89bf5..0517b2ae 100644 --- a/gui/package.json +++ b/gui/package.json @@ -2,7 +2,7 @@ "name": "FreeDATA", "description": "FreeDATA", "private": true, - "version": "0.12.0-alpha", + "version": "0.12.1-alpha", "main": "dist-electron/main/index.js", "scripts": { "start": "vite", diff --git a/gui/src/js/chatHandler.ts b/gui/src/js/chatHandler.ts index 258673b4..4265abd0 100644 --- a/gui/src/js/chatHandler.ts +++ b/gui/src/js/chatHandler.ts @@ -14,6 +14,7 @@ import { useStateStore } from "../store/stateStore.js"; const state = useStateStore(pinia); import { settingsStore as settings } from "../store/settingsStore.js"; +import {getAppDataPath} from "../js/freedata"; import { displayToast } from "./popupHandler.js"; @@ -99,34 +100,8 @@ PouchDB.plugin(require("pouchdb-find")); //PouchDB.plugin(require('pouchdb-replication')); PouchDB.plugin(require("pouchdb-upsert")); -// https://stackoverflow.com/a/26227660 -if (typeof process.env["APPDATA"] !== "undefined") { - var appDataFolder = process.env["APPDATA"]; - console.log(appDataFolder); -} else { - var appDataFolder: string; - - switch (process.platform) { - case "darwin": - appDataFolder = process.env["HOME"] + "/Library/Application Support"; - console.log(appDataFolder); - break; - case "linux": - appDataFolder = process.env["HOME"] + "/.config"; - console.log(appDataFolder); - break; - case "win32": - appDataFolder = "undefined"; - break; - default: - appDataFolder = "undefined"; - break; - } -} -console.log("loading chat database..."); -console.log("appdata folder:" + appDataFolder); -var configFolder = path.join(appDataFolder, "FreeDATA"); -console.log("config folder:" + configFolder); +var appDataPath = getAppDataPath() +var configFolder = path.join(appDataPath, "FreeDATA"); var chatDB = path.join(configFolder, "chatDB"); console.log("database path:" + chatDB); diff --git a/gui/src/js/freedata.ts b/gui/src/js/freedata.ts index 8e8ffd10..a42d0edb 100644 --- a/gui/src/js/freedata.ts +++ b/gui/src/js/freedata.ts @@ -1,3 +1,6 @@ +const os = require('os'); +const path = require('path'); + /** * Binary to ASCII replacement * @param {string} data in normal/usual utf-8 format @@ -97,3 +100,25 @@ export function validateCallsignWithoutSSID(callsign: string) { } return true; } + +export function getAppDataPath(){ + const platform = os.platform(); + let appDataPath; + + switch (platform) { + case 'darwin': // macOS + appDataPath = path.join(os.homedir(), 'Library', 'Application Support'); + break; + case 'win32': // Windows + appDataPath = process.env.APPDATA || path.join(os.homedir(), 'AppData', 'Roaming'); + break; + case 'linux': // Linux + appDataPath = path.join(os.homedir(), '.config'); + break; + default: + throw new Error('Unsupported platform'); + } + + return appDataPath; + +} \ No newline at end of file diff --git a/gui/src/store/settingsStore.js b/gui/src/store/settingsStore.js index 9ee10885..a3ae3040 100644 --- a/gui/src/store/settingsStore.js +++ b/gui/src/store/settingsStore.js @@ -1,9 +1,19 @@ import { reactive, ref, watch } from "vue"; - import { getConfig, setConfig } from "../js/api"; +import {getAppDataPath} from "../js/freedata"; +import fs from "fs"; +const path = require('path'); +const nconf = require("nconf"); -var nconf = require("nconf"); -nconf.file({ file: "config/config.json" }); +var appDataPath = getAppDataPath() +var configFolder = path.join(appDataPath, "FreeDATA"); +var configPath = path.join(configFolder, "config.json"); + +console.log('AppData Path:', appDataPath); +console.log(configFolder); +console.log(configPath); + +nconf.file({ file: configPath }); // +++ //GUI DEFAULT SETTINGS........ diff --git a/modem/server.py b/modem/server.py index de16a470..3a5e0971 100644 --- a/modem/server.py +++ b/modem/server.py @@ -22,7 +22,7 @@ app = Flask(__name__) CORS(app) CORS(app, resources={r"/*": {"origins": "*"}}) sock = Sock(app) -MODEM_VERSION = "0.12.0-alpha" +MODEM_VERSION = "0.12.1-alpha" # set config file to use def set_config(): From 98f48295d08e16263b9c84fd5e5dd4c9001e474d Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Wed, 17 Jan 2024 19:48:14 +0100 Subject: [PATCH 04/15] removed hamlib response --- modem/rigctld.py | 1 - 1 file changed, 1 deletion(-) diff --git a/modem/rigctld.py b/modem/rigctld.py index b4e5eda7..64996794 100644 --- a/modem/rigctld.py +++ b/modem/rigctld.py @@ -190,7 +190,6 @@ class radio: try: mode, bandwidth = response.split('\n', 1) # Split the response into mode and bandwidth except ValueError: - print(response) mode = 'err' bandwidth = 'err' From 552a5a9ac8cde78e3b58320746ea01ff75c44c86 Mon Sep 17 00:00:00 2001 From: codefactor-io Date: Wed, 17 Jan 2024 18:49:14 +0000 Subject: [PATCH 05/15] [CodeFactor] Apply fixes --- gui/src/js/chatHandler.ts | 4 ++-- gui/src/js/freedata.ts | 42 +++++++++++++++++----------------- gui/src/store/settingsStore.js | 8 +++---- 3 files changed, 27 insertions(+), 27 deletions(-) diff --git a/gui/src/js/chatHandler.ts b/gui/src/js/chatHandler.ts index 4265abd0..50c664ba 100644 --- a/gui/src/js/chatHandler.ts +++ b/gui/src/js/chatHandler.ts @@ -14,7 +14,7 @@ import { useStateStore } from "../store/stateStore.js"; const state = useStateStore(pinia); import { settingsStore as settings } from "../store/settingsStore.js"; -import {getAppDataPath} from "../js/freedata"; +import { getAppDataPath } from "../js/freedata"; import { displayToast } from "./popupHandler.js"; @@ -100,7 +100,7 @@ PouchDB.plugin(require("pouchdb-find")); //PouchDB.plugin(require('pouchdb-replication')); PouchDB.plugin(require("pouchdb-upsert")); -var appDataPath = getAppDataPath() +var appDataPath = getAppDataPath(); var configFolder = path.join(appDataPath, "FreeDATA"); var chatDB = path.join(configFolder, "chatDB"); diff --git a/gui/src/js/freedata.ts b/gui/src/js/freedata.ts index a42d0edb..5d9158f9 100644 --- a/gui/src/js/freedata.ts +++ b/gui/src/js/freedata.ts @@ -1,5 +1,5 @@ -const os = require('os'); -const path = require('path'); +const os = require("os"); +const path = require("path"); /** * Binary to ASCII replacement @@ -101,24 +101,24 @@ export function validateCallsignWithoutSSID(callsign: string) { return true; } -export function getAppDataPath(){ - const platform = os.platform(); - let appDataPath; +export function getAppDataPath() { + const platform = os.platform(); + let appDataPath; - switch (platform) { - case 'darwin': // macOS - appDataPath = path.join(os.homedir(), 'Library', 'Application Support'); - break; - case 'win32': // Windows - appDataPath = process.env.APPDATA || path.join(os.homedir(), 'AppData', 'Roaming'); - break; - case 'linux': // Linux - appDataPath = path.join(os.homedir(), '.config'); - break; - default: - throw new Error('Unsupported platform'); - } + switch (platform) { + case "darwin": // macOS + appDataPath = path.join(os.homedir(), "Library", "Application Support"); + break; + case "win32": // Windows + appDataPath = + process.env.APPDATA || path.join(os.homedir(), "AppData", "Roaming"); + break; + case "linux": // Linux + appDataPath = path.join(os.homedir(), ".config"); + break; + default: + throw new Error("Unsupported platform"); + } - return appDataPath; - -} \ No newline at end of file + return appDataPath; +} diff --git a/gui/src/store/settingsStore.js b/gui/src/store/settingsStore.js index a3ae3040..421853cf 100644 --- a/gui/src/store/settingsStore.js +++ b/gui/src/store/settingsStore.js @@ -1,15 +1,15 @@ import { reactive, ref, watch } from "vue"; import { getConfig, setConfig } from "../js/api"; -import {getAppDataPath} from "../js/freedata"; +import { getAppDataPath } from "../js/freedata"; import fs from "fs"; -const path = require('path'); +const path = require("path"); const nconf = require("nconf"); -var appDataPath = getAppDataPath() +var appDataPath = getAppDataPath(); var configFolder = path.join(appDataPath, "FreeDATA"); var configPath = path.join(configFolder, "config.json"); -console.log('AppData Path:', appDataPath); +console.log("AppData Path:", appDataPath); console.log(configFolder); console.log(configPath); From 8d81e89d098254edd9a15fb3ba6cd8f92e2eb92c Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Fri, 19 Jan 2024 22:28:45 +0000 Subject: [PATCH 06/15] Bump vite from 5.0.10 to 5.0.12 in /gui Bumps [vite](https://github.com/vitejs/vite/tree/HEAD/packages/vite) from 5.0.10 to 5.0.12. - [Release notes](https://github.com/vitejs/vite/releases) - [Changelog](https://github.com/vitejs/vite/blob/v5.0.12/packages/vite/CHANGELOG.md) - [Commits](https://github.com/vitejs/vite/commits/v5.0.12/packages/vite) --- updated-dependencies: - dependency-name: vite dependency-type: direct:development ... Signed-off-by: dependabot[bot] --- gui/package.json | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gui/package.json b/gui/package.json index 0517b2ae..e3ff3deb 100644 --- a/gui/package.json +++ b/gui/package.json @@ -80,7 +80,7 @@ "eslint-plugin-promise": "6.1.1", "eslint-plugin-vue": "9.20.1", "typescript": "5.3.3", - "vite": "5.0.10", + "vite": "5.0.12", "vite-plugin-electron": "0.28.0", "vite-plugin-electron-renderer": "0.14.5", "vitest": "1.0.2", From a31fce3301d0ebe47b00b221cab8bd6ada4e0720 Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Sat, 20 Jan 2024 13:52:35 +0100 Subject: [PATCH 07/15] work on data dispatcher --- modem/arq_session_irs.py | 4 ++- modem/data_dispatcher.py | 52 +++++++++++++++++++++++++++++++++++ tests/test_data_dispatcher.py | 33 ++++++++++++++++++++++ 3 files changed, 88 insertions(+), 1 deletion(-) create mode 100644 modem/data_dispatcher.py create mode 100644 tests/test_data_dispatcher.py diff --git a/modem/arq_session_irs.py b/modem/arq_session_irs.py index 4b52a959..cbdb02e1 100644 --- a/modem/arq_session_irs.py +++ b/modem/arq_session_irs.py @@ -5,6 +5,8 @@ from modem_frametypes import FRAME_TYPE from codec2 import FREEDV_MODE from enum import Enum import time +from data_dispatcher import DataDispatcher + class IRS_State(Enum): NEW = 0 OPEN_ACK_SENT = 1 @@ -191,7 +193,7 @@ class ARQSessionIRS(arq_session.ARQSession): self.set_state(IRS_State.ENDED) self.event_manager.send_arq_session_finished( False, self.id, self.dxcall, True, self.state.name, data=self.received_data, statistics=self.calculate_session_statistics()) - + DataDispatcher().dispatch(self.received_data) else: ack = self.frame_factory.build_arq_burst_ack(self.id, diff --git a/modem/data_dispatcher.py b/modem/data_dispatcher.py new file mode 100644 index 00000000..43b87a4c --- /dev/null +++ b/modem/data_dispatcher.py @@ -0,0 +1,52 @@ +import json +import structlog +class DataDispatcher: + def __init__(self): + self.logger = structlog.get_logger(type(self).__name__) + + # Hardcoded endpoints + self.endpoints = { + "p2pmsg": self.handle_p2pmsg, + "test": self.handle_test, + } + self.default_handler = self.handle_raw # Default handler for unrecognized types + + def log(self, message, isWarning = False): + msg = f"[{type(self).__name__}]: {message}" + logger = self.logger.warn if isWarning else self.logger.info + logger(msg) + + def encapsulate(self, data, type_key="p2pmsg"): + """Encapsulate data into the specified format with the given type key.""" + formatted_data = {type_key: data} + return json.dumps(formatted_data) + + def decapsulate(self, byte_data): + """Decapsulate data from the specified format, returning both the data and the type.""" + try: + json_data = byte_data.decode('utf-8') # Decode byte array to string + parsed_data = json.loads(json_data) + if parsed_data and isinstance(parsed_data, dict): + for key, value in parsed_data.items(): + return key, value # Return type and data + return "raw", byte_data # Treat as raw data if no matching type is found + except (json.JSONDecodeError, UnicodeDecodeError): + return "raw", byte_data # Return original data as raw if there's an error + + def dispatch(self, byte_data): + """Decapsulate and dispatch data to the appropriate endpoint based on its type.""" + type_key, data = self.decapsulate(byte_data) + if type_key in self.endpoints: + self.endpoints[type_key](data) + else: + # Use the default handler for unrecognized types + self.default_handler(data) + + def handle_p2pmsg(self, data): + self.log(f"Handling p2pmsg: {data}") + + def handle_raw(self, data): + self.log(f"Handling raw data: {data}") + + def handle_test(self, data): + self.log(f"Handling test data: {data}") \ No newline at end of file diff --git a/tests/test_data_dispatcher.py b/tests/test_data_dispatcher.py new file mode 100644 index 00000000..6b30c8e8 --- /dev/null +++ b/tests/test_data_dispatcher.py @@ -0,0 +1,33 @@ +import sys +sys.path.append('modem') + +import unittest +from data_dispatcher import DataDispatcher + +class TestDispatcher(unittest.TestCase): + + @classmethod + def setUpClass(cls): + cls.data_dispatcher = DataDispatcher() + + + def testEncapsulator(self): + message_type = "p2pmsg" + message_data = {"message": "Hello, P2P World!"} + + encapsulated = self.data_dispatcher.encapsulate(message_data, message_type) + type, decapsulated = self.data_dispatcher.decapsulate(encapsulated.encode('utf-8')) + self.assertEqual(type, message_type) + self.assertEqual(decapsulated, message_data) + + def testDispatcher(self): + message_type = "test" + message_data = {"message": "Hello, P2P World!"} + + encapsulated = self.data_dispatcher.encapsulate(message_data, message_type) + self.data_dispatcher.dispatch(encapsulated.encode('utf-8')) + + + +if __name__ == '__main__': + unittest.main() From 47363b2521e90494d507d503c256ea9dc1375fd9 Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Sat, 20 Jan 2024 14:35:04 +0100 Subject: [PATCH 08/15] small adjustments --- modem/data_dispatcher.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/modem/data_dispatcher.py b/modem/data_dispatcher.py index 43b87a4c..f3593f05 100644 --- a/modem/data_dispatcher.py +++ b/modem/data_dispatcher.py @@ -4,12 +4,11 @@ class DataDispatcher: def __init__(self): self.logger = structlog.get_logger(type(self).__name__) - # Hardcoded endpoints + # endpoints self.endpoints = { "p2pmsg": self.handle_p2pmsg, "test": self.handle_test, } - self.default_handler = self.handle_raw # Default handler for unrecognized types def log(self, message, isWarning = False): msg = f"[{type(self).__name__}]: {message}" @@ -40,7 +39,7 @@ class DataDispatcher: self.endpoints[type_key](data) else: # Use the default handler for unrecognized types - self.default_handler(data) + self.handle_raw(data) def handle_p2pmsg(self, data): self.log(f"Handling p2pmsg: {data}") From 26478ef0a4d46acf00c2f182e79d7c0af9e8cec6 Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Sat, 20 Jan 2024 21:47:21 +0100 Subject: [PATCH 09/15] adjusted and splitted dispatcher --- modem/arq_data_formatter.py | 24 +++++++++++++ modem/arq_received_data_dispatcher.py | 35 ++++++++++++++++++ modem/arq_session.py | 10 ++++-- modem/arq_session_irs.py | 5 ++- modem/data_dispatcher.py | 51 --------------------------- tests/test_data_dispatcher.py | 12 ++++--- 6 files changed, 76 insertions(+), 61 deletions(-) create mode 100644 modem/arq_data_formatter.py create mode 100644 modem/arq_received_data_dispatcher.py delete mode 100644 modem/data_dispatcher.py diff --git a/modem/arq_data_formatter.py b/modem/arq_data_formatter.py new file mode 100644 index 00000000..d81b1fd6 --- /dev/null +++ b/modem/arq_data_formatter.py @@ -0,0 +1,24 @@ +# File: arq_data_formatter.py + +import json + +class ARQDataFormatter: + def __init__(self): + pass + + def encapsulate(self, data, type_key="p2pmsg"): + """Encapsulate data into the specified format with the given type key.""" + formatted_data = {type_key: data} + return json.dumps(formatted_data) + + def decapsulate(self, byte_data): + """Decapsulate data from the specified format, returning both the data and the type.""" + try: + json_data = byte_data.decode('utf-8') # Decode byte array to string + parsed_data = json.loads(json_data) + if parsed_data and isinstance(parsed_data, dict): + for key, value in parsed_data.items(): + return key, value + return "raw", byte_data + except (json.JSONDecodeError, UnicodeDecodeError): + return "raw", byte_data diff --git a/modem/arq_received_data_dispatcher.py b/modem/arq_received_data_dispatcher.py new file mode 100644 index 00000000..b8572841 --- /dev/null +++ b/modem/arq_received_data_dispatcher.py @@ -0,0 +1,35 @@ +# File: arq_received_data_dispatcher.py + +import structlog +from arq_data_formatter import ARQDataFormatter + +class ARQReceivedDataDispatcher: + def __init__(self): + self.logger = structlog.get_logger(type(self).__name__) + self.arq_data_formatter = ARQDataFormatter() + self.endpoints = { + "p2pmsg": self.handle_p2pmsg, + "test": self.handle_test, + } + + def log(self, message, isWarning=False): + msg = f"[{type(self).__name__}]: {message}" + logger = self.logger.warn if isWarning else self.logger.info + logger(msg) + + def dispatch(self, byte_data): + """Use the data formatter to decapsulate and then dispatch data to the appropriate endpoint.""" + type_key, data = self.arq_data_formatter.decapsulate(byte_data) + if type_key in self.endpoints: + self.endpoints[type_key](data) + else: + self.handle_raw(data) + + def handle_p2pmsg(self, data): + self.log(f"Handling p2pmsg: {data}") + + def handle_raw(self, data): + self.log(f"Handling raw data: {data}") + + def handle_test(self, data): + self.log(f"Handling test data: {data}") diff --git a/modem/arq_session.py b/modem/arq_session.py index 9df7cc68..71ef28e3 100644 --- a/modem/arq_session.py +++ b/modem/arq_session.py @@ -5,6 +5,8 @@ import structlog from event_manager import EventManager from modem_frametypes import FRAME_TYPE import time +from arq_received_data_dispatcher import ARQReceivedDataDispatcher + class ARQSession(): @@ -44,6 +46,7 @@ class ARQSession(): self.frame_factory = data_frame_factory.DataFrameFactory(self.config) self.event_frame_received = threading.Event() + self.arq_received_data_dispatcher = ARQReceivedDataDispatcher() self.id = None self.session_started = time.time() self.session_ended = 0 @@ -88,10 +91,13 @@ class ARQSession(): if self.state in self.STATE_TRANSITION: if frame_type in self.STATE_TRANSITION[self.state]: action_name = self.STATE_TRANSITION[self.state][frame_type] - getattr(self, action_name)(frame) + received_data = getattr(self, action_name)(frame) + if received_data: + self.arq_received_data_dispatcher.dispatch(received_data) + return - self.log(f"Ignoring unknow transition from state {self.state.name} with frame {frame['frame_type']}") + self.log(f"Ignoring unknown transition from state {self.state.name} with frame {frame['frame_type']}") def is_session_outdated(self): session_alivetime = time.time() - self.session_max_age diff --git a/modem/arq_session_irs.py b/modem/arq_session_irs.py index cbdb02e1..7eb7f821 100644 --- a/modem/arq_session_irs.py +++ b/modem/arq_session_irs.py @@ -5,7 +5,6 @@ from modem_frametypes import FRAME_TYPE from codec2 import FREEDV_MODE from enum import Enum import time -from data_dispatcher import DataDispatcher class IRS_State(Enum): NEW = 0 @@ -193,7 +192,7 @@ class ARQSessionIRS(arq_session.ARQSession): self.set_state(IRS_State.ENDED) self.event_manager.send_arq_session_finished( False, self.id, self.dxcall, True, self.state.name, data=self.received_data, statistics=self.calculate_session_statistics()) - DataDispatcher().dispatch(self.received_data) + return self.received_data else: ack = self.frame_factory.build_arq_burst_ack(self.id, @@ -209,7 +208,7 @@ class ARQSessionIRS(arq_session.ARQSession): self.set_state(IRS_State.FAILED) self.event_manager.send_arq_session_finished( False, self.id, self.dxcall, False, self.state.name, statistics=self.calculate_session_statistics()) - + return False def calibrate_speed_settings(self): self.speed_level = 0 # for now stay at lowest speed level diff --git a/modem/data_dispatcher.py b/modem/data_dispatcher.py deleted file mode 100644 index f3593f05..00000000 --- a/modem/data_dispatcher.py +++ /dev/null @@ -1,51 +0,0 @@ -import json -import structlog -class DataDispatcher: - def __init__(self): - self.logger = structlog.get_logger(type(self).__name__) - - # endpoints - self.endpoints = { - "p2pmsg": self.handle_p2pmsg, - "test": self.handle_test, - } - - def log(self, message, isWarning = False): - msg = f"[{type(self).__name__}]: {message}" - logger = self.logger.warn if isWarning else self.logger.info - logger(msg) - - def encapsulate(self, data, type_key="p2pmsg"): - """Encapsulate data into the specified format with the given type key.""" - formatted_data = {type_key: data} - return json.dumps(formatted_data) - - def decapsulate(self, byte_data): - """Decapsulate data from the specified format, returning both the data and the type.""" - try: - json_data = byte_data.decode('utf-8') # Decode byte array to string - parsed_data = json.loads(json_data) - if parsed_data and isinstance(parsed_data, dict): - for key, value in parsed_data.items(): - return key, value # Return type and data - return "raw", byte_data # Treat as raw data if no matching type is found - except (json.JSONDecodeError, UnicodeDecodeError): - return "raw", byte_data # Return original data as raw if there's an error - - def dispatch(self, byte_data): - """Decapsulate and dispatch data to the appropriate endpoint based on its type.""" - type_key, data = self.decapsulate(byte_data) - if type_key in self.endpoints: - self.endpoints[type_key](data) - else: - # Use the default handler for unrecognized types - self.handle_raw(data) - - def handle_p2pmsg(self, data): - self.log(f"Handling p2pmsg: {data}") - - def handle_raw(self, data): - self.log(f"Handling raw data: {data}") - - def handle_test(self, data): - self.log(f"Handling test data: {data}") \ No newline at end of file diff --git a/tests/test_data_dispatcher.py b/tests/test_data_dispatcher.py index 6b30c8e8..90b64fa9 100644 --- a/tests/test_data_dispatcher.py +++ b/tests/test_data_dispatcher.py @@ -2,21 +2,23 @@ import sys sys.path.append('modem') import unittest -from data_dispatcher import DataDispatcher +from arq_data_formatter import ARQDataFormatter +from arq_received_data_dispatcher import ARQReceivedDataDispatcher class TestDispatcher(unittest.TestCase): @classmethod def setUpClass(cls): - cls.data_dispatcher = DataDispatcher() + cls.data_dispatcher = ARQReceivedDataDispatcher() + cls.data_formatter = ARQDataFormatter() def testEncapsulator(self): message_type = "p2pmsg" message_data = {"message": "Hello, P2P World!"} - encapsulated = self.data_dispatcher.encapsulate(message_data, message_type) - type, decapsulated = self.data_dispatcher.decapsulate(encapsulated.encode('utf-8')) + encapsulated = self.data_formatter.encapsulate(message_data, message_type) + type, decapsulated = self.data_formatter.decapsulate(encapsulated.encode('utf-8')) self.assertEqual(type, message_type) self.assertEqual(decapsulated, message_data) @@ -24,7 +26,7 @@ class TestDispatcher(unittest.TestCase): message_type = "test" message_data = {"message": "Hello, P2P World!"} - encapsulated = self.data_dispatcher.encapsulate(message_data, message_type) + encapsulated = self.data_formatter.encapsulate(message_data, message_type) self.data_dispatcher.dispatch(encapsulated.encode('utf-8')) From 857916285d399ec822e351910cf26389db245917 Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Sun, 21 Jan 2024 20:34:01 +0100 Subject: [PATCH 10/15] changed dispatcher to a data type handler --- modem/arq_data_type_handler.py | 83 +++++++++++++++++++++++++++ modem/arq_received_data_dispatcher.py | 35 ----------- modem/arq_session.py | 10 ++-- modem/arq_session_irs.py | 15 +++-- modem/arq_session_iss.py | 18 ++++-- modem/command.py | 3 + modem/command_arq_raw.py | 9 ++- modem/data_frame_factory.py | 10 ++-- modem/frame_handler.py | 1 - tests/test_arq_session.py | 4 +- tests/test_data_dispatcher.py | 35 ----------- tests/test_data_type_handler.py | 37 ++++++++++++ 12 files changed, 165 insertions(+), 95 deletions(-) create mode 100644 modem/arq_data_type_handler.py delete mode 100644 modem/arq_received_data_dispatcher.py delete mode 100644 tests/test_data_dispatcher.py create mode 100644 tests/test_data_type_handler.py diff --git a/modem/arq_data_type_handler.py b/modem/arq_data_type_handler.py new file mode 100644 index 00000000..8b724844 --- /dev/null +++ b/modem/arq_data_type_handler.py @@ -0,0 +1,83 @@ +# File: arq_data_type_handler.py + +import structlog +import lzma +import gzip + +class ARQDataTypeHandler: + def __init__(self): + self.logger = structlog.get_logger(type(self).__name__) + self.handlers = { + "raw": { + 'prepare': self.prepare_raw, + 'handle': self.handle_raw + }, + "raw_lzma": { + 'prepare': self.prepare_raw_lzma, + 'handle': self.handle_raw_lzma + }, + "raw_gzip": { + 'prepare': self.prepare_raw_gzip, + 'handle': self.handle_raw_gzip + }, + "p2pmsg_lzma": { + 'prepare': self.prepare_p2pmsg_lzma, + 'handle': self.handle_p2pmsg_lzma + }, + } + + def dispatch(self, type_byte: int, data: bytearray): + endpoint_name = list(self.handlers.keys())[type_byte] + if endpoint_name in self.handlers and 'handle' in self.handlers[endpoint_name]: + return self.handlers[endpoint_name]['handle'](data) + else: + self.log(f"Unknown handling endpoint: {endpoint_name}", isWarning=True) + + def prepare(self, data: bytearray, endpoint_name="raw" ): + if endpoint_name in self.handlers and 'prepare' in self.handlers[endpoint_name]: + return self.handlers[endpoint_name]['prepare'](data), list(self.handlers.keys()).index(endpoint_name) + else: + self.log(f"Unknown preparation endpoint: {endpoint_name}", isWarning=True) + + def log(self, message, isWarning=False): + msg = f"[{type(self).__name__}]: {message}" + logger = self.logger.warn if isWarning else self.logger.info + logger(msg) + + def prepare_raw(self, data): + self.log(f"Preparing uncompressed data: {len(data)} Bytes") + return data + + def handle_raw(self, data): + self.log(f"Handling uncompressed data: {len(data)} Bytes") + return data + + def prepare_raw_lzma(self, data): + compressed_data = lzma.compress(data) + self.log(f"Preparing LZMA compressed data: {len(data)} Bytes >>> {len(compressed_data)} Bytes") + return compressed_data + + def handle_raw_lzma(self, data): + decompressed_data = lzma.decompress(data) + self.log(f"Handling LZMA compressed data: {len(decompressed_data)} Bytes from {len(data)} Bytes") + return decompressed_data + + def prepare_raw_gzip(self, data): + compressed_data = gzip.compress(data) + self.log(f"Preparing GZIP compressed data: {len(data)} Bytes >>> {len(compressed_data)} Bytes") + return compressed_data + + def handle_raw_gzip(self, data): + decompressed_data = gzip.decompress(data) + self.log(f"Handling GZIP compressed data: {len(decompressed_data)} Bytes from {len(data)} Bytes") + return decompressed_data + + def prepare_p2pmsg_lzma(self, data): + compressed_data = lzma.compress(data) + self.log(f"Preparing LZMA compressed P2PMSG data: {len(data)} Bytes >>> {len(compressed_data)} Bytes") + return compressed_data + + def handle_p2pmsg_lzma(self, data): + decompressed_data = lzma.decompress(data) + self.log(f"Handling LZMA compressed P2PMSG data: {len(decompressed_data)} Bytes from {len(data)} Bytes") + return decompressed_data diff --git a/modem/arq_received_data_dispatcher.py b/modem/arq_received_data_dispatcher.py deleted file mode 100644 index b8572841..00000000 --- a/modem/arq_received_data_dispatcher.py +++ /dev/null @@ -1,35 +0,0 @@ -# File: arq_received_data_dispatcher.py - -import structlog -from arq_data_formatter import ARQDataFormatter - -class ARQReceivedDataDispatcher: - def __init__(self): - self.logger = structlog.get_logger(type(self).__name__) - self.arq_data_formatter = ARQDataFormatter() - self.endpoints = { - "p2pmsg": self.handle_p2pmsg, - "test": self.handle_test, - } - - def log(self, message, isWarning=False): - msg = f"[{type(self).__name__}]: {message}" - logger = self.logger.warn if isWarning else self.logger.info - logger(msg) - - def dispatch(self, byte_data): - """Use the data formatter to decapsulate and then dispatch data to the appropriate endpoint.""" - type_key, data = self.arq_data_formatter.decapsulate(byte_data) - if type_key in self.endpoints: - self.endpoints[type_key](data) - else: - self.handle_raw(data) - - def handle_p2pmsg(self, data): - self.log(f"Handling p2pmsg: {data}") - - def handle_raw(self, data): - self.log(f"Handling raw data: {data}") - - def handle_test(self, data): - self.log(f"Handling test data: {data}") diff --git a/modem/arq_session.py b/modem/arq_session.py index 71ef28e3..26756d90 100644 --- a/modem/arq_session.py +++ b/modem/arq_session.py @@ -5,7 +5,7 @@ import structlog from event_manager import EventManager from modem_frametypes import FRAME_TYPE import time -from arq_received_data_dispatcher import ARQReceivedDataDispatcher +from arq_data_type_handler import ARQDataTypeHandler class ARQSession(): @@ -46,7 +46,7 @@ class ARQSession(): self.frame_factory = data_frame_factory.DataFrameFactory(self.config) self.event_frame_received = threading.Event() - self.arq_received_data_dispatcher = ARQReceivedDataDispatcher() + self.arq_data_type_handler = ARQDataTypeHandler() self.id = None self.session_started = time.time() self.session_ended = 0 @@ -91,9 +91,9 @@ class ARQSession(): if self.state in self.STATE_TRANSITION: if frame_type in self.STATE_TRANSITION[self.state]: action_name = self.STATE_TRANSITION[self.state][frame_type] - received_data = getattr(self, action_name)(frame) - if received_data: - self.arq_received_data_dispatcher.dispatch(received_data) + received_data, type_byte = getattr(self, action_name)(frame) + if isinstance(received_data, bytearray) and isinstance(type_byte, int): + self.arq_data_type_handler.dispatch(type_byte, received_data) return diff --git a/modem/arq_session_irs.py b/modem/arq_session_irs.py index 7eb7f821..8e0b461f 100644 --- a/modem/arq_session_irs.py +++ b/modem/arq_session_irs.py @@ -69,6 +69,7 @@ class ARQSessionIRS(arq_session.ARQSession): self.state = IRS_State.NEW self.state_enum = IRS_State # needed for access State enum from outside + self.type_byte = None self.total_length = 0 self.total_crc = '' self.received_data = None @@ -115,6 +116,7 @@ class ARQSessionIRS(arq_session.ARQSession): self.launch_transmit_and_wait(ack_frame, self.TIMEOUT_CONNECT, mode=FREEDV_MODE.signalling) if not self.abort: self.set_state(IRS_State.OPEN_ACK_SENT) + return None, None def send_info_ack(self, info_frame): # Get session info from ISS @@ -122,6 +124,7 @@ class ARQSessionIRS(arq_session.ARQSession): self.total_length = info_frame['total_length'] self.total_crc = info_frame['total_crc'] self.dx_snr.append(info_frame['snr']) + self.type_byte = info_frame['type'] self.log(f"New transfer of {self.total_length} bytes") self.event_manager.send_arq_session_new(False, self.id, self.dxcall, self.total_length, self.state.name) @@ -135,7 +138,7 @@ class ARQSessionIRS(arq_session.ARQSession): self.launch_transmit_and_wait(info_ack, self.TIMEOUT_CONNECT, mode=FREEDV_MODE.signalling) if not self.abort: self.set_state(IRS_State.INFO_ACK_SENT) - + return None, None def process_incoming_data(self, frame): if frame['offset'] != self.received_bytes: @@ -175,7 +178,7 @@ class ARQSessionIRS(arq_session.ARQSession): # self.transmitted_acks += 1 self.set_state(IRS_State.BURST_REPLY_SENT) self.launch_transmit_and_wait(ack, self.TIMEOUT_DATA, mode=FREEDV_MODE.signalling) - return + return None, None if self.final_crc_matches(): self.log("All data received successfully!") @@ -192,7 +195,8 @@ class ARQSessionIRS(arq_session.ARQSession): self.set_state(IRS_State.ENDED) self.event_manager.send_arq_session_finished( False, self.id, self.dxcall, True, self.state.name, data=self.received_data, statistics=self.calculate_session_statistics()) - return self.received_data + + return self.received_data, self.type_byte else: ack = self.frame_factory.build_arq_burst_ack(self.id, @@ -208,7 +212,7 @@ class ARQSessionIRS(arq_session.ARQSession): self.set_state(IRS_State.FAILED) self.event_manager.send_arq_session_finished( False, self.id, self.dxcall, False, self.state.name, statistics=self.calculate_session_statistics()) - return False + return False, False def calibrate_speed_settings(self): self.speed_level = 0 # for now stay at lowest speed level @@ -231,4 +235,5 @@ class ARQSessionIRS(arq_session.ARQSession): self.launch_transmit_and_wait(stop_ack, self.TIMEOUT_CONNECT, mode=FREEDV_MODE.signalling) self.set_state(IRS_State.ABORTED) self.event_manager.send_arq_session_finished( - False, self.id, self.dxcall, False, self.state.name, statistics=self.calculate_session_statistics()) \ No newline at end of file + False, self.id, self.dxcall, False, self.state.name, statistics=self.calculate_session_statistics()) + return None, None \ No newline at end of file diff --git a/modem/arq_session_iss.py b/modem/arq_session_iss.py index 5edc47e4..14970262 100644 --- a/modem/arq_session_iss.py +++ b/modem/arq_session_iss.py @@ -53,13 +53,13 @@ class ARQSessionISS(arq_session.ARQSession): } } - def __init__(self, config: dict, modem, dxcall: str, data: bytearray, state_manager): + def __init__(self, config: dict, modem, dxcall: str, state_manager, data: bytearray, type_byte: bytes): super().__init__(config, modem, dxcall) self.state_manager = state_manager self.data = data self.total_length = len(data) self.data_crc = '' - + self.type_byte = type_byte self.confirmed_bytes = 0 self.state = ISS_State.NEW @@ -119,11 +119,13 @@ class ARQSessionISS(arq_session.ARQSession): info_frame = self.frame_factory.build_arq_session_info(self.id, self.total_length, helpers.get_crc_32(self.data), - self.snr[0]) + self.snr[0], self.type_byte) self.launch_twr(info_frame, self.TIMEOUT_CONNECT_ACK, self.RETRIES_CONNECT, mode=FREEDV_MODE.signalling) self.set_state(ISS_State.INFO_SENT) + return None, None + def send_data(self, irs_frame): self.set_speed_and_frames_per_burst(irs_frame) @@ -137,15 +139,15 @@ class ARQSessionISS(arq_session.ARQSession): # check if we received an abort flag if irs_frame["flag"]["ABORT"]: self.transmission_aborted(irs_frame) - return + return None, None if irs_frame["flag"]["FINAL"]: if self.confirmed_bytes == self.total_length and irs_frame["flag"]["CHECKSUM"]: self.transmission_ended(irs_frame) - return + else: self.transmission_failed() - return + return None, None payload_size = self.get_data_payload_size() burst = [] @@ -158,6 +160,7 @@ class ARQSessionISS(arq_session.ARQSession): burst.append(data_frame) self.launch_twr(burst, self.TIMEOUT_TRANSFER, self.RETRIES_CONNECT, mode='auto') self.set_state(ISS_State.BURST_SENT) + return None, None def transmission_ended(self, irs_frame): # final function for sucessfully ended transmissions @@ -166,6 +169,7 @@ class ARQSessionISS(arq_session.ARQSession): self.log(f"All data transfered! flag_final={irs_frame['flag']['FINAL']}, flag_checksum={irs_frame['flag']['CHECKSUM']}") self.event_manager.send_arq_session_finished(True, self.id, self.dxcall,True, self.state.name, statistics=self.calculate_session_statistics()) self.state_manager.remove_arq_iss_session(self.id) + return None, None def transmission_failed(self, irs_frame=None): # final function for failed transmissions @@ -173,6 +177,7 @@ class ARQSessionISS(arq_session.ARQSession): self.set_state(ISS_State.FAILED) self.log(f"Transmission failed!") self.event_manager.send_arq_session_finished(True, self.id, self.dxcall,False, self.state.name, statistics=self.calculate_session_statistics()) + return None, None def abort_transmission(self, irs_frame=None): # function for starting the abort sequence @@ -202,4 +207,5 @@ class ARQSessionISS(arq_session.ARQSession): self.event_manager.send_arq_session_finished( True, self.id, self.dxcall, False, self.state.name, statistics=self.calculate_session_statistics()) self.state_manager.remove_arq_iss_session(self.id) + return None, None diff --git a/modem/command.py b/modem/command.py index 9bcb76f4..331e3fa8 100644 --- a/modem/command.py +++ b/modem/command.py @@ -3,6 +3,8 @@ import queue from codec2 import FREEDV_MODE import structlog from state_manager import StateManager +from arq_data_type_handler import ARQDataTypeHandler + class TxCommand(): @@ -13,6 +15,7 @@ class TxCommand(): self.event_manager = event_manager self.set_params_from_api(apiParams) self.frame_factory = DataFrameFactory(config) + self.arq_data_type_handler = ARQDataTypeHandler() def set_params_from_api(self, apiParams): pass diff --git a/modem/command_arq_raw.py b/modem/command_arq_raw.py index 7544db71..4d640bd0 100644 --- a/modem/command_arq_raw.py +++ b/modem/command_arq_raw.py @@ -13,13 +13,20 @@ class ARQRawCommand(TxCommand): if not api_validations.validate_freedata_callsign(self.dxcall): self.dxcall = f"{self.dxcall}-0" + try: + self.type = apiParams['type'] + except KeyError: + self.type = "raw" + self.data = base64.b64decode(apiParams['data']) def run(self, event_queue: Queue, modem): self.emit_event(event_queue) self.logger.info(self.log_message()) - iss = ARQSessionISS(self.config, modem, self.dxcall, self.data, self.state_manager) + prepared_data, type_byte = self.arq_data_type_handler.prepare(self.data, self.type) + + iss = ARQSessionISS(self.config, modem, self.dxcall, self.state_manager, prepared_data, type_byte) if iss.id: self.state_manager.register_arq_iss_session(iss) iss.start() diff --git a/modem/data_frame_factory.py b/modem/data_frame_factory.py index 29c2f460..b62ba11b 100644 --- a/modem/data_frame_factory.py +++ b/modem/data_frame_factory.py @@ -15,7 +15,6 @@ class DataFrameFactory: 'FINAL': 0, # Bit-position for indicating the FINAL state 'ABORT': 1, # Bit-position for indicating the ABORT request 'CHECKSUM': 2, # Bit-position for indicating the CHECKSUM is correct or not - 'ENABLE_COMPRESSION': 3 # Bit-position for indicating compression is enabled } def __init__(self, config): @@ -118,6 +117,7 @@ class DataFrameFactory: "total_crc": 4, "snr": 1, "flag": 1, + "type": 1, } self.template_list[FR_TYPE.ARQ_SESSION_INFO_ACK.value] = { @@ -218,7 +218,7 @@ class DataFrameFactory: elif key in ["session_id", "speed_level", "frames_per_burst", "version", - "offset", "total_length", "state"]: + "offset", "total_length", "state", "type"]: extracted_data[key] = int.from_bytes(data, 'big') elif key in ["snr"]: @@ -350,10 +350,8 @@ class DataFrameFactory: } return self.construct(FR_TYPE.ARQ_SESSION_OPEN_ACK, payload) - def build_arq_session_info(self, session_id: int, total_length: int, total_crc: bytes, snr, flag_compression=False): + def build_arq_session_info(self, session_id: int, total_length: int, total_crc: bytes, snr, type): flag = 0b00000000 - if flag_compression: - flag = helpers.set_flag(flag, 'ENABLE_COMPRESSION', True, self.ARQ_FLAGS) payload = { "session_id": session_id.to_bytes(1, 'big'), @@ -361,6 +359,7 @@ class DataFrameFactory: "total_crc": total_crc, "snr": helpers.snr_to_bytes(1), "flag": flag.to_bytes(1, 'big'), + "type": type.to_bytes(1, 'big'), } return self.construct(FR_TYPE.ARQ_SESSION_INFO, payload) @@ -377,7 +376,6 @@ class DataFrameFactory: } return self.construct(FR_TYPE.ARQ_STOP_ACK, payload) - def build_arq_session_info_ack(self, session_id, total_crc, snr, speed_level, frames_per_burst, flag_final=False, flag_abort=False): flag = 0b00000000 if flag_final: diff --git a/modem/frame_handler.py b/modem/frame_handler.py index 3d454782..d11ba742 100644 --- a/modem/frame_handler.py +++ b/modem/frame_handler.py @@ -31,7 +31,6 @@ class FrameHandler(): def is_frame_for_me(self): call_with_ssid = self.config['STATION']['mycall'] + "-" + str(self.config['STATION']['myssid']) ft = self.details['frame']['frame_type'] - print(self.details) valid = False # Check for callsign checksum if ft in ['ARQ_SESSION_OPEN', 'ARQ_SESSION_OPEN_ACK', 'PING', 'PING_ACK']: diff --git a/tests/test_arq_session.py b/tests/test_arq_session.py index 4bf66ad9..ecfc0b02 100644 --- a/tests/test_arq_session.py +++ b/tests/test_arq_session.py @@ -126,12 +126,13 @@ class TestARQSession(unittest.TestCase): def testARQSessionSmallPayload(self): # set Packet Error Rate (PER) / frame loss probability - self.loss_probability = 50 + self.loss_probability = 0 self.establishChannels() params = { 'dxcall': "XX1XXX-1", 'data': base64.b64encode(bytes("Hello world!", encoding="utf-8")), + 'type': "raw_lzma" } cmd = ARQRawCommand(self.config, self.iss_state_manager, self.iss_event_queue, params) cmd.run(self.iss_event_queue, self.iss_modem) @@ -146,6 +147,7 @@ class TestARQSession(unittest.TestCase): params = { 'dxcall': "XX1XXX-1", 'data': base64.b64encode(np.random.bytes(1000)), + 'type': "raw_lzma" } cmd = ARQRawCommand(self.config, self.iss_state_manager, self.iss_event_queue, params) cmd.run(self.iss_event_queue, self.iss_modem) diff --git a/tests/test_data_dispatcher.py b/tests/test_data_dispatcher.py deleted file mode 100644 index 90b64fa9..00000000 --- a/tests/test_data_dispatcher.py +++ /dev/null @@ -1,35 +0,0 @@ -import sys -sys.path.append('modem') - -import unittest -from arq_data_formatter import ARQDataFormatter -from arq_received_data_dispatcher import ARQReceivedDataDispatcher - -class TestDispatcher(unittest.TestCase): - - @classmethod - def setUpClass(cls): - cls.data_dispatcher = ARQReceivedDataDispatcher() - cls.data_formatter = ARQDataFormatter() - - - def testEncapsulator(self): - message_type = "p2pmsg" - message_data = {"message": "Hello, P2P World!"} - - encapsulated = self.data_formatter.encapsulate(message_data, message_type) - type, decapsulated = self.data_formatter.decapsulate(encapsulated.encode('utf-8')) - self.assertEqual(type, message_type) - self.assertEqual(decapsulated, message_data) - - def testDispatcher(self): - message_type = "test" - message_data = {"message": "Hello, P2P World!"} - - encapsulated = self.data_formatter.encapsulate(message_data, message_type) - self.data_dispatcher.dispatch(encapsulated.encode('utf-8')) - - - -if __name__ == '__main__': - unittest.main() diff --git a/tests/test_data_type_handler.py b/tests/test_data_type_handler.py new file mode 100644 index 00000000..b7b8cc26 --- /dev/null +++ b/tests/test_data_type_handler.py @@ -0,0 +1,37 @@ +import sys +sys.path.append('modem') + +import unittest +from arq_data_type_handler import ARQDataTypeHandler + +class TestDispatcher(unittest.TestCase): + + @classmethod + def setUpClass(cls): + cls.arq_data_type_handler = ARQDataTypeHandler() + + + def testDataTypeHandlerRaw(self): + # Example usage + example_data = b"Hello FreeDATA!" + formatted_data, type_byte = self.arq_data_type_handler.prepare(example_data, "raw") + dispatched_data = self.arq_data_type_handler.dispatch(type_byte, formatted_data) + self.assertEqual(example_data, dispatched_data) + + def testDataTypeHandlerLZMA(self): + # Example usage + example_data = b"Hello FreeDATA!" + formatted_data, type_byte = self.arq_data_type_handler.prepare(example_data, "raw_lzma") + dispatched_data = self.arq_data_type_handler.dispatch(type_byte, formatted_data) + self.assertEqual(example_data, dispatched_data) + + def testDataTypeHandlerGZIP(self): + # Example usage + example_data = b"Hello FreeDATA!" + formatted_data, type_byte = self.arq_data_type_handler.prepare(example_data, "raw_gzip") + dispatched_data = self.arq_data_type_handler.dispatch(type_byte, formatted_data) + self.assertEqual(example_data, dispatched_data) + + +if __name__ == '__main__': + unittest.main() From f83751cc8086df07f38e7a11f65d068968ce8fc0 Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Tue, 23 Jan 2024 07:32:03 +0100 Subject: [PATCH 11/15] removed data formatter --- modem/arq_data_formatter.py | 24 ------------------------ 1 file changed, 24 deletions(-) delete mode 100644 modem/arq_data_formatter.py diff --git a/modem/arq_data_formatter.py b/modem/arq_data_formatter.py deleted file mode 100644 index d81b1fd6..00000000 --- a/modem/arq_data_formatter.py +++ /dev/null @@ -1,24 +0,0 @@ -# File: arq_data_formatter.py - -import json - -class ARQDataFormatter: - def __init__(self): - pass - - def encapsulate(self, data, type_key="p2pmsg"): - """Encapsulate data into the specified format with the given type key.""" - formatted_data = {type_key: data} - return json.dumps(formatted_data) - - def decapsulate(self, byte_data): - """Decapsulate data from the specified format, returning both the data and the type.""" - try: - json_data = byte_data.decode('utf-8') # Decode byte array to string - parsed_data = json.loads(json_data) - if parsed_data and isinstance(parsed_data, dict): - for key, value in parsed_data.items(): - return key, value - return "raw", byte_data - except (json.JSONDecodeError, UnicodeDecodeError): - return "raw", byte_data From 965dd5e29d5151a51728ff6612f09babbc2f491e Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Tue, 23 Jan 2024 11:39:16 +0100 Subject: [PATCH 12/15] attempt fixing github test --- gui/src/js/freedata.ts | 6 ++++++ gui/src/store/settingsStore.js | 8 +++++++- 2 files changed, 13 insertions(+), 1 deletion(-) diff --git a/gui/src/js/freedata.ts b/gui/src/js/freedata.ts index 5d9158f9..c26154cf 100644 --- a/gui/src/js/freedata.ts +++ b/gui/src/js/freedata.ts @@ -105,6 +105,12 @@ export function getAppDataPath() { const platform = os.platform(); let appDataPath; + // Check if running in GitHub Actions + const isGitHubActions = process.env.GITHUB_ACTIONS === 'true'; + if (isGitHubActions) { + return "/home/runner/work/FreeDATA/FreeDATA/gui/config"; + } + switch (platform) { case "darwin": // macOS appDataPath = path.join(os.homedir(), "Library", "Application Support"); diff --git a/gui/src/store/settingsStore.js b/gui/src/store/settingsStore.js index 421853cf..a73db5b0 100644 --- a/gui/src/store/settingsStore.js +++ b/gui/src/store/settingsStore.js @@ -7,7 +7,13 @@ const nconf = require("nconf"); var appDataPath = getAppDataPath(); var configFolder = path.join(appDataPath, "FreeDATA"); -var configPath = path.join(configFolder, "config.json"); + +let configFile = "config.json" +const isGitHubActions = process.env.GITHUB_ACTIONS === 'true'; +if (isGitHubActions) { + configFile = "example.json"; +} +var configPath = path.join(configFolder, configFile); console.log("AppData Path:", appDataPath); console.log(configFolder); From 9d2332477f757b7e707d5ccd936d1d9600166dfd Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Tue, 23 Jan 2024 11:42:47 +0100 Subject: [PATCH 13/15] attempt fixing github test --- gui/src/store/settingsStore.js | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/gui/src/store/settingsStore.js b/gui/src/store/settingsStore.js index a73db5b0..a73eaabc 100644 --- a/gui/src/store/settingsStore.js +++ b/gui/src/store/settingsStore.js @@ -7,12 +7,14 @@ const nconf = require("nconf"); var appDataPath = getAppDataPath(); var configFolder = path.join(appDataPath, "FreeDATA"); - let configFile = "config.json" + const isGitHubActions = process.env.GITHUB_ACTIONS === 'true'; if (isGitHubActions) { configFile = "example.json"; + configFolder = appDataPath; } + var configPath = path.join(configFolder, configFile); console.log("AppData Path:", appDataPath); From 53a34eaaa27b442285831ef28dc9b4b67ad77780 Mon Sep 17 00:00:00 2001 From: codefactor-io Date: Tue, 23 Jan 2024 10:44:34 +0000 Subject: [PATCH 14/15] [CodeFactor] Apply fixes --- gui/src/js/freedata.ts | 4 ++-- gui/src/store/settingsStore.js | 8 ++++---- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/gui/src/js/freedata.ts b/gui/src/js/freedata.ts index c26154cf..17c816bd 100644 --- a/gui/src/js/freedata.ts +++ b/gui/src/js/freedata.ts @@ -105,8 +105,8 @@ export function getAppDataPath() { const platform = os.platform(); let appDataPath; - // Check if running in GitHub Actions - const isGitHubActions = process.env.GITHUB_ACTIONS === 'true'; + // Check if running in GitHub Actions + const isGitHubActions = process.env.GITHUB_ACTIONS === "true"; if (isGitHubActions) { return "/home/runner/work/FreeDATA/FreeDATA/gui/config"; } diff --git a/gui/src/store/settingsStore.js b/gui/src/store/settingsStore.js index a73eaabc..f7ff0865 100644 --- a/gui/src/store/settingsStore.js +++ b/gui/src/store/settingsStore.js @@ -7,12 +7,12 @@ const nconf = require("nconf"); var appDataPath = getAppDataPath(); var configFolder = path.join(appDataPath, "FreeDATA"); -let configFile = "config.json" +let configFile = "config.json"; -const isGitHubActions = process.env.GITHUB_ACTIONS === 'true'; +const isGitHubActions = process.env.GITHUB_ACTIONS === "true"; if (isGitHubActions) { - configFile = "example.json"; - configFolder = appDataPath; + configFile = "example.json"; + configFolder = appDataPath; } var configPath = path.join(configFolder, configFile); From f4de64d3be199d868369dfd1c140c88dbc8bb0a9 Mon Sep 17 00:00:00 2001 From: DJ2LS Date: Tue, 23 Jan 2024 19:12:04 +0100 Subject: [PATCH 15/15] attempt fixing gui build process --- gui/electron-builder.json5 | 14 -------------- 1 file changed, 14 deletions(-) diff --git a/gui/electron-builder.json5 b/gui/electron-builder.json5 index 5bcf15e6..13f06861 100644 --- a/gui/electron-builder.json5 +++ b/gui/electron-builder.json5 @@ -19,22 +19,8 @@ "files": [ "dist", "dist-electron", - "../modem/server.dist/", ], - "extraResources": [ - - { - "from": "../modem/server.dist/", - "to": "modem", - "filter": [ - "**/*", - "!**/.git" - ] - - } - ], - "mac": { "target": [