- Removed unnecessary blank lines in cx_setup.py. - Reformatted conditional checks in meshchat.py for better readability. - Updated callback variable names in NomadnetDownloader for clarity. - Adjusted version string formatting in version.py for consistency. - Reordered import statements in prepare_frontend_dir.py for better organization.
5439 lines
210 KiB
Python
5439 lines
210 KiB
Python
#!/usr/bin/env python
|
|
|
|
import argparse
|
|
import asyncio
|
|
import base64
|
|
import copy
|
|
import io
|
|
import json
|
|
import os
|
|
import platform
|
|
import sys
|
|
import threading
|
|
import time
|
|
import webbrowser
|
|
from collections.abc import Callable
|
|
from datetime import UTC, datetime
|
|
|
|
import LXMF
|
|
import psutil
|
|
import RNS
|
|
import RNS.vendor.umsgpack as msgpack
|
|
from aiohttp import WSCloseCode, WSMessage, WSMsgType, web
|
|
from LXMF import LXMRouter
|
|
from peewee import SqliteDatabase
|
|
from serial.tools import list_ports
|
|
|
|
from meshchatx import database
|
|
from meshchatx.src.backend.announce_handler import AnnounceHandler
|
|
from meshchatx.src.backend.async_utils import AsyncUtils
|
|
from meshchatx.src.backend.audio_call_manager import AudioCall, AudioCallManager
|
|
from meshchatx.src.backend.colour_utils import ColourUtils
|
|
from meshchatx.src.backend.interface_config_parser import InterfaceConfigParser
|
|
from meshchatx.src.backend.interface_editor import InterfaceEditor
|
|
from meshchatx.src.backend.lxmf_message_fields import (
|
|
LxmfAudioField,
|
|
LxmfFileAttachment,
|
|
LxmfFileAttachmentsField,
|
|
LxmfImageField,
|
|
)
|
|
from meshchatx.src.backend.sideband_commands import SidebandCommands
|
|
from meshchatx.src.version import __version__ as app_version
|
|
|
|
|
|
# NOTE: this is required to be able to pack our app with cxfreeze as an exe, otherwise it can't access bundled assets
|
|
# this returns a file path based on if we are running meshchat.py directly, or if we have packed it as an exe with cxfreeze
|
|
# https://cx-freeze.readthedocs.io/en/latest/faq.html#using-data-files
|
|
# bearer:disable python_lang_path_traversal
|
|
def get_file_path(filename):
|
|
if getattr(sys, "frozen", False):
|
|
datadir = os.path.dirname(sys.executable)
|
|
return os.path.join(datadir, filename)
|
|
|
|
# Assets live inside the meshchatx package when installed from a wheel
|
|
package_dir = os.path.dirname(__file__)
|
|
package_path = os.path.join(package_dir, filename)
|
|
if os.path.exists(package_path):
|
|
return package_path
|
|
|
|
# When running from the repository, fall back to the project root
|
|
repo_root = os.path.dirname(package_dir)
|
|
repo_path = os.path.join(repo_root, filename)
|
|
if os.path.exists(repo_path):
|
|
return repo_path
|
|
|
|
return package_path
|
|
|
|
|
|
class ReticulumMeshChat:
|
|
def __init__(self, identity: RNS.Identity, storage_dir, reticulum_config_dir):
|
|
# when providing a custom storage_dir, files will be saved as
|
|
# <storage_dir>/identities/<identity_hex>/
|
|
# <storage_dir>/identities/<identity_hex>/database.db
|
|
|
|
# if storage_dir is not provided, we will use ./storage instead
|
|
# ./storage/identities/<identity_hex>/
|
|
# ./storage/identities/<identity_hex>/database.db
|
|
|
|
# ensure a storage path exists for the loaded identity
|
|
self.storage_dir = storage_dir or os.path.join("storage")
|
|
self.storage_path = os.path.join(
|
|
self.storage_dir,
|
|
"identities",
|
|
identity.hash.hex(),
|
|
)
|
|
print(f"Using Storage Path: {self.storage_path}")
|
|
os.makedirs(self.storage_path, exist_ok=True)
|
|
|
|
# define path to files based on storage path
|
|
self.database_path = os.path.join(self.storage_path, "database.db")
|
|
lxmf_router_path = os.path.join(self.storage_path, "lxmf_router")
|
|
|
|
# check if database already exists, before initialization
|
|
database_already_exists = os.path.exists(self.database_path)
|
|
|
|
# init database
|
|
sqlite_database = SqliteDatabase(self.database_path)
|
|
database.database.initialize(sqlite_database)
|
|
self.db = database.database
|
|
self.db.connect()
|
|
self.db.create_tables(
|
|
[
|
|
database.Config,
|
|
database.Announce,
|
|
database.CustomDestinationDisplayName,
|
|
database.FavouriteDestination,
|
|
database.LxmfMessage,
|
|
database.LxmfConversationReadState,
|
|
database.LxmfUserIcon,
|
|
database.BlockedDestination,
|
|
database.SpamKeyword,
|
|
],
|
|
)
|
|
|
|
# init config
|
|
self.config = Config()
|
|
|
|
# if database already existed before init, and we don't have a previous version set, we are on version 1
|
|
if database_already_exists and self.config.database_version.get() is None:
|
|
self.config.database_version.set(1)
|
|
|
|
# if database didn't already exist, it was just fully migrated when it was created, so set the current version
|
|
if not database_already_exists:
|
|
self.config.database_version.set(database.latest_version)
|
|
|
|
# migrate database
|
|
current_database_version = self.config.database_version.get()
|
|
migrated_database_version = database.migrate(
|
|
current_version=current_database_version,
|
|
)
|
|
self.config.database_version.set(migrated_database_version)
|
|
|
|
# vacuum database on start to shrink its file size
|
|
sqlite_database.execute_sql("VACUUM")
|
|
|
|
# lxmf messages in outbound or sending state should be marked as failed when app starts as they are no longer being processed
|
|
(
|
|
database.LxmfMessage.update(state="failed")
|
|
.where(database.LxmfMessage.state == "outbound")
|
|
.orwhere(
|
|
(database.LxmfMessage.state == "sent")
|
|
& (database.LxmfMessage.method == "opportunistic"),
|
|
)
|
|
.orwhere(database.LxmfMessage.state == "sending")
|
|
.execute()
|
|
)
|
|
|
|
# init reticulum
|
|
self.reticulum = RNS.Reticulum(reticulum_config_dir)
|
|
self.identity = identity
|
|
|
|
# init lxmf router
|
|
# get propagation node stamp cost from config (only used if running a propagation node)
|
|
propagation_stamp_cost = self.config.lxmf_propagation_node_stamp_cost.get()
|
|
self.message_router = LXMF.LXMRouter(
|
|
identity=self.identity,
|
|
storagepath=lxmf_router_path,
|
|
propagation_cost=propagation_stamp_cost,
|
|
)
|
|
self.message_router.PROCESSING_INTERVAL = 1
|
|
|
|
# increase limit for incoming lxmf messages (received over a resource), to allow receiving larger attachments
|
|
# the lxmf router expects delivery_per_transfer_limit to be provided in kilobytes, so we will do that...
|
|
self.message_router.delivery_per_transfer_limit = (
|
|
self.config.lxmf_delivery_transfer_limit_in_bytes.get() / 1000
|
|
)
|
|
|
|
# register lxmf identity
|
|
inbound_stamp_cost = self.config.lxmf_inbound_stamp_cost.get()
|
|
self.local_lxmf_destination = self.message_router.register_delivery_identity(
|
|
identity=self.identity,
|
|
display_name=self.config.display_name.get(),
|
|
stamp_cost=inbound_stamp_cost,
|
|
)
|
|
|
|
# set a callback for when an lxmf message is received
|
|
self.message_router.register_delivery_callback(self.on_lxmf_delivery)
|
|
|
|
# update active propagation node
|
|
self.set_active_propagation_node(
|
|
self.config.lxmf_preferred_propagation_node_destination_hash.get(),
|
|
)
|
|
|
|
# enable propagation node (we don't call with false if disabled, as no need to announce disabled state every launch)
|
|
if self.config.lxmf_local_propagation_node_enabled.get():
|
|
self.enable_local_propagation_node()
|
|
|
|
# handle received announces based on aspect
|
|
RNS.Transport.register_announce_handler(
|
|
AnnounceHandler("call.audio", self.on_audio_call_announce_received),
|
|
)
|
|
RNS.Transport.register_announce_handler(
|
|
AnnounceHandler("lxmf.delivery", self.on_lxmf_announce_received),
|
|
)
|
|
RNS.Transport.register_announce_handler(
|
|
AnnounceHandler(
|
|
"lxmf.propagation",
|
|
self.on_lxmf_propagation_announce_received,
|
|
),
|
|
)
|
|
RNS.Transport.register_announce_handler(
|
|
AnnounceHandler(
|
|
"nomadnetwork.node",
|
|
self.on_nomadnet_node_announce_received,
|
|
),
|
|
)
|
|
|
|
# remember websocket clients
|
|
self.websocket_clients: list[web.WebSocketResponse] = []
|
|
|
|
# track announce timestamps for rate calculation
|
|
self.announce_timestamps = []
|
|
|
|
# track download speeds for nomadnetwork files (list of tuples: (file_size_bytes, duration_seconds))
|
|
self.download_speeds = []
|
|
|
|
# track active downloads (download_id -> downloader instance)
|
|
self.active_downloads = {}
|
|
self.download_id_counter = 0
|
|
|
|
# register audio call identity
|
|
self.audio_call_manager = AudioCallManager(
|
|
identity=self.identity,
|
|
is_destination_blocked_callback=self.is_destination_blocked,
|
|
)
|
|
self.audio_call_manager.register_incoming_call_callback(
|
|
self.on_incoming_audio_call,
|
|
)
|
|
|
|
# start background thread for auto announce loop
|
|
thread = threading.Thread(target=asyncio.run, args=(self.announce_loop(),))
|
|
thread.daemon = True
|
|
thread.start()
|
|
|
|
# start background thread for auto syncing propagation nodes
|
|
thread = threading.Thread(
|
|
target=asyncio.run,
|
|
args=(self.announce_sync_propagation_nodes(),),
|
|
)
|
|
thread.daemon = True
|
|
thread.start()
|
|
|
|
# gets app version from the synchronized Python version helper
|
|
@staticmethod
|
|
def get_app_version() -> str:
|
|
return app_version
|
|
|
|
# automatically announces based on user config
|
|
async def announce_loop(self):
|
|
while True:
|
|
should_announce = False
|
|
|
|
# check if auto announce is enabled
|
|
if self.config.auto_announce_enabled.get():
|
|
# check if we have announced recently
|
|
last_announced_at = self.config.last_announced_at.get()
|
|
if last_announced_at is not None:
|
|
# determine when next announce should be sent
|
|
auto_announce_interval_seconds = (
|
|
self.config.auto_announce_interval_seconds.get()
|
|
)
|
|
next_announce_at = (
|
|
last_announced_at + auto_announce_interval_seconds
|
|
)
|
|
|
|
# we should announce if current time has passed next announce at timestamp
|
|
if time.time() > next_announce_at:
|
|
should_announce = True
|
|
|
|
else:
|
|
# last announced at is null, so we have never announced, lets do it now
|
|
should_announce = True
|
|
|
|
# announce
|
|
if should_announce:
|
|
await self.announce()
|
|
|
|
# wait 1 second before next loop
|
|
await asyncio.sleep(1)
|
|
|
|
# automatically syncs propagation nodes based on user config
|
|
async def announce_sync_propagation_nodes(self):
|
|
while True:
|
|
should_sync = False
|
|
|
|
# check if auto sync is enabled
|
|
auto_sync_interval_seconds = self.config.lxmf_preferred_propagation_node_auto_sync_interval_seconds.get()
|
|
if auto_sync_interval_seconds > 0:
|
|
# check if we have synced recently
|
|
last_synced_at = (
|
|
self.config.lxmf_preferred_propagation_node_last_synced_at.get()
|
|
)
|
|
if last_synced_at is not None:
|
|
# determine when next sync should happen
|
|
next_sync_at = last_synced_at + auto_sync_interval_seconds
|
|
|
|
# we should sync if current time has passed next sync at timestamp
|
|
if time.time() > next_sync_at:
|
|
should_sync = True
|
|
|
|
else:
|
|
# last synced at is null, so we have never synced, lets do it now
|
|
should_sync = True
|
|
|
|
# sync
|
|
if should_sync:
|
|
await self.sync_propagation_nodes()
|
|
|
|
# wait 1 second before next loop
|
|
await asyncio.sleep(1)
|
|
|
|
# uses the provided destination hash as the active propagation node
|
|
def set_active_propagation_node(self, destination_hash: str | None):
|
|
# set outbound propagation node
|
|
if destination_hash is not None and destination_hash != "":
|
|
try:
|
|
self.message_router.set_outbound_propagation_node(
|
|
bytes.fromhex(destination_hash),
|
|
)
|
|
except Exception:
|
|
# failed to set propagation node, clear it to ensure we don't use an old one by mistake
|
|
self.remove_active_propagation_node()
|
|
|
|
# stop using propagation node
|
|
else:
|
|
self.remove_active_propagation_node()
|
|
|
|
# stops the in progress propagation node sync
|
|
def stop_propagation_node_sync(self):
|
|
self.message_router.cancel_propagation_node_requests()
|
|
|
|
# stops and removes the active propagation node
|
|
def remove_active_propagation_node(self):
|
|
# fixme: it's possible for internal transfer state to get stuck if we change propagation node during a sync
|
|
# this still happens even if we cancel the propagation node requests
|
|
# for now, the user can just manually cancel syncing in the ui if they think it's stuck...
|
|
self.stop_propagation_node_sync()
|
|
self.message_router.outbound_propagation_node = None
|
|
|
|
# enables or disables the local lxmf propagation node
|
|
def enable_local_propagation_node(self, enabled: bool = True):
|
|
try:
|
|
if enabled:
|
|
self.message_router.enable_propagation()
|
|
else:
|
|
self.message_router.disable_propagation()
|
|
except Exception:
|
|
print("failed to enable or disable propagation node")
|
|
|
|
def _get_reticulum_section(self):
|
|
try:
|
|
reticulum_config = self.reticulum.config["reticulum"]
|
|
except Exception:
|
|
reticulum_config = None
|
|
|
|
if not isinstance(reticulum_config, dict):
|
|
reticulum_config = {}
|
|
self.reticulum.config["reticulum"] = reticulum_config
|
|
|
|
return reticulum_config
|
|
|
|
def _get_interfaces_section(self):
|
|
try:
|
|
interfaces = self.reticulum.config["interfaces"]
|
|
except Exception:
|
|
interfaces = None
|
|
|
|
if not isinstance(interfaces, dict):
|
|
interfaces = {}
|
|
self.reticulum.config["interfaces"] = interfaces
|
|
|
|
return interfaces
|
|
|
|
def _get_interfaces_snapshot(self):
|
|
snapshot = {}
|
|
interfaces = self._get_interfaces_section()
|
|
for name, interface in interfaces.items():
|
|
try:
|
|
snapshot[name] = copy.deepcopy(dict(interface))
|
|
except Exception:
|
|
try:
|
|
snapshot[name] = copy.deepcopy(interface)
|
|
except Exception:
|
|
snapshot[name] = {}
|
|
return snapshot
|
|
|
|
def _write_reticulum_config(self):
|
|
try:
|
|
self.reticulum.config.write()
|
|
return True
|
|
except Exception as e:
|
|
print(f"Failed to write Reticulum config: {e}")
|
|
return False
|
|
|
|
def build_user_guidance_messages(self):
|
|
guidance = []
|
|
|
|
interfaces = self._get_interfaces_section()
|
|
if len(interfaces) == 0:
|
|
guidance.append(
|
|
{
|
|
"id": "no_interfaces",
|
|
"title": "No Reticulum interfaces configured",
|
|
"description": "Add at least one Reticulum interface so MeshChat can talk to your radio or transport.",
|
|
"action_route": "/interfaces/add",
|
|
"action_label": "Add Interface",
|
|
"severity": "warning",
|
|
},
|
|
)
|
|
|
|
if not self.reticulum.transport_enabled():
|
|
guidance.append(
|
|
{
|
|
"id": "transport_disabled",
|
|
"title": "Transport mode is disabled",
|
|
"description": "Enable transport to allow MeshChat to relay traffic over your configured interfaces.",
|
|
"action_route": "/settings",
|
|
"action_label": "Open Settings",
|
|
"severity": "info",
|
|
},
|
|
)
|
|
|
|
if not self.config.auto_announce_enabled.get():
|
|
guidance.append(
|
|
{
|
|
"id": "announce_disabled",
|
|
"title": "Auto announcements are turned off",
|
|
"description": "Automatic announces make it easier for other peers to discover you. Enable them if you want to stay visible.",
|
|
"action_route": "/settings",
|
|
"action_label": "Manage Announce Settings",
|
|
"severity": "info",
|
|
},
|
|
)
|
|
|
|
return guidance
|
|
|
|
def _conversation_messages_query(self, destination_hash: str):
|
|
local_hash = self.local_lxmf_destination.hexhash
|
|
return database.LxmfMessage.select().where(
|
|
(
|
|
(database.LxmfMessage.source_hash == local_hash)
|
|
& (database.LxmfMessage.destination_hash == destination_hash)
|
|
)
|
|
| (
|
|
(database.LxmfMessage.destination_hash == local_hash)
|
|
& (database.LxmfMessage.source_hash == destination_hash)
|
|
),
|
|
)
|
|
|
|
def get_conversation_latest_message(self, destination_hash: str):
|
|
return (
|
|
self._conversation_messages_query(destination_hash)
|
|
.order_by(database.LxmfMessage.id.desc())
|
|
.get_or_none()
|
|
)
|
|
|
|
def conversation_has_attachments(self, destination_hash: str):
|
|
query = (
|
|
self._conversation_messages_query(destination_hash)
|
|
.where(
|
|
database.LxmfMessage.fields.contains('"image"')
|
|
| database.LxmfMessage.fields.contains('"audio"')
|
|
| database.LxmfMessage.fields.contains('"file_attachments"'),
|
|
)
|
|
.limit(1)
|
|
)
|
|
return query.exists()
|
|
|
|
@staticmethod
|
|
def message_fields_have_attachments(fields_json: str | None):
|
|
if not fields_json:
|
|
return False
|
|
try:
|
|
fields = json.loads(fields_json)
|
|
except Exception:
|
|
return False
|
|
if "image" in fields or "audio" in fields:
|
|
return True
|
|
if "file_attachments" in fields and isinstance(
|
|
fields["file_attachments"],
|
|
list,
|
|
):
|
|
return len(fields["file_attachments"]) > 0
|
|
return False
|
|
|
|
def search_destination_hashes_by_message(self, search_term: str):
|
|
if search_term is None or search_term.strip() == "":
|
|
return set()
|
|
|
|
local_hash = self.local_lxmf_destination.hexhash
|
|
like_term = f"%{search_term}%"
|
|
|
|
matches = set()
|
|
query = database.LxmfMessage.select(
|
|
database.LxmfMessage.source_hash,
|
|
database.LxmfMessage.destination_hash,
|
|
).where(
|
|
(
|
|
(database.LxmfMessage.source_hash == local_hash)
|
|
| (database.LxmfMessage.destination_hash == local_hash)
|
|
)
|
|
& (
|
|
database.LxmfMessage.title**like_term
|
|
| database.LxmfMessage.content**like_term
|
|
),
|
|
)
|
|
|
|
for message in query:
|
|
if message.source_hash == local_hash:
|
|
matches.add(message.destination_hash)
|
|
else:
|
|
matches.add(message.source_hash)
|
|
|
|
return matches
|
|
|
|
@staticmethod
|
|
def parse_bool_query_param(value: str | None) -> bool:
|
|
if value is None:
|
|
return False
|
|
value = value.lower()
|
|
return value in {"1", "true", "yes", "on"}
|
|
|
|
# handle receiving a new audio call
|
|
def on_incoming_audio_call(self, audio_call: AudioCall):
|
|
print(f"on_incoming_audio_call: {audio_call.link.hash.hex()}")
|
|
AsyncUtils.run_async(
|
|
self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "incoming_audio_call",
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# web server has shutdown, likely ctrl+c, but if we don't do the following, the script never exits
|
|
async def shutdown(self, app):
|
|
# force close websocket clients
|
|
for websocket_client in self.websocket_clients:
|
|
await websocket_client.close(code=WSCloseCode.GOING_AWAY)
|
|
|
|
# stop reticulum
|
|
RNS.Transport.detach_interfaces()
|
|
self.reticulum.exit_handler()
|
|
RNS.exit()
|
|
|
|
def run(self, host, port, launch_browser: bool):
|
|
# create route table
|
|
routes = web.RouteTableDef()
|
|
|
|
# serve index.html
|
|
@routes.get("/")
|
|
async def index(request):
|
|
return web.FileResponse(
|
|
path=get_file_path("public/index.html"),
|
|
headers={
|
|
# don't allow browser to store page in cache, otherwise new app versions may get stale ui
|
|
"Cache-Control": "no-cache, no-store",
|
|
},
|
|
)
|
|
|
|
# serve ping
|
|
@routes.get("/api/v1/status")
|
|
async def status(request):
|
|
return web.json_response(
|
|
{
|
|
"status": "ok",
|
|
},
|
|
)
|
|
|
|
# fetch com ports
|
|
@routes.get("/api/v1/comports")
|
|
async def comports(request):
|
|
comports = []
|
|
for comport in list_ports.comports():
|
|
comports.append(
|
|
{
|
|
"device": comport.device,
|
|
"product": comport.product,
|
|
"serial_number": comport.serial_number,
|
|
},
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"comports": comports,
|
|
},
|
|
)
|
|
|
|
# fetch reticulum interfaces
|
|
@routes.get("/api/v1/reticulum/interfaces")
|
|
async def reticulum_interfaces(request):
|
|
interfaces = self._get_interfaces_snapshot()
|
|
|
|
processed_interfaces = {}
|
|
for interface_name, interface in interfaces.items():
|
|
interface_data = copy.deepcopy(interface)
|
|
|
|
# handle sub-interfaces for RNodeMultiInterface
|
|
if interface_data.get("type") == "RNodeMultiInterface":
|
|
sub_interfaces = []
|
|
for sub_name, sub_config in interface_data.items():
|
|
if sub_name not in {
|
|
"type",
|
|
"port",
|
|
"interface_enabled",
|
|
"selected_interface_mode",
|
|
"configured_bitrate",
|
|
}:
|
|
if isinstance(sub_config, dict):
|
|
sub_config["name"] = sub_name
|
|
sub_interfaces.append(sub_config)
|
|
|
|
# add sub-interfaces to the main interface data
|
|
interface_data["sub_interfaces"] = sub_interfaces
|
|
|
|
for sub in sub_interfaces:
|
|
del interface_data[sub["name"]]
|
|
|
|
processed_interfaces[interface_name] = interface_data
|
|
|
|
return web.json_response(
|
|
{
|
|
"interfaces": processed_interfaces,
|
|
},
|
|
)
|
|
|
|
# enable reticulum interface
|
|
@routes.post("/api/v1/reticulum/interfaces/enable")
|
|
async def reticulum_interfaces_enable(request):
|
|
# get request data
|
|
data = await request.json()
|
|
interface_name = data.get("name")
|
|
|
|
if interface_name is None or interface_name == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface name is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# enable interface
|
|
interfaces = self._get_interfaces_section()
|
|
if interface_name not in interfaces:
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface not found",
|
|
},
|
|
status=404,
|
|
)
|
|
interface = interfaces[interface_name]
|
|
if "enabled" in interface:
|
|
interface["enabled"] = "true"
|
|
if "interface_enabled" in interface:
|
|
interface["interface_enabled"] = "true"
|
|
|
|
keys_to_remove = []
|
|
for key, value in interface.items():
|
|
if value is None:
|
|
keys_to_remove.append(key)
|
|
for key in keys_to_remove:
|
|
del interface[key]
|
|
|
|
# save config
|
|
if not self._write_reticulum_config():
|
|
return web.json_response(
|
|
{
|
|
"message": "Failed to write Reticulum config",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface is now enabled",
|
|
},
|
|
)
|
|
|
|
# disable reticulum interface
|
|
@routes.post("/api/v1/reticulum/interfaces/disable")
|
|
async def reticulum_interfaces_disable(request):
|
|
# get request data
|
|
data = await request.json()
|
|
interface_name = data.get("name")
|
|
|
|
if interface_name is None or interface_name == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface name is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# disable interface
|
|
interfaces = self._get_interfaces_section()
|
|
if interface_name not in interfaces:
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface not found",
|
|
},
|
|
status=404,
|
|
)
|
|
interface = interfaces[interface_name]
|
|
if "enabled" in interface:
|
|
interface["enabled"] = "false"
|
|
if "interface_enabled" in interface:
|
|
interface["interface_enabled"] = "false"
|
|
|
|
keys_to_remove = []
|
|
for key, value in interface.items():
|
|
if value is None:
|
|
keys_to_remove.append(key)
|
|
for key in keys_to_remove:
|
|
del interface[key]
|
|
|
|
# save config
|
|
if not self._write_reticulum_config():
|
|
return web.json_response(
|
|
{
|
|
"message": "Failed to write Reticulum config",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface is now disabled",
|
|
},
|
|
)
|
|
|
|
# delete reticulum interface
|
|
@routes.post("/api/v1/reticulum/interfaces/delete")
|
|
async def reticulum_interfaces_delete(request):
|
|
# get request data
|
|
data = await request.json()
|
|
interface_name = data.get("name")
|
|
|
|
if interface_name is None or interface_name == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface name is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
interfaces = self._get_interfaces_section()
|
|
if interface_name not in interfaces:
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface not found",
|
|
},
|
|
status=404,
|
|
)
|
|
|
|
# delete interface
|
|
del interfaces[interface_name]
|
|
|
|
# save config
|
|
if not self._write_reticulum_config():
|
|
return web.json_response(
|
|
{
|
|
"message": "Failed to write Reticulum config",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface has been deleted",
|
|
},
|
|
)
|
|
|
|
# add reticulum interface
|
|
@routes.post("/api/v1/reticulum/interfaces/add")
|
|
async def reticulum_interfaces_add(request):
|
|
# get request data
|
|
data = await request.json()
|
|
interface_name = data.get("name")
|
|
interface_type = data.get("type")
|
|
allow_overwriting_interface = data.get("allow_overwriting_interface", False)
|
|
|
|
# ensure name is provided
|
|
if interface_name is None or interface_name == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Name is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure type name provided
|
|
if interface_type is None or interface_type == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Type is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# get existing interfaces
|
|
interfaces = self._get_interfaces_section()
|
|
|
|
# ensure name is not for an existing interface, to prevent overwriting
|
|
if allow_overwriting_interface is False and interface_name in interfaces:
|
|
return web.json_response(
|
|
{
|
|
"message": "Name is already in use by another interface",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# get existing interface details if available
|
|
interface_details = {}
|
|
if interface_name in interfaces:
|
|
interface_details = interfaces[interface_name]
|
|
|
|
# update interface details
|
|
interface_details["type"] = interface_type
|
|
|
|
# if interface doesn't have enabled or interface_enabled setting already, enable it by default
|
|
if (
|
|
"enabled" not in interface_details
|
|
and "interface_enabled" not in interface_details
|
|
):
|
|
interface_details["interface_enabled"] = "true"
|
|
|
|
# handle AutoInterface
|
|
if interface_type == "AutoInterface":
|
|
# set optional AutoInterface options
|
|
InterfaceEditor.update_value(interface_details, data, "group_id")
|
|
InterfaceEditor.update_value(
|
|
interface_details,
|
|
data,
|
|
"multicast_address_type",
|
|
)
|
|
InterfaceEditor.update_value(interface_details, data, "devices")
|
|
InterfaceEditor.update_value(interface_details, data, "ignored_devices")
|
|
InterfaceEditor.update_value(interface_details, data, "discovery_scope")
|
|
InterfaceEditor.update_value(interface_details, data, "discovery_port")
|
|
InterfaceEditor.update_value(interface_details, data, "data_port")
|
|
|
|
# handle TCPClientInterface
|
|
if interface_type == "TCPClientInterface":
|
|
# ensure target host provided
|
|
interface_target_host = data.get("target_host")
|
|
if interface_target_host is None or interface_target_host == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Target Host is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure target port provided
|
|
interface_target_port = data.get("target_port")
|
|
if interface_target_port is None or interface_target_port == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Target Port is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# set required TCPClientInterface options
|
|
interface_details["target_host"] = interface_target_host
|
|
interface_details["target_port"] = interface_target_port
|
|
|
|
# set optional TCPClientInterface options
|
|
InterfaceEditor.update_value(interface_details, data, "kiss_framing")
|
|
InterfaceEditor.update_value(interface_details, data, "i2p_tunneled")
|
|
|
|
# handle I2P interface
|
|
if interface_type == "I2PInterface":
|
|
interface_details["connectable"] = "True"
|
|
InterfaceEditor.update_value(interface_details, data, "peers")
|
|
|
|
# handle tcp server interface
|
|
if interface_type == "TCPServerInterface":
|
|
# ensure listen ip provided
|
|
interface_listen_ip = data.get("listen_ip")
|
|
if interface_listen_ip is None or interface_listen_ip == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Listen IP is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure listen port provided
|
|
interface_listen_port = data.get("listen_port")
|
|
if interface_listen_port is None or interface_listen_port == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Listen Port is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# set required TCPServerInterface options
|
|
interface_details["listen_ip"] = interface_listen_ip
|
|
interface_details["listen_port"] = interface_listen_port
|
|
|
|
# set optional TCPServerInterface options
|
|
InterfaceEditor.update_value(interface_details, data, "device")
|
|
InterfaceEditor.update_value(interface_details, data, "prefer_ipv6")
|
|
|
|
# handle udp interface
|
|
if interface_type == "UDPInterface":
|
|
# ensure listen ip provided
|
|
interface_listen_ip = data.get("listen_ip")
|
|
if interface_listen_ip is None or interface_listen_ip == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Listen IP is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure listen port provided
|
|
interface_listen_port = data.get("listen_port")
|
|
if interface_listen_port is None or interface_listen_port == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Listen Port is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure forward ip provided
|
|
interface_forward_ip = data.get("forward_ip")
|
|
if interface_forward_ip is None or interface_forward_ip == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Forward IP is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure forward port provided
|
|
interface_forward_port = data.get("forward_port")
|
|
if interface_forward_port is None or interface_forward_port == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Forward Port is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# set required UDPInterface options
|
|
interface_details["listen_ip"] = interface_listen_ip
|
|
interface_details["listen_port"] = interface_listen_port
|
|
interface_details["forward_ip"] = interface_forward_ip
|
|
interface_details["forward_port"] = interface_forward_port
|
|
|
|
# set optional UDPInterface options
|
|
InterfaceEditor.update_value(interface_details, data, "device")
|
|
|
|
# handle RNodeInterface
|
|
if interface_type == "RNodeInterface":
|
|
# ensure port provided
|
|
interface_port = data.get("port")
|
|
if interface_port is None or interface_port == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Port is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure frequency provided
|
|
interface_frequency = data.get("frequency")
|
|
if interface_frequency is None or interface_frequency == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Frequency is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure bandwidth provided
|
|
interface_bandwidth = data.get("bandwidth")
|
|
if interface_bandwidth is None or interface_bandwidth == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Bandwidth is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure txpower provided
|
|
interface_txpower = data.get("txpower")
|
|
if interface_txpower is None or interface_txpower == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "TX power is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure spreading factor provided
|
|
interface_spreadingfactor = data.get("spreadingfactor")
|
|
if interface_spreadingfactor is None or interface_spreadingfactor == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Spreading Factor is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure coding rate provided
|
|
interface_codingrate = data.get("codingrate")
|
|
if interface_codingrate is None or interface_codingrate == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Coding Rate is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# set required RNodeInterface options
|
|
interface_details["port"] = interface_port
|
|
interface_details["frequency"] = interface_frequency
|
|
interface_details["bandwidth"] = interface_bandwidth
|
|
interface_details["txpower"] = interface_txpower
|
|
interface_details["spreadingfactor"] = interface_spreadingfactor
|
|
interface_details["codingrate"] = interface_codingrate
|
|
|
|
# set optional RNodeInterface options
|
|
InterfaceEditor.update_value(interface_details, data, "callsign")
|
|
InterfaceEditor.update_value(interface_details, data, "id_interval")
|
|
InterfaceEditor.update_value(
|
|
interface_details,
|
|
data,
|
|
"airtime_limit_long",
|
|
)
|
|
InterfaceEditor.update_value(
|
|
interface_details,
|
|
data,
|
|
"airtime_limit_short",
|
|
)
|
|
|
|
# handle RNodeMultiInterface
|
|
if interface_type == "RNodeMultiInterface":
|
|
# required settings
|
|
interface_port = data.get("port")
|
|
sub_interfaces = data.get("sub_interfaces", [])
|
|
|
|
# ensure port provided
|
|
if interface_port is None or interface_port == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Port is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure sub interfaces provided
|
|
if not isinstance(sub_interfaces, list) or not sub_interfaces:
|
|
return web.json_response(
|
|
{
|
|
"message": "At least one sub-interface is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# set required RNodeMultiInterface options
|
|
interface_details["port"] = interface_port
|
|
|
|
# remove any existing sub interfaces, which can be found by finding keys that contain a dict value
|
|
# this allows us to replace all sub interfaces with the ones we are about to add, while also ensuring
|
|
# that we do not remove any existing config values from the main interface config
|
|
for key in list(interface_details.keys()):
|
|
value = interface_details[key]
|
|
if isinstance(value, dict):
|
|
del interface_details[key]
|
|
|
|
# process each provided sub interface
|
|
for idx, sub_interface in enumerate(sub_interfaces):
|
|
# ensure required fields for sub-interface provided
|
|
missing_fields = []
|
|
required_subinterface_fields = [
|
|
"name",
|
|
"frequency",
|
|
"bandwidth",
|
|
"txpower",
|
|
"spreadingfactor",
|
|
"codingrate",
|
|
"vport",
|
|
]
|
|
for field in required_subinterface_fields:
|
|
if (
|
|
field not in sub_interface
|
|
or sub_interface.get(field) is None
|
|
or sub_interface.get(field) == ""
|
|
):
|
|
missing_fields.append(field)
|
|
if missing_fields:
|
|
return web.json_response(
|
|
{
|
|
"message": f"Sub-interface {idx + 1} is missing required field(s): {', '.join(missing_fields)}",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
sub_interface_name = sub_interface.get("name")
|
|
interface_details[sub_interface_name] = {
|
|
"interface_enabled": "true",
|
|
"frequency": int(sub_interface["frequency"]),
|
|
"bandwidth": int(sub_interface["bandwidth"]),
|
|
"txpower": int(sub_interface["txpower"]),
|
|
"spreadingfactor": int(sub_interface["spreadingfactor"]),
|
|
"codingrate": int(sub_interface["codingrate"]),
|
|
"vport": int(sub_interface["vport"]),
|
|
}
|
|
|
|
interfaces[interface_name] = interface_details
|
|
|
|
# handle SerialInterface, KISSInterface, and AX25KISSInterface
|
|
if interface_type in (
|
|
"SerialInterface",
|
|
"KISSInterface",
|
|
"AX25KISSInterface",
|
|
):
|
|
# ensure port provided
|
|
interface_port = data.get("port")
|
|
if interface_port is None or interface_port == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Port is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# set required options
|
|
interface_details["port"] = interface_port
|
|
|
|
# set optional options
|
|
InterfaceEditor.update_value(interface_details, data, "speed")
|
|
InterfaceEditor.update_value(interface_details, data, "databits")
|
|
InterfaceEditor.update_value(interface_details, data, "parity")
|
|
InterfaceEditor.update_value(interface_details, data, "stopbits")
|
|
|
|
# Handle KISS and AX25KISS specific options
|
|
if interface_type in ("KISSInterface", "AX25KISSInterface"):
|
|
# set optional options
|
|
InterfaceEditor.update_value(interface_details, data, "preamble")
|
|
InterfaceEditor.update_value(interface_details, data, "txtail")
|
|
InterfaceEditor.update_value(interface_details, data, "persistence")
|
|
InterfaceEditor.update_value(interface_details, data, "slottime")
|
|
InterfaceEditor.update_value(interface_details, data, "callsign")
|
|
InterfaceEditor.update_value(interface_details, data, "ssid")
|
|
|
|
# FIXME: move to own sections
|
|
# RNode Airtime limits and station ID
|
|
InterfaceEditor.update_value(interface_details, data, "callsign")
|
|
InterfaceEditor.update_value(interface_details, data, "id_interval")
|
|
InterfaceEditor.update_value(interface_details, data, "airtime_limit_long")
|
|
InterfaceEditor.update_value(interface_details, data, "airtime_limit_short")
|
|
|
|
# handle Pipe Interface
|
|
if interface_type == "PipeInterface":
|
|
# ensure command provided
|
|
interface_command = data.get("command")
|
|
if interface_command is None or interface_command == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Command is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# ensure command provided
|
|
interface_respawn_delay = data.get("respawn_delay")
|
|
if interface_respawn_delay is None or interface_respawn_delay == "":
|
|
return web.json_response(
|
|
{
|
|
"message": "Respawn delay is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# set required options
|
|
interface_details["command"] = interface_command
|
|
interface_details["respawn_delay"] = interface_respawn_delay
|
|
|
|
# set common interface options
|
|
InterfaceEditor.update_value(interface_details, data, "bitrate")
|
|
InterfaceEditor.update_value(interface_details, data, "mode")
|
|
InterfaceEditor.update_value(interface_details, data, "network_name")
|
|
InterfaceEditor.update_value(interface_details, data, "passphrase")
|
|
InterfaceEditor.update_value(interface_details, data, "ifac_size")
|
|
|
|
# merge new interface into existing interfaces
|
|
interfaces[interface_name] = interface_details
|
|
# save config
|
|
if not self._write_reticulum_config():
|
|
return web.json_response(
|
|
{
|
|
"message": "Failed to write Reticulum config",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
if allow_overwriting_interface:
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface has been saved. Please restart MeshChat for these changes to take effect.",
|
|
},
|
|
)
|
|
return web.json_response(
|
|
{
|
|
"message": "Interface has been added. Please restart MeshChat for these changes to take effect.",
|
|
},
|
|
)
|
|
|
|
# export interfaces
|
|
@routes.post("/api/v1/reticulum/interfaces/export")
|
|
async def export_interfaces(request):
|
|
try:
|
|
# get request data
|
|
selected_interface_names = None
|
|
try:
|
|
data = await request.json()
|
|
selected_interface_names = data.get("selected_interface_names")
|
|
except Exception:
|
|
# request data was not json, but we don't care
|
|
pass
|
|
|
|
# format interfaces for export
|
|
output = []
|
|
interfaces = self._get_interfaces_snapshot()
|
|
for interface_name, interface in interfaces.items():
|
|
# skip interface if not selected
|
|
if (
|
|
selected_interface_names is not None
|
|
and selected_interface_names != ""
|
|
):
|
|
if interface_name not in selected_interface_names:
|
|
continue
|
|
|
|
# add interface to output
|
|
output.append(f"[[{interface_name}]]")
|
|
for key, value in interface.items():
|
|
if not isinstance(value, dict):
|
|
output.append(f" {key} = {value}")
|
|
output.append("")
|
|
|
|
# Handle sub-interfaces for RNodeMultiInterface
|
|
if interface.get("type") == "RNodeMultiInterface":
|
|
for sub_name, sub_config in interface.items():
|
|
if sub_name in {"type", "port", "interface_enabled"}:
|
|
continue
|
|
if isinstance(sub_config, dict):
|
|
output.append(f" [[[{sub_name}]]]")
|
|
for sub_key, sub_value in sub_config.items():
|
|
output.append(f" {sub_key} = {sub_value}")
|
|
output.append("")
|
|
|
|
return web.Response(
|
|
text="\n".join(output),
|
|
content_type="text/plain",
|
|
headers={
|
|
"Content-Disposition": "attachment; filename=meshchat_interfaces",
|
|
},
|
|
)
|
|
|
|
except Exception as e:
|
|
return web.json_response(
|
|
{
|
|
"message": f"Failed to export interfaces: {e!s}",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
# preview importable interfaces
|
|
@routes.post("/api/v1/reticulum/interfaces/import-preview")
|
|
async def import_interfaces_preview(request):
|
|
try:
|
|
# get request data
|
|
data = await request.json()
|
|
config = data.get("config")
|
|
|
|
# parse interfaces from config
|
|
interfaces = InterfaceConfigParser.parse(config)
|
|
|
|
return web.json_response(
|
|
{
|
|
"interfaces": interfaces,
|
|
},
|
|
)
|
|
|
|
except Exception as e:
|
|
return web.json_response(
|
|
{
|
|
"message": f"Failed to parse config file: {e!s}",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
# import interfaces from config
|
|
@routes.post("/api/v1/reticulum/interfaces/import")
|
|
async def import_interfaces(request):
|
|
try:
|
|
# get request data
|
|
data = await request.json()
|
|
config = data.get("config")
|
|
selected_interface_names = data.get("selected_interface_names")
|
|
|
|
# parse interfaces from config
|
|
interfaces = InterfaceConfigParser.parse(config)
|
|
|
|
# find selected interfaces
|
|
selected_interfaces = []
|
|
for interface in interfaces:
|
|
if interface["name"] in selected_interface_names:
|
|
selected_interfaces.append(interface)
|
|
|
|
# convert interfaces to object
|
|
interface_config = {}
|
|
for interface in selected_interfaces:
|
|
# add interface and keys/values
|
|
interface_name = interface["name"]
|
|
interface_config[interface_name] = {}
|
|
for key, value in interface.items():
|
|
interface_config[interface_name][key] = value
|
|
|
|
# unset name which isn't part of the config
|
|
del interface_config[interface_name]["name"]
|
|
|
|
# force imported interface to be enabled by default
|
|
interface_config[interface_name]["interface_enabled"] = "true"
|
|
|
|
# remove enabled config value in favour of interface_enabled
|
|
if "enabled" in interface_config[interface_name]:
|
|
del interface_config[interface_name]["enabled"]
|
|
|
|
# update reticulum config with new interfaces
|
|
interfaces = self._get_interfaces_section()
|
|
interfaces.update(interface_config)
|
|
if not self._write_reticulum_config():
|
|
return web.json_response(
|
|
{
|
|
"message": "Failed to write Reticulum config",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Interfaces imported successfully",
|
|
},
|
|
)
|
|
|
|
except Exception as e:
|
|
return web.json_response(
|
|
{
|
|
"message": f"Failed to import interfaces: {e!s}",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
# handle websocket clients
|
|
@routes.get("/ws")
|
|
async def ws(request):
|
|
# prepare websocket response
|
|
websocket_response = web.WebSocketResponse(
|
|
# set max message size accepted by server to 50 megabytes
|
|
max_msg_size=50 * 1024 * 1024,
|
|
)
|
|
await websocket_response.prepare(request)
|
|
|
|
# add client to connected clients list
|
|
self.websocket_clients.append(websocket_response)
|
|
|
|
# send config to all clients
|
|
await self.send_config_to_websocket_clients()
|
|
|
|
# handle websocket messages until disconnected
|
|
async for msg in websocket_response:
|
|
msg: WSMessage = msg
|
|
if msg.type == WSMsgType.TEXT:
|
|
try:
|
|
data = json.loads(msg.data)
|
|
await self.on_websocket_data_received(websocket_response, data)
|
|
except Exception as e:
|
|
# ignore errors while handling message
|
|
print("failed to process client message")
|
|
print(e)
|
|
elif msg.type == WSMsgType.ERROR:
|
|
# ignore errors while handling message
|
|
print(f"ws connection error {websocket_response.exception()}")
|
|
|
|
# websocket closed
|
|
self.websocket_clients.remove(websocket_response)
|
|
|
|
return websocket_response
|
|
|
|
# get app info
|
|
@routes.get("/api/v1/app/info")
|
|
async def app_info(request):
|
|
# Get memory usage for current process
|
|
process = psutil.Process()
|
|
memory_info = process.memory_info()
|
|
|
|
# Get network I/O statistics
|
|
net_io = psutil.net_io_counters()
|
|
|
|
# Get total paths
|
|
path_table = self.reticulum.get_path_table()
|
|
total_paths = len(path_table)
|
|
|
|
# Calculate announce rates
|
|
current_time = time.time()
|
|
announces_per_second = len(
|
|
[t for t in self.announce_timestamps if current_time - t <= 1.0],
|
|
)
|
|
announces_per_minute = len(
|
|
[t for t in self.announce_timestamps if current_time - t <= 60.0],
|
|
)
|
|
announces_per_hour = len(
|
|
[t for t in self.announce_timestamps if current_time - t <= 3600.0],
|
|
)
|
|
|
|
# Clean up old announce timestamps (older than 1 hour)
|
|
self.announce_timestamps = [
|
|
t for t in self.announce_timestamps if current_time - t <= 3600.0
|
|
]
|
|
|
|
# Calculate average download speed
|
|
avg_download_speed_bps = None
|
|
if self.download_speeds:
|
|
total_bytes = sum(size for size, _ in self.download_speeds)
|
|
total_duration = sum(duration for _, duration in self.download_speeds)
|
|
if total_duration > 0:
|
|
avg_download_speed_bps = total_bytes / total_duration
|
|
|
|
return web.json_response(
|
|
{
|
|
"app_info": {
|
|
"version": self.get_app_version(),
|
|
"lxmf_version": LXMF.__version__,
|
|
"rns_version": RNS.__version__,
|
|
"python_version": platform.python_version(),
|
|
"storage_path": self.storage_path,
|
|
"database_path": self.database_path,
|
|
"database_file_size": os.path.getsize(self.database_path),
|
|
"reticulum_config_path": self.reticulum.configpath,
|
|
"is_connected_to_shared_instance": self.reticulum.is_connected_to_shared_instance,
|
|
"is_transport_enabled": self.reticulum.transport_enabled(),
|
|
"memory_usage": {
|
|
"rss": memory_info.rss, # Resident Set Size (bytes)
|
|
"vms": memory_info.vms, # Virtual Memory Size (bytes)
|
|
},
|
|
"network_stats": {
|
|
"bytes_sent": net_io.bytes_sent,
|
|
"bytes_recv": net_io.bytes_recv,
|
|
"packets_sent": net_io.packets_sent,
|
|
"packets_recv": net_io.packets_recv,
|
|
},
|
|
"reticulum_stats": {
|
|
"total_paths": total_paths,
|
|
"announces_per_second": announces_per_second,
|
|
"announces_per_minute": announces_per_minute,
|
|
"announces_per_hour": announces_per_hour,
|
|
},
|
|
"download_stats": {
|
|
"avg_download_speed_bps": avg_download_speed_bps,
|
|
},
|
|
"user_guidance": self.build_user_guidance_messages(),
|
|
},
|
|
},
|
|
)
|
|
|
|
# get config
|
|
@routes.get("/api/v1/config")
|
|
async def config_get(request):
|
|
return web.json_response(
|
|
{
|
|
"config": self.get_config_dict(),
|
|
},
|
|
)
|
|
|
|
# update config
|
|
@routes.patch("/api/v1/config")
|
|
async def config_update(request):
|
|
# get request body as json
|
|
data = await request.json()
|
|
|
|
# update config
|
|
await self.update_config(data)
|
|
|
|
return web.json_response(
|
|
{
|
|
"config": self.get_config_dict(),
|
|
},
|
|
)
|
|
|
|
# enable transport mode
|
|
@routes.post("/api/v1/reticulum/enable-transport")
|
|
async def reticulum_enable_transport(request):
|
|
# enable transport mode
|
|
reticulum_config = self._get_reticulum_section()
|
|
reticulum_config["enable_transport"] = True
|
|
if not self._write_reticulum_config():
|
|
return web.json_response(
|
|
{
|
|
"message": "Failed to write Reticulum config",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Transport has been enabled. MeshChat must be restarted for this change to take effect.",
|
|
},
|
|
)
|
|
|
|
# disable transport mode
|
|
@routes.post("/api/v1/reticulum/disable-transport")
|
|
async def reticulum_disable_transport(request):
|
|
# disable transport mode
|
|
reticulum_config = self._get_reticulum_section()
|
|
reticulum_config["enable_transport"] = False
|
|
if not self._write_reticulum_config():
|
|
return web.json_response(
|
|
{
|
|
"message": "Failed to write Reticulum config",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Transport has been disabled. MeshChat must be restarted for this change to take effect.",
|
|
},
|
|
)
|
|
|
|
# get calls
|
|
@routes.get("/api/v1/calls")
|
|
async def calls_get(request):
|
|
# get audio calls
|
|
audio_calls = []
|
|
for audio_call in self.audio_call_manager.audio_calls:
|
|
audio_calls.append(self.convert_audio_call_to_dict(audio_call))
|
|
|
|
return web.json_response(
|
|
{
|
|
"audio_calls": audio_calls,
|
|
},
|
|
)
|
|
|
|
# clear call history
|
|
@routes.post("/api/v1/calls/clear-call-history")
|
|
async def calls_clear_history(request):
|
|
# delete inactive calls, which are classed as call history
|
|
for audio_call in self.audio_call_manager.audio_calls:
|
|
if audio_call.is_active() is False:
|
|
self.audio_call_manager.delete_audio_call(audio_call)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Call history has been cleared",
|
|
},
|
|
)
|
|
|
|
# hangup all calls
|
|
@routes.get("/api/v1/calls/hangup-all")
|
|
async def calls_hangup_all(request):
|
|
self.audio_call_manager.hangup_all()
|
|
return web.json_response(
|
|
{
|
|
"message": "All calls have been hungup",
|
|
},
|
|
)
|
|
|
|
# get call
|
|
@routes.get("/api/v1/calls/{audio_call_link_hash}")
|
|
async def calls_get_by_hash(request):
|
|
# get path params
|
|
audio_call_link_hash = request.match_info.get("audio_call_link_hash", "")
|
|
|
|
# convert hash to bytes
|
|
audio_call_link_hash = bytes.fromhex(audio_call_link_hash)
|
|
|
|
# find audio call
|
|
audio_call = self.audio_call_manager.find_audio_call_by_link_hash(
|
|
audio_call_link_hash,
|
|
)
|
|
if audio_call is None:
|
|
return web.json_response(
|
|
{
|
|
"message": "audio call not found",
|
|
},
|
|
status=404,
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"audio_call": self.convert_audio_call_to_dict(audio_call),
|
|
},
|
|
)
|
|
|
|
# delete call
|
|
@routes.delete("/api/v1/calls/{audio_call_link_hash}")
|
|
async def calls_delete(request):
|
|
# get path params
|
|
audio_call_link_hash = request.match_info.get("audio_call_link_hash", "")
|
|
|
|
# convert hash to bytes
|
|
audio_call_link_hash = bytes.fromhex(audio_call_link_hash)
|
|
|
|
# delete audio call
|
|
self.audio_call_manager.delete_audio_call_by_link_hash(audio_call_link_hash)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "audio call deleted",
|
|
},
|
|
)
|
|
|
|
# initiate a call to the provided destination
|
|
@routes.get("/api/v1/calls/initiate/{destination_hash}")
|
|
async def calls_initiate(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
timeout_seconds = int(request.query.get("timeout", 15))
|
|
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# initiate audio call
|
|
try:
|
|
audio_call = await self.audio_call_manager.initiate(
|
|
destination_hash,
|
|
timeout_seconds,
|
|
)
|
|
return web.json_response(
|
|
{
|
|
"audio_call": self.convert_audio_call_to_dict(audio_call),
|
|
},
|
|
)
|
|
|
|
except Exception as e:
|
|
return web.json_response(
|
|
{
|
|
"message": f"Call Failed: {e!s}",
|
|
},
|
|
status=503,
|
|
)
|
|
|
|
# handle websocket client for sending and receiving audio packets in a call
|
|
@routes.get("/api/v1/calls/{audio_call_link_hash}/audio")
|
|
async def calls_audio_ws(request):
|
|
# get path params
|
|
audio_call_link_hash = request.match_info.get("audio_call_link_hash", "")
|
|
|
|
# convert hash to bytes
|
|
audio_call_link_hash = bytes.fromhex(audio_call_link_hash)
|
|
|
|
# find audio call, this will be null until the link is established
|
|
audio_call = self.audio_call_manager.find_audio_call_by_link_hash(
|
|
audio_call_link_hash,
|
|
)
|
|
if audio_call is None:
|
|
# fixme: web browser expects websocket, so this won't be useful
|
|
return web.json_response(
|
|
{
|
|
"message": "audio call not found",
|
|
},
|
|
status=404,
|
|
)
|
|
|
|
# send audio received from call initiator to call receiver websocket
|
|
def on_audio_packet(data):
|
|
if websocket_response.closed is False:
|
|
try:
|
|
AsyncUtils.run_async(websocket_response.send_bytes(data))
|
|
except Exception:
|
|
# ignore errors sending audio packets to websocket
|
|
pass
|
|
|
|
# close websocket when call is hungup
|
|
def on_hangup():
|
|
if websocket_response.closed is False:
|
|
try:
|
|
AsyncUtils.run_async(
|
|
websocket_response.close(code=WSCloseCode.GOING_AWAY),
|
|
)
|
|
except Exception:
|
|
# ignore errors closing websocket
|
|
pass
|
|
|
|
# register audio packet listener
|
|
audio_call.register_audio_packet_listener(on_audio_packet)
|
|
audio_call.register_hangup_listener(on_hangup)
|
|
|
|
# prepare websocket response
|
|
websocket_response = web.WebSocketResponse()
|
|
await websocket_response.prepare(request)
|
|
|
|
# handle websocket messages until disconnected
|
|
# FIXME: we should send a type with the message, so we can send other data as well
|
|
async for msg in websocket_response:
|
|
msg: WSMessage = msg
|
|
if msg.type == WSMsgType.BINARY:
|
|
try:
|
|
audio_call.send_audio_packet(msg.data)
|
|
except Exception as e:
|
|
# ignore errors while handling message
|
|
print("failed to process client message")
|
|
print(e)
|
|
elif msg.type == WSMsgType.ERROR:
|
|
# ignore errors while handling message
|
|
print(f"ws connection error {websocket_response.exception()}")
|
|
|
|
# unregister audio packet handler now that the websocket has been closed
|
|
audio_call.register_audio_packet_listener(on_audio_packet)
|
|
|
|
return websocket_response
|
|
|
|
# hangup calls
|
|
@routes.get("/api/v1/calls/{audio_call_link_hash}/hangup")
|
|
async def calls_hangup(request):
|
|
# get path params
|
|
audio_call_link_hash = request.match_info.get("audio_call_link_hash", "")
|
|
|
|
# convert hash to bytes
|
|
audio_call_link_hash = bytes.fromhex(audio_call_link_hash)
|
|
|
|
# find audio call
|
|
audio_call = self.audio_call_manager.find_audio_call_by_link_hash(
|
|
audio_call_link_hash,
|
|
)
|
|
if audio_call is None:
|
|
return web.json_response(
|
|
{
|
|
"message": "audio call not found",
|
|
},
|
|
status=404,
|
|
)
|
|
|
|
# hangup the call
|
|
audio_call.hangup()
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Call has been hungup",
|
|
},
|
|
)
|
|
|
|
# announce
|
|
@routes.get("/api/v1/announce")
|
|
async def announce_trigger(request):
|
|
await self.announce()
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "announcing",
|
|
},
|
|
)
|
|
|
|
# serve announces
|
|
@routes.get("/api/v1/announces")
|
|
async def announces_get(request):
|
|
# get query params
|
|
aspect = request.query.get("aspect", None)
|
|
identity_hash = request.query.get("identity_hash", None)
|
|
destination_hash = request.query.get("destination_hash", None)
|
|
limit = request.query.get("limit", None)
|
|
|
|
# build announces database query
|
|
query = database.Announce.select()
|
|
|
|
# filter by provided aspect
|
|
if aspect is not None:
|
|
query = query.where(database.Announce.aspect == aspect)
|
|
|
|
# filter by provided identity hash
|
|
if identity_hash is not None:
|
|
query = query.where(database.Announce.identity_hash == identity_hash)
|
|
|
|
# filter by provided destination hash
|
|
if destination_hash is not None:
|
|
query = query.where(
|
|
database.Announce.destination_hash == destination_hash,
|
|
)
|
|
|
|
# limit results
|
|
if limit is not None:
|
|
query = query.limit(limit)
|
|
|
|
# order announces latest to oldest
|
|
query_results = query.order_by(database.Announce.updated_at.desc())
|
|
|
|
# process announces
|
|
announces = []
|
|
for announce in query_results:
|
|
announces.append(self.convert_db_announce_to_dict(announce))
|
|
|
|
return web.json_response(
|
|
{
|
|
"announces": announces,
|
|
},
|
|
)
|
|
|
|
# serve favourites
|
|
@routes.get("/api/v1/favourites")
|
|
async def favourites_get(request):
|
|
# get query params
|
|
aspect = request.query.get("aspect", None)
|
|
|
|
# build favourites database query
|
|
query = database.FavouriteDestination.select()
|
|
|
|
# filter by provided aspect
|
|
if aspect is not None:
|
|
query = query.where(database.FavouriteDestination.aspect == aspect)
|
|
|
|
# order favourites alphabetically
|
|
query_results = query.order_by(
|
|
database.FavouriteDestination.display_name.asc(),
|
|
)
|
|
|
|
# process favourites
|
|
favourites = []
|
|
for favourite in query_results:
|
|
favourites.append(self.convert_db_favourite_to_dict(favourite))
|
|
|
|
return web.json_response(
|
|
{
|
|
"favourites": favourites,
|
|
},
|
|
)
|
|
|
|
# add favourite
|
|
@routes.post("/api/v1/favourites/add")
|
|
async def favourites_add(request):
|
|
# get request data
|
|
data = await request.json()
|
|
destination_hash = data.get("destination_hash", None)
|
|
display_name = data.get("display_name", None)
|
|
aspect = data.get("aspect", None)
|
|
|
|
# destination hash is required
|
|
if destination_hash is None:
|
|
return web.json_response(
|
|
{
|
|
"message": "destination_hash is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# display name is required
|
|
if display_name is None:
|
|
return web.json_response(
|
|
{
|
|
"message": "display_name is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# aspect is required
|
|
if aspect is None:
|
|
return web.json_response(
|
|
{
|
|
"message": "aspect is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# upsert favourite
|
|
self.db_upsert_favourite(destination_hash, display_name, aspect)
|
|
return web.json_response(
|
|
{
|
|
"message": "Favourite has been added!",
|
|
},
|
|
)
|
|
|
|
# rename favourite
|
|
@routes.post("/api/v1/favourites/{destination_hash}/rename")
|
|
async def favourites_rename(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# get request data
|
|
data = await request.json()
|
|
display_name = data.get("display_name")
|
|
|
|
# update display name if provided
|
|
if len(display_name) > 0:
|
|
database.FavouriteDestination.update(display_name=display_name).where(
|
|
database.FavouriteDestination.destination_hash == destination_hash,
|
|
).execute()
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Favourite has been renamed",
|
|
},
|
|
)
|
|
|
|
# delete favourite
|
|
@routes.delete("/api/v1/favourites/{destination_hash}")
|
|
async def favourites_delete(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# delete favourite
|
|
database.FavouriteDestination.delete().where(
|
|
database.FavouriteDestination.destination_hash == destination_hash,
|
|
).execute()
|
|
return web.json_response(
|
|
{
|
|
"message": "Favourite has been added!",
|
|
},
|
|
)
|
|
|
|
# propagation node status
|
|
@routes.get("/api/v1/lxmf/propagation-node/status")
|
|
async def propagation_node_status(request):
|
|
return web.json_response(
|
|
{
|
|
"propagation_node_status": {
|
|
"state": self.convert_propagation_node_state_to_string(
|
|
self.message_router.propagation_transfer_state,
|
|
),
|
|
"progress": self.message_router.propagation_transfer_progress
|
|
* 100, # convert to percentage
|
|
"messages_received": self.message_router.propagation_transfer_last_result,
|
|
},
|
|
},
|
|
)
|
|
|
|
# sync propagation node
|
|
@routes.get("/api/v1/lxmf/propagation-node/sync")
|
|
async def propagation_node_sync(request):
|
|
# ensure propagation node is configured before attempting to sync
|
|
if self.message_router.get_outbound_propagation_node() is None:
|
|
return web.json_response(
|
|
{
|
|
"message": "A propagation node must be configured to sync messages.",
|
|
},
|
|
status=400,
|
|
)
|
|
|
|
# request messages from propagation node
|
|
await self.sync_propagation_nodes()
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Sync is starting",
|
|
},
|
|
)
|
|
|
|
# stop syncing propagation node
|
|
@routes.get("/api/v1/lxmf/propagation-node/stop-sync")
|
|
async def propagation_node_stop_sync(request):
|
|
self.stop_propagation_node_sync()
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Sync is stopping",
|
|
},
|
|
)
|
|
|
|
# serve propagation nodes
|
|
@routes.get("/api/v1/lxmf/propagation-nodes")
|
|
async def propagation_nodes_get(request):
|
|
# get query params
|
|
limit = request.query.get("limit", None)
|
|
|
|
# get lxmf.propagation announces
|
|
query = database.Announce.select().where(
|
|
database.Announce.aspect == "lxmf.propagation",
|
|
)
|
|
|
|
# limit results
|
|
if limit is not None:
|
|
query = query.limit(limit)
|
|
|
|
# order announces latest to oldest
|
|
query_results = query.order_by(database.Announce.updated_at.desc())
|
|
|
|
# process announces
|
|
lxmf_propagation_nodes = []
|
|
for announce in query_results:
|
|
# find an lxmf.delivery announce for the same identity hash, so we can use that as an "operater by" name
|
|
lxmf_delivery_announce = (
|
|
database.Announce.select()
|
|
.where(database.Announce.aspect == "lxmf.delivery")
|
|
.where(database.Announce.identity_hash == announce.identity_hash)
|
|
.get_or_none()
|
|
)
|
|
|
|
# find a nomadnetwork.node announce for the same identity hash, so we can use that as an "operated by" name
|
|
nomadnetwork_node_announce = (
|
|
database.Announce.select()
|
|
.where(database.Announce.aspect == "nomadnetwork.node")
|
|
.where(database.Announce.identity_hash == announce.identity_hash)
|
|
.get_or_none()
|
|
)
|
|
|
|
# get a display name from other announces belonging to the propagation nodes identity
|
|
operator_display_name = None
|
|
if (
|
|
lxmf_delivery_announce is not None
|
|
and lxmf_delivery_announce.app_data is not None
|
|
):
|
|
operator_display_name = self.parse_lxmf_display_name(
|
|
lxmf_delivery_announce.app_data,
|
|
None,
|
|
)
|
|
elif (
|
|
nomadnetwork_node_announce is not None
|
|
and nomadnetwork_node_announce.app_data is not None
|
|
):
|
|
operator_display_name = (
|
|
ReticulumMeshChat.parse_nomadnetwork_node_display_name(
|
|
nomadnetwork_node_announce.app_data,
|
|
None,
|
|
)
|
|
)
|
|
|
|
# parse app_data so we can see if propagation is enabled or disabled for this node
|
|
is_propagation_enabled = None
|
|
per_transfer_limit = None
|
|
propagation_node_data = (
|
|
ReticulumMeshChat.parse_lxmf_propagation_node_app_data(
|
|
announce.app_data,
|
|
)
|
|
)
|
|
if propagation_node_data is not None:
|
|
is_propagation_enabled = propagation_node_data["enabled"]
|
|
per_transfer_limit = propagation_node_data["per_transfer_limit"]
|
|
|
|
lxmf_propagation_nodes.append(
|
|
{
|
|
"destination_hash": announce.destination_hash,
|
|
"identity_hash": announce.identity_hash,
|
|
"operator_display_name": operator_display_name,
|
|
"is_propagation_enabled": is_propagation_enabled,
|
|
"per_transfer_limit": per_transfer_limit,
|
|
"created_at": announce.created_at,
|
|
"updated_at": announce.updated_at,
|
|
},
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"lxmf_propagation_nodes": lxmf_propagation_nodes,
|
|
},
|
|
)
|
|
|
|
# get path to destination
|
|
@routes.get("/api/v1/destination/{destination_hash}/path")
|
|
async def destination_path(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# check if user wants to request the path from the network right now
|
|
request_query_param = request.query.get("request", "false")
|
|
should_request_now = request_query_param in ("true", "1")
|
|
if should_request_now:
|
|
# determine how long we should wait for a path response
|
|
timeout_seconds = int(request.query.get("timeout", 15))
|
|
timeout_after_seconds = time.time() + timeout_seconds
|
|
|
|
# request path if we don't have it
|
|
if not RNS.Transport.has_path(destination_hash):
|
|
RNS.Transport.request_path(destination_hash)
|
|
|
|
# wait until we have a path, or give up after the configured timeout
|
|
while (
|
|
not RNS.Transport.has_path(destination_hash)
|
|
and time.time() < timeout_after_seconds
|
|
):
|
|
await asyncio.sleep(0.1)
|
|
|
|
# ensure path is known
|
|
if not RNS.Transport.has_path(destination_hash):
|
|
return web.json_response(
|
|
{
|
|
"path": None,
|
|
},
|
|
)
|
|
|
|
# determine next hop and hop count
|
|
hops = RNS.Transport.hops_to(destination_hash)
|
|
next_hop_bytes = self.reticulum.get_next_hop(destination_hash)
|
|
|
|
# ensure next hop provided
|
|
if next_hop_bytes is None:
|
|
return web.json_response(
|
|
{
|
|
"path": None,
|
|
},
|
|
)
|
|
|
|
next_hop = next_hop_bytes.hex()
|
|
next_hop_interface = self.reticulum.get_next_hop_if_name(destination_hash)
|
|
|
|
return web.json_response(
|
|
{
|
|
"path": {
|
|
"hops": hops,
|
|
"next_hop": next_hop,
|
|
"next_hop_interface": next_hop_interface,
|
|
},
|
|
},
|
|
)
|
|
|
|
# drop path to destination
|
|
@routes.post("/api/v1/destination/{destination_hash}/drop-path")
|
|
async def destination_drop_path(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# drop path
|
|
self.reticulum.drop_path(destination_hash)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "Path has been dropped",
|
|
},
|
|
)
|
|
|
|
# get signal metrics for a destination by checking the latest announce or lxmf message received from them
|
|
@routes.get("/api/v1/destination/{destination_hash}/signal-metrics")
|
|
async def destination_signal_metrics(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# signal metrics to return
|
|
snr = None
|
|
rssi = None
|
|
quality = None
|
|
updated_at = None
|
|
|
|
# get latest announce from database for the provided destination hash
|
|
latest_announce = (
|
|
database.Announce.select()
|
|
.where(database.Announce.destination_hash == destination_hash)
|
|
.get_or_none()
|
|
)
|
|
|
|
# get latest lxmf message from database sent to us from the provided destination hash
|
|
latest_lxmf_message = (
|
|
database.LxmfMessage.select()
|
|
.where(
|
|
database.LxmfMessage.destination_hash
|
|
== self.local_lxmf_destination.hexhash,
|
|
)
|
|
.where(database.LxmfMessage.source_hash == destination_hash)
|
|
.order_by(database.LxmfMessage.id.desc())
|
|
.get_or_none()
|
|
)
|
|
|
|
# determine when latest announce was received
|
|
latest_announce_at = None
|
|
if latest_announce is not None:
|
|
latest_announce_at = datetime.fromisoformat(latest_announce.updated_at)
|
|
|
|
# determine when latest lxmf message was received
|
|
latest_lxmf_message_at = None
|
|
if latest_lxmf_message is not None:
|
|
latest_lxmf_message_at = datetime.fromisoformat(
|
|
latest_lxmf_message.created_at,
|
|
)
|
|
|
|
# get signal metrics from latest announce
|
|
if latest_announce is not None:
|
|
snr = latest_announce.snr
|
|
rssi = latest_announce.rssi
|
|
quality = latest_announce.quality
|
|
# using updated_at from announce because this is when the latest announce was received
|
|
updated_at = latest_announce.updated_at
|
|
|
|
# get signal metrics from latest lxmf message if it's more recent than the announce
|
|
if latest_lxmf_message is not None and (
|
|
latest_announce_at is None
|
|
or latest_lxmf_message_at > latest_announce_at
|
|
):
|
|
snr = latest_lxmf_message.snr
|
|
rssi = latest_lxmf_message.rssi
|
|
quality = latest_lxmf_message.quality
|
|
# using created_at from lxmf message because this is when the message was received
|
|
updated_at = latest_lxmf_message.created_at
|
|
|
|
return web.json_response(
|
|
{
|
|
"signal_metrics": {
|
|
"snr": snr,
|
|
"rssi": rssi,
|
|
"quality": quality,
|
|
"updated_at": updated_at,
|
|
},
|
|
},
|
|
)
|
|
|
|
# pings an lxmf.delivery destination by sending empty data and waiting for the recipient to send a proof back
|
|
# the lxmf router proves all received packets, then drops them if they can't be decoded as lxmf messages
|
|
# this allows us to ping/probe any active lxmf.delivery destination and get rtt/snr/rssi data on demand
|
|
# https://github.com/markqvist/LXMF/blob/9ff76c0473e9d4107e079f266dd08144bb74c7c8/LXMF/LXMRouter.py#L234
|
|
# https://github.com/markqvist/LXMF/blob/9ff76c0473e9d4107e079f266dd08144bb74c7c8/LXMF/LXMRouter.py#L1374
|
|
@routes.get("/api/v1/ping/{destination_hash}/lxmf.delivery")
|
|
async def ping_lxmf_delivery(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# determine how long until we should time out
|
|
timeout_seconds = int(request.query.get("timeout", 15))
|
|
timeout_after_seconds = time.time() + timeout_seconds
|
|
|
|
# request path if we don't have it
|
|
if not RNS.Transport.has_path(destination_hash):
|
|
RNS.Transport.request_path(destination_hash)
|
|
|
|
# wait until we have a path, or give up after the configured timeout
|
|
while (
|
|
not RNS.Transport.has_path(destination_hash)
|
|
and time.time() < timeout_after_seconds
|
|
):
|
|
await asyncio.sleep(0.1)
|
|
|
|
# find destination identity
|
|
destination_identity = RNS.Identity.recall(destination_hash)
|
|
if destination_identity is None:
|
|
return web.json_response(
|
|
{
|
|
"message": "Ping failed. Could not find path to destination.",
|
|
},
|
|
status=503,
|
|
)
|
|
|
|
# create outbound destination
|
|
request_destination = RNS.Destination(
|
|
destination_identity,
|
|
RNS.Destination.OUT,
|
|
RNS.Destination.SINGLE,
|
|
"lxmf",
|
|
"delivery",
|
|
)
|
|
|
|
# send empty packet to destination
|
|
packet = RNS.Packet(request_destination, b"")
|
|
receipt = packet.send()
|
|
|
|
# wait until delivered, or give up after time out
|
|
while (
|
|
receipt.status != RNS.PacketReceipt.DELIVERED
|
|
and time.time() < timeout_after_seconds
|
|
):
|
|
await asyncio.sleep(0.1)
|
|
|
|
# ping failed if not delivered
|
|
if receipt.status != RNS.PacketReceipt.DELIVERED:
|
|
return web.json_response(
|
|
{
|
|
"message": f"Ping failed. Timed out after {timeout_seconds} seconds.",
|
|
},
|
|
status=503,
|
|
)
|
|
|
|
# get number of hops to destination and back from destination
|
|
hops_there = RNS.Transport.hops_to(destination_hash)
|
|
hops_back = receipt.proof_packet.hops
|
|
|
|
# get rssi
|
|
rssi = receipt.proof_packet.rssi
|
|
if rssi is None:
|
|
rssi = self.reticulum.get_packet_rssi(receipt.proof_packet.packet_hash)
|
|
|
|
# get snr
|
|
snr = receipt.proof_packet.snr
|
|
if snr is None:
|
|
snr = self.reticulum.get_packet_snr(receipt.proof_packet.packet_hash)
|
|
|
|
# get signal quality
|
|
quality = receipt.proof_packet.q
|
|
if quality is None:
|
|
quality = self.reticulum.get_packet_q(receipt.proof_packet.packet_hash)
|
|
|
|
# get and format round trip time
|
|
rtt = receipt.get_rtt()
|
|
rtt_milliseconds = round(rtt * 1000, 3)
|
|
rtt_duration_string = f"{rtt_milliseconds} ms"
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": f"Valid reply from {receipt.destination.hash.hex()}\nDuration: {rtt_duration_string}\nHops There: {hops_there}\nHops Back: {hops_back}",
|
|
"ping_result": {
|
|
"rtt": rtt,
|
|
"hops_there": hops_there,
|
|
"hops_back": hops_back,
|
|
"rssi": rssi,
|
|
"snr": snr,
|
|
"quality": quality,
|
|
"receiving_interface": str(
|
|
receipt.proof_packet.receiving_interface,
|
|
),
|
|
},
|
|
},
|
|
)
|
|
|
|
# get custom destination display name
|
|
@routes.get("/api/v1/destination/{destination_hash}/custom-display-name")
|
|
async def destination_custom_display_name_get(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
return web.json_response(
|
|
{
|
|
"custom_display_name": self.get_custom_destination_display_name(
|
|
destination_hash,
|
|
),
|
|
},
|
|
)
|
|
|
|
# set custom destination display name
|
|
@routes.post(
|
|
"/api/v1/destination/{destination_hash}/custom-display-name/update",
|
|
)
|
|
async def destination_custom_display_name_update(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# get request data
|
|
data = await request.json()
|
|
display_name = data.get("display_name")
|
|
|
|
# update display name if provided
|
|
if len(display_name) > 0:
|
|
self.db_upsert_custom_destination_display_name(
|
|
destination_hash,
|
|
display_name,
|
|
)
|
|
return web.json_response(
|
|
{
|
|
"message": "Custom display name has been updated",
|
|
},
|
|
)
|
|
|
|
# otherwise remove display name
|
|
database.CustomDestinationDisplayName.delete().where(
|
|
database.CustomDestinationDisplayName.destination_hash
|
|
== destination_hash,
|
|
).execute()
|
|
return web.json_response(
|
|
{
|
|
"message": "Custom display name has been removed",
|
|
},
|
|
)
|
|
|
|
# get lxmf stamp cost for the provided lxmf.delivery destination hash
|
|
@routes.get("/api/v1/destination/{destination_hash}/lxmf-stamp-info")
|
|
async def destination_lxmf_stamp_info(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# get lxmf stamp cost from announce in database
|
|
lxmf_stamp_cost = None
|
|
announce = database.Announce.get_or_none(
|
|
database.Announce.destination_hash == destination_hash.hex(),
|
|
)
|
|
if announce is not None:
|
|
lxmf_stamp_cost = self.parse_lxmf_stamp_cost(announce.app_data)
|
|
|
|
# get outbound ticket expiry for this lxmf destination
|
|
lxmf_outbound_ticket_expiry = (
|
|
self.message_router.get_outbound_ticket_expiry(destination_hash)
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"lxmf_stamp_info": {
|
|
"stamp_cost": lxmf_stamp_cost,
|
|
"outbound_ticket_expiry": lxmf_outbound_ticket_expiry,
|
|
},
|
|
},
|
|
)
|
|
|
|
# get interface stats
|
|
@routes.get("/api/v1/interface-stats")
|
|
async def interface_stats(request):
|
|
# get interface stats
|
|
interface_stats = self.reticulum.get_interface_stats()
|
|
|
|
# ensure transport_id is hex as json_response can't serialize bytes
|
|
if "transport_id" in interface_stats:
|
|
interface_stats["transport_id"] = interface_stats["transport_id"].hex()
|
|
|
|
# ensure probe_responder is hex as json_response can't serialize bytes
|
|
if (
|
|
"probe_responder" in interface_stats
|
|
and interface_stats["probe_responder"] is not None
|
|
):
|
|
interface_stats["probe_responder"] = interface_stats[
|
|
"probe_responder"
|
|
].hex()
|
|
|
|
# ensure ifac_signature is hex as json_response can't serialize bytes
|
|
for interface in interface_stats["interfaces"]:
|
|
if "short_name" in interface:
|
|
interface["interface_name"] = interface["short_name"]
|
|
|
|
if (
|
|
"parent_interface_name" in interface
|
|
and interface["parent_interface_name"] is not None
|
|
):
|
|
interface["parent_interface_hash"] = interface[
|
|
"parent_interface_hash"
|
|
].hex()
|
|
|
|
if interface.get("ifac_signature"):
|
|
interface["ifac_signature"] = interface["ifac_signature"].hex()
|
|
|
|
if interface.get("hash"):
|
|
interface["hash"] = interface["hash"].hex()
|
|
|
|
return web.json_response(
|
|
{
|
|
"interface_stats": interface_stats,
|
|
},
|
|
)
|
|
|
|
# get path table
|
|
@routes.get("/api/v1/path-table")
|
|
async def path_table(request):
|
|
# get path table, making sure hash and via are in hex as json_response can't serialize bytes
|
|
path_table = []
|
|
for path in self.reticulum.get_path_table():
|
|
path["hash"] = path["hash"].hex()
|
|
path["via"] = path["via"].hex()
|
|
path_table.append(path)
|
|
|
|
return web.json_response(
|
|
{
|
|
"path_table": path_table,
|
|
},
|
|
)
|
|
|
|
# send lxmf message
|
|
@routes.post("/api/v1/lxmf-messages/send")
|
|
async def lxmf_messages_send(request):
|
|
# get request body as json
|
|
data = await request.json()
|
|
|
|
# get delivery method
|
|
delivery_method = None
|
|
if "delivery_method" in data:
|
|
delivery_method = data["delivery_method"]
|
|
|
|
# get data from json
|
|
destination_hash = data["lxmf_message"]["destination_hash"]
|
|
content = data["lxmf_message"]["content"]
|
|
fields = {}
|
|
if "fields" in data["lxmf_message"]:
|
|
fields = data["lxmf_message"]["fields"]
|
|
|
|
# parse image field
|
|
image_field = None
|
|
if "image" in fields:
|
|
image_type = data["lxmf_message"]["fields"]["image"]["image_type"]
|
|
image_bytes = base64.b64decode(
|
|
data["lxmf_message"]["fields"]["image"]["image_bytes"],
|
|
)
|
|
image_field = LxmfImageField(image_type, image_bytes)
|
|
|
|
# parse audio field
|
|
audio_field = None
|
|
if "audio" in fields:
|
|
audio_mode = data["lxmf_message"]["fields"]["audio"]["audio_mode"]
|
|
audio_bytes = base64.b64decode(
|
|
data["lxmf_message"]["fields"]["audio"]["audio_bytes"],
|
|
)
|
|
audio_field = LxmfAudioField(audio_mode, audio_bytes)
|
|
|
|
# parse file attachments field
|
|
file_attachments_field = None
|
|
if "file_attachments" in fields:
|
|
file_attachments = []
|
|
for file_attachment in data["lxmf_message"]["fields"][
|
|
"file_attachments"
|
|
]:
|
|
file_name = file_attachment["file_name"]
|
|
file_bytes = base64.b64decode(file_attachment["file_bytes"])
|
|
file_attachments.append(LxmfFileAttachment(file_name, file_bytes))
|
|
|
|
file_attachments_field = LxmfFileAttachmentsField(file_attachments)
|
|
|
|
try:
|
|
# send lxmf message to destination
|
|
lxmf_message = await self.send_message(
|
|
destination_hash=destination_hash,
|
|
content=content,
|
|
image_field=image_field,
|
|
audio_field=audio_field,
|
|
file_attachments_field=file_attachments_field,
|
|
delivery_method=delivery_method,
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"lxmf_message": self.convert_lxmf_message_to_dict(lxmf_message),
|
|
},
|
|
)
|
|
|
|
except Exception as e:
|
|
return web.json_response(
|
|
{
|
|
"message": f"Sending Failed: {e!s}",
|
|
},
|
|
status=503,
|
|
)
|
|
|
|
# cancel sending lxmf message
|
|
@routes.post("/api/v1/lxmf-messages/{hash}/cancel")
|
|
async def lxmf_messages_cancel(request):
|
|
# get path params
|
|
hash = request.match_info.get("hash", None)
|
|
|
|
# convert hash to bytes
|
|
hash_as_bytes = bytes.fromhex(hash)
|
|
|
|
# cancel outbound message by lxmf message hash
|
|
self.message_router.cancel_outbound(hash_as_bytes)
|
|
|
|
# get lxmf message from database
|
|
lxmf_message = None
|
|
db_lxmf_message = database.LxmfMessage.get_or_none(
|
|
database.LxmfMessage.hash == hash,
|
|
)
|
|
if db_lxmf_message is not None:
|
|
lxmf_message = self.convert_db_lxmf_message_to_dict(db_lxmf_message)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "ok",
|
|
"lxmf_message": lxmf_message,
|
|
},
|
|
)
|
|
|
|
# identify self on existing nomadnetwork link
|
|
@routes.post("/api/v1/nomadnetwork/{destination_hash}/identify")
|
|
async def nomadnetwork_identify(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# identify to existing active link
|
|
if destination_hash in nomadnet_cached_links:
|
|
link = nomadnet_cached_links[destination_hash]
|
|
if link.status is RNS.Link.ACTIVE:
|
|
link.identify(self.identity)
|
|
return web.json_response(
|
|
{
|
|
"message": "Identity has been sent!",
|
|
},
|
|
)
|
|
|
|
# failed to identify
|
|
return web.json_response(
|
|
{
|
|
"message": "Failed to identify. No active link to destination.",
|
|
},
|
|
status=500,
|
|
)
|
|
|
|
# delete lxmf message
|
|
@routes.delete("/api/v1/lxmf-messages/{hash}")
|
|
async def lxmf_messages_delete(request):
|
|
# get path params
|
|
hash = request.match_info.get("hash", None)
|
|
|
|
# hash is required
|
|
if hash is None:
|
|
return web.json_response(
|
|
{
|
|
"message": "hash is required",
|
|
},
|
|
status=422,
|
|
)
|
|
|
|
# delete lxmf messages from db where hash matches
|
|
database.LxmfMessage.delete().where(
|
|
database.LxmfMessage.hash == hash,
|
|
).execute()
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "ok",
|
|
},
|
|
)
|
|
|
|
# serve lxmf messages for conversation
|
|
@routes.get("/api/v1/lxmf-messages/conversation/{destination_hash}")
|
|
async def lxmf_messages_conversation(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
order = request.query.get("order", "asc")
|
|
count = request.query.get("count")
|
|
after_id = request.query.get("after_id")
|
|
|
|
# get source hash from local lxmf destination
|
|
source_hash = self.local_lxmf_destination.hash.hex()
|
|
|
|
# get lxmf messages from db where "source to destination" or "destination to source" and ordered by oldest to newest
|
|
db_lxmf_messages = (
|
|
database.LxmfMessage.select()
|
|
.where(
|
|
(database.LxmfMessage.source_hash == source_hash)
|
|
& (database.LxmfMessage.destination_hash == destination_hash),
|
|
)
|
|
.orwhere(
|
|
(database.LxmfMessage.destination_hash == source_hash)
|
|
& (database.LxmfMessage.source_hash == destination_hash),
|
|
)
|
|
.order_by(
|
|
database.LxmfMessage.id.asc()
|
|
if order == "asc"
|
|
else database.LxmfMessage.id.desc(),
|
|
)
|
|
)
|
|
|
|
# limit how many messages to return
|
|
if count is not None:
|
|
db_lxmf_messages = db_lxmf_messages.limit(count)
|
|
|
|
# only get records after provided id, based on query order
|
|
if after_id is not None:
|
|
if order == "asc":
|
|
db_lxmf_messages = db_lxmf_messages.where(
|
|
database.LxmfMessage.id > after_id,
|
|
)
|
|
else:
|
|
db_lxmf_messages = db_lxmf_messages.where(
|
|
database.LxmfMessage.id < after_id,
|
|
)
|
|
|
|
# convert to response json
|
|
lxmf_messages = []
|
|
for db_lxmf_message in db_lxmf_messages:
|
|
lxmf_messages.append(
|
|
self.convert_db_lxmf_message_to_dict(db_lxmf_message),
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"lxmf_messages": lxmf_messages,
|
|
},
|
|
)
|
|
|
|
# delete lxmf messages for conversation
|
|
@routes.delete("/api/v1/lxmf-messages/conversation/{destination_hash}")
|
|
async def lxmf_messages_conversation_delete(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# get source hash from local lxmf destination
|
|
source_hash = self.local_lxmf_destination.hash.hex()
|
|
|
|
# delete lxmf messages from db where "source to destination" or "destination to source"
|
|
(
|
|
database.LxmfMessage.delete()
|
|
.where(
|
|
(database.LxmfMessage.source_hash == source_hash)
|
|
& (database.LxmfMessage.destination_hash == destination_hash),
|
|
)
|
|
.orwhere(
|
|
(database.LxmfMessage.destination_hash == source_hash)
|
|
& (database.LxmfMessage.source_hash == destination_hash),
|
|
)
|
|
.execute()
|
|
)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "ok",
|
|
},
|
|
)
|
|
|
|
# get unqiue lxmf conversations
|
|
@routes.get("/api/v1/lxmf/conversations")
|
|
async def lxmf_conversations_get(request):
|
|
search_query = request.query.get("search", None)
|
|
filter_unread = self.parse_bool_query_param(
|
|
request.query.get("filter_unread"),
|
|
)
|
|
filter_failed = self.parse_bool_query_param(
|
|
request.query.get("filter_failed"),
|
|
)
|
|
filter_has_attachments = self.parse_bool_query_param(
|
|
request.query.get("filter_has_attachments"),
|
|
)
|
|
|
|
search_destination_hashes = set()
|
|
if search_query is not None and search_query != "":
|
|
search_destination_hashes = self.search_destination_hashes_by_message(
|
|
search_query,
|
|
)
|
|
|
|
# sql query to fetch unique source/destination hash pairs ordered by the most recently updated message
|
|
query = """
|
|
WITH NormalizedMessages AS (
|
|
SELECT
|
|
CASE WHEN source_hash < destination_hash THEN source_hash ELSE destination_hash END AS normalized_source,
|
|
CASE WHEN source_hash < destination_hash THEN destination_hash ELSE source_hash END AS normalized_destination,
|
|
MAX(created_at) AS most_recent_created_at
|
|
FROM lxmf_messages
|
|
GROUP BY normalized_source, normalized_destination
|
|
)
|
|
SELECT
|
|
normalized_source AS source_hash,
|
|
normalized_destination AS destination_hash,
|
|
most_recent_created_at
|
|
FROM NormalizedMessages
|
|
ORDER BY most_recent_created_at DESC;
|
|
"""
|
|
|
|
# execute sql query
|
|
cursor = database.database.execute_sql(query)
|
|
|
|
# parse results to get a list of conversations we have sent or received a message from
|
|
conversations = []
|
|
for row in cursor.fetchall():
|
|
# get data from row
|
|
source_hash = row[0]
|
|
destination_hash = row[1]
|
|
created_at = row[2]
|
|
|
|
# determine destination hash of other user
|
|
if source_hash == self.local_lxmf_destination.hexhash:
|
|
other_user_hash = destination_hash
|
|
else:
|
|
other_user_hash = source_hash
|
|
|
|
latest_message = self.get_conversation_latest_message(other_user_hash)
|
|
latest_message_title = None
|
|
latest_message_preview = None
|
|
latest_message_created_at = None
|
|
latest_message_has_attachments = False
|
|
if latest_message is not None:
|
|
latest_message_title = latest_message.title
|
|
latest_message_preview = latest_message.content
|
|
latest_message_created_at = latest_message.created_at
|
|
latest_message_has_attachments = (
|
|
self.message_fields_have_attachments(latest_message.fields)
|
|
)
|
|
|
|
has_attachments = self.conversation_has_attachments(other_user_hash)
|
|
|
|
# find lxmf user icon from database
|
|
lxmf_user_icon = None
|
|
db_lxmf_user_icon = database.LxmfUserIcon.get_or_none(
|
|
database.LxmfUserIcon.destination_hash == other_user_hash,
|
|
)
|
|
if db_lxmf_user_icon is not None:
|
|
lxmf_user_icon = {
|
|
"icon_name": db_lxmf_user_icon.icon_name,
|
|
"foreground_colour": db_lxmf_user_icon.foreground_colour,
|
|
"background_colour": db_lxmf_user_icon.background_colour,
|
|
}
|
|
|
|
# add to conversations
|
|
conversations.append(
|
|
{
|
|
"display_name": self.get_lxmf_conversation_name(
|
|
other_user_hash,
|
|
),
|
|
"custom_display_name": self.get_custom_destination_display_name(
|
|
other_user_hash,
|
|
),
|
|
"destination_hash": other_user_hash,
|
|
"is_unread": ReticulumMeshChat.is_lxmf_conversation_unread(
|
|
other_user_hash,
|
|
),
|
|
"failed_messages_count": ReticulumMeshChat.lxmf_conversation_failed_messages_count(
|
|
other_user_hash,
|
|
),
|
|
"has_attachments": has_attachments,
|
|
"latest_message_title": latest_message_title,
|
|
"latest_message_preview": latest_message_preview,
|
|
"latest_message_created_at": latest_message_created_at,
|
|
"latest_message_has_attachments": latest_message_has_attachments,
|
|
"lxmf_user_icon": lxmf_user_icon,
|
|
# we say the conversation was updated when the latest message was created
|
|
# otherwise this will go crazy when sending a message, as the updated_at on the latest message changes very frequently
|
|
"updated_at": created_at,
|
|
},
|
|
)
|
|
|
|
if search_query is not None and search_query != "":
|
|
lowered_query = search_query.lower()
|
|
filtered = []
|
|
for conversation in conversations:
|
|
matches_display = (
|
|
conversation["display_name"]
|
|
and lowered_query in conversation["display_name"].lower()
|
|
)
|
|
matches_custom = (
|
|
conversation["custom_display_name"]
|
|
and lowered_query in conversation["custom_display_name"].lower()
|
|
)
|
|
matches_destination = (
|
|
conversation["destination_hash"]
|
|
and lowered_query in conversation["destination_hash"].lower()
|
|
)
|
|
matches_latest_title = (
|
|
conversation["latest_message_title"]
|
|
and lowered_query
|
|
in conversation["latest_message_title"].lower()
|
|
)
|
|
matches_latest_preview = (
|
|
conversation["latest_message_preview"]
|
|
and lowered_query
|
|
in conversation["latest_message_preview"].lower()
|
|
)
|
|
matches_history = (
|
|
conversation["destination_hash"] in search_destination_hashes
|
|
)
|
|
if (
|
|
matches_display
|
|
or matches_custom
|
|
or matches_destination
|
|
or matches_latest_title
|
|
or matches_latest_preview
|
|
or matches_history
|
|
):
|
|
filtered.append(conversation)
|
|
conversations = filtered
|
|
|
|
if filter_unread:
|
|
conversations = [c for c in conversations if c["is_unread"]]
|
|
|
|
if filter_failed:
|
|
conversations = [
|
|
c for c in conversations if c["failed_messages_count"] > 0
|
|
]
|
|
|
|
if filter_has_attachments:
|
|
conversations = [c for c in conversations if c["has_attachments"]]
|
|
|
|
return web.json_response(
|
|
{
|
|
"conversations": conversations,
|
|
},
|
|
)
|
|
|
|
# mark lxmf conversation as read
|
|
@routes.get("/api/v1/lxmf/conversations/{destination_hash}/mark-as-read")
|
|
async def lxmf_conversations_mark_read(request):
|
|
# get path params
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
|
|
# mark lxmf conversation as read
|
|
self.db_mark_lxmf_conversation_as_read(destination_hash)
|
|
|
|
return web.json_response(
|
|
{
|
|
"message": "ok",
|
|
},
|
|
)
|
|
|
|
# get blocked destinations
|
|
@routes.get("/api/v1/blocked-destinations")
|
|
async def blocked_destinations_get(request):
|
|
blocked = database.BlockedDestination.select()
|
|
blocked_list = []
|
|
for b in blocked:
|
|
created_at = b.created_at
|
|
if isinstance(created_at, str):
|
|
created_at_str = created_at
|
|
else:
|
|
created_at_str = (
|
|
created_at.isoformat()
|
|
if hasattr(created_at, "isoformat")
|
|
else str(created_at)
|
|
)
|
|
blocked_list.append(
|
|
{
|
|
"destination_hash": b.destination_hash,
|
|
"created_at": created_at_str,
|
|
},
|
|
)
|
|
return web.json_response(
|
|
{
|
|
"blocked_destinations": blocked_list,
|
|
},
|
|
)
|
|
|
|
# add blocked destination
|
|
@routes.post("/api/v1/blocked-destinations")
|
|
async def blocked_destinations_add(request):
|
|
data = await request.json()
|
|
destination_hash = data.get("destination_hash", "")
|
|
if not destination_hash or len(destination_hash) != 32:
|
|
return web.json_response(
|
|
{"error": "Invalid destination hash"},
|
|
status=400,
|
|
)
|
|
|
|
try:
|
|
database.BlockedDestination.create(destination_hash=destination_hash)
|
|
# drop any existing paths to this destination
|
|
try:
|
|
RNS.Transport.drop_path(bytes.fromhex(destination_hash))
|
|
except Exception:
|
|
pass
|
|
return web.json_response({"message": "ok"})
|
|
except Exception:
|
|
return web.json_response(
|
|
{"error": "Destination already blocked"},
|
|
status=400,
|
|
)
|
|
|
|
# remove blocked destination
|
|
@routes.delete("/api/v1/blocked-destinations/{destination_hash}")
|
|
async def blocked_destinations_delete(request):
|
|
destination_hash = request.match_info.get("destination_hash", "")
|
|
if not destination_hash or len(destination_hash) != 32:
|
|
return web.json_response(
|
|
{"error": "Invalid destination hash"},
|
|
status=400,
|
|
)
|
|
|
|
try:
|
|
blocked = database.BlockedDestination.get_or_none(
|
|
database.BlockedDestination.destination_hash == destination_hash,
|
|
)
|
|
if blocked:
|
|
blocked.delete_instance()
|
|
return web.json_response({"message": "ok"})
|
|
return web.json_response(
|
|
{"error": "Destination not blocked"},
|
|
status=404,
|
|
)
|
|
except Exception as e:
|
|
return web.json_response({"error": str(e)}, status=500)
|
|
|
|
# get spam keywords
|
|
@routes.get("/api/v1/spam-keywords")
|
|
async def spam_keywords_get(request):
|
|
keywords = database.SpamKeyword.select()
|
|
keyword_list = [
|
|
{
|
|
"id": k.id,
|
|
"keyword": k.keyword,
|
|
"created_at": k.created_at.isoformat(),
|
|
}
|
|
for k in keywords
|
|
]
|
|
return web.json_response(
|
|
{
|
|
"spam_keywords": keyword_list,
|
|
},
|
|
)
|
|
|
|
# add spam keyword
|
|
@routes.post("/api/v1/spam-keywords")
|
|
async def spam_keywords_add(request):
|
|
data = await request.json()
|
|
keyword = data.get("keyword", "").strip()
|
|
if not keyword:
|
|
return web.json_response({"error": "Keyword is required"}, status=400)
|
|
|
|
try:
|
|
database.SpamKeyword.create(keyword=keyword)
|
|
return web.json_response({"message": "ok"})
|
|
except Exception:
|
|
return web.json_response(
|
|
{"error": "Keyword already exists"},
|
|
status=400,
|
|
)
|
|
|
|
# remove spam keyword
|
|
@routes.delete("/api/v1/spam-keywords/{keyword_id}")
|
|
async def spam_keywords_delete(request):
|
|
keyword_id = request.match_info.get("keyword_id", "")
|
|
try:
|
|
keyword_id = int(keyword_id)
|
|
except (ValueError, TypeError):
|
|
return web.json_response({"error": "Invalid keyword ID"}, status=400)
|
|
|
|
try:
|
|
keyword = database.SpamKeyword.get_or_none(
|
|
database.SpamKeyword.id == keyword_id,
|
|
)
|
|
if keyword:
|
|
keyword.delete_instance()
|
|
return web.json_response({"message": "ok"})
|
|
return web.json_response({"error": "Keyword not found"}, status=404)
|
|
except Exception as e:
|
|
return web.json_response({"error": str(e)}, status=500)
|
|
|
|
# mark message as spam or not spam
|
|
@routes.post("/api/v1/lxmf-messages/{hash}/spam")
|
|
async def lxmf_messages_spam(request):
|
|
message_hash = request.match_info.get("hash", "")
|
|
data = await request.json()
|
|
is_spam = data.get("is_spam", False)
|
|
|
|
try:
|
|
message = database.LxmfMessage.get_or_none(
|
|
database.LxmfMessage.hash == message_hash,
|
|
)
|
|
if message:
|
|
message.is_spam = is_spam
|
|
message.updated_at = datetime.now(UTC)
|
|
message.save()
|
|
return web.json_response({"message": "ok"})
|
|
return web.json_response({"error": "Message not found"}, status=404)
|
|
except Exception as e:
|
|
return web.json_response({"error": str(e)}, status=500)
|
|
|
|
# security headers middleware
|
|
@web.middleware
|
|
async def security_middleware(request, handler):
|
|
response = await handler(request)
|
|
# Add security headers to all responses
|
|
response.headers["X-Content-Type-Options"] = "nosniff"
|
|
response.headers["X-Frame-Options"] = "DENY"
|
|
response.headers["X-XSS-Protection"] = "1; mode=block"
|
|
response.headers["Referrer-Policy"] = "strict-origin-when-cross-origin"
|
|
# CSP: allow localhost for development and Electron, websockets, and blob URLs
|
|
csp = (
|
|
"default-src 'self'; "
|
|
"script-src 'self' 'unsafe-inline' 'unsafe-eval'; "
|
|
"style-src 'self' 'unsafe-inline'; "
|
|
"img-src 'self' data: blob:; "
|
|
"font-src 'self' data:; "
|
|
"connect-src 'self' ws://localhost:* wss://localhost:* blob:; "
|
|
"media-src 'self' blob:; "
|
|
"worker-src 'self' blob:; "
|
|
"object-src 'none'; "
|
|
"base-uri 'self';"
|
|
)
|
|
response.headers["Content-Security-Policy"] = csp
|
|
return response
|
|
|
|
# called when web app has started
|
|
async def on_startup(app):
|
|
# remember main event loop
|
|
AsyncUtils.set_main_loop(asyncio.get_event_loop())
|
|
|
|
# auto launch web browser
|
|
if launch_browser:
|
|
try:
|
|
webbrowser.open(f"http://127.0.0.1:{port}")
|
|
except Exception:
|
|
print("failed to launch web browser")
|
|
|
|
# create and run web app
|
|
app = web.Application(
|
|
client_max_size=1024 * 1024 * 50,
|
|
middlewares=[security_middleware],
|
|
) # allow uploading files up to 50mb
|
|
app.add_routes(routes)
|
|
app.add_routes(
|
|
[web.static("/", get_file_path("public/"))],
|
|
) # serve anything in public folder
|
|
app.on_shutdown.append(
|
|
self.shutdown,
|
|
) # need to force close websockets and stop reticulum now
|
|
app.on_startup.append(on_startup)
|
|
web.run_app(app, host=host, port=port)
|
|
|
|
# handle announcing
|
|
async def announce(self):
|
|
# update last announced at timestamp
|
|
self.config.last_announced_at.set(int(time.time()))
|
|
|
|
# send announce for lxmf (ensuring name is updated before announcing)
|
|
self.local_lxmf_destination.display_name = self.config.display_name.get()
|
|
self.message_router.announce(destination_hash=self.local_lxmf_destination.hash)
|
|
|
|
# send announce for local propagation node (if enabled)
|
|
if self.config.lxmf_local_propagation_node_enabled.get():
|
|
self.message_router.announce_propagation_node()
|
|
|
|
# send announce for audio call
|
|
self.audio_call_manager.announce(
|
|
app_data=self.config.display_name.get().encode("utf-8"),
|
|
)
|
|
|
|
# tell websocket clients we just announced
|
|
await self.send_announced_to_websocket_clients()
|
|
|
|
# handle syncing propagation nodes
|
|
async def sync_propagation_nodes(self):
|
|
# update last synced at timestamp
|
|
self.config.lxmf_preferred_propagation_node_last_synced_at.set(int(time.time()))
|
|
|
|
# request messages from propagation node
|
|
self.message_router.request_messages_from_propagation_node(self.identity)
|
|
|
|
# send config to websocket clients (used to tell ui last synced at)
|
|
await self.send_config_to_websocket_clients()
|
|
|
|
async def update_config(self, data):
|
|
# update display name in config
|
|
if "display_name" in data and data["display_name"] != "":
|
|
self.config.display_name.set(data["display_name"])
|
|
|
|
# update theme in config
|
|
if "theme" in data and data["theme"] != "":
|
|
self.config.theme.set(data["theme"])
|
|
|
|
# update auto announce interval
|
|
if "auto_announce_interval_seconds" in data:
|
|
# auto auto announce interval
|
|
auto_announce_interval_seconds = int(data["auto_announce_interval_seconds"])
|
|
self.config.auto_announce_interval_seconds.set(
|
|
data["auto_announce_interval_seconds"],
|
|
)
|
|
|
|
# enable or disable auto announce based on interval
|
|
if auto_announce_interval_seconds > 0:
|
|
self.config.auto_announce_enabled.set(True)
|
|
else:
|
|
self.config.auto_announce_enabled.set(False)
|
|
|
|
if "auto_resend_failed_messages_when_announce_received" in data:
|
|
value = bool(data["auto_resend_failed_messages_when_announce_received"])
|
|
self.config.auto_resend_failed_messages_when_announce_received.set(value)
|
|
|
|
if "allow_auto_resending_failed_messages_with_attachments" in data:
|
|
value = bool(data["allow_auto_resending_failed_messages_with_attachments"])
|
|
self.config.allow_auto_resending_failed_messages_with_attachments.set(value)
|
|
|
|
if "auto_send_failed_messages_to_propagation_node" in data:
|
|
value = bool(data["auto_send_failed_messages_to_propagation_node"])
|
|
self.config.auto_send_failed_messages_to_propagation_node.set(value)
|
|
|
|
if "show_suggested_community_interfaces" in data:
|
|
value = bool(data["show_suggested_community_interfaces"])
|
|
self.config.show_suggested_community_interfaces.set(value)
|
|
|
|
if "lxmf_preferred_propagation_node_destination_hash" in data:
|
|
# update config value
|
|
value = data["lxmf_preferred_propagation_node_destination_hash"]
|
|
self.config.lxmf_preferred_propagation_node_destination_hash.set(value)
|
|
|
|
# update active propagation node
|
|
self.set_active_propagation_node(value)
|
|
|
|
# update inbound stamp cost (for direct delivery messages)
|
|
if "lxmf_inbound_stamp_cost" in data:
|
|
value = int(data["lxmf_inbound_stamp_cost"])
|
|
# validate stamp cost (must be between 1 and 254)
|
|
if value < 1:
|
|
value = None
|
|
elif value >= 255:
|
|
value = 254
|
|
self.config.lxmf_inbound_stamp_cost.set(value)
|
|
# update the inbound stamp cost on the delivery destination
|
|
self.message_router.set_inbound_stamp_cost(
|
|
self.local_lxmf_destination.hash,
|
|
value,
|
|
)
|
|
# re-announce to update the stamp cost in announces
|
|
self.local_lxmf_destination.display_name = self.config.display_name.get()
|
|
self.message_router.announce(
|
|
destination_hash=self.local_lxmf_destination.hash,
|
|
)
|
|
|
|
# update propagation node stamp cost (for messages propagated through your node)
|
|
if "lxmf_propagation_node_stamp_cost" in data:
|
|
value = int(data["lxmf_propagation_node_stamp_cost"])
|
|
# validate stamp cost (must be at least 13, per LXMF minimum)
|
|
if value < 13:
|
|
value = 13
|
|
elif value >= 255:
|
|
value = 254
|
|
self.config.lxmf_propagation_node_stamp_cost.set(value)
|
|
# update the propagation stamp cost on the router
|
|
self.message_router.propagation_stamp_cost = value
|
|
# re-announce propagation node if enabled
|
|
if self.config.lxmf_local_propagation_node_enabled.get():
|
|
self.message_router.announce_propagation_node()
|
|
|
|
# update auto sync interval
|
|
if "lxmf_preferred_propagation_node_auto_sync_interval_seconds" in data:
|
|
value = int(
|
|
data["lxmf_preferred_propagation_node_auto_sync_interval_seconds"],
|
|
)
|
|
self.config.lxmf_preferred_propagation_node_auto_sync_interval_seconds.set(
|
|
value,
|
|
)
|
|
|
|
if "lxmf_local_propagation_node_enabled" in data:
|
|
# update config value
|
|
value = bool(data["lxmf_local_propagation_node_enabled"])
|
|
self.config.lxmf_local_propagation_node_enabled.set(value)
|
|
|
|
# enable or disable local propagation node
|
|
self.enable_local_propagation_node(value)
|
|
|
|
# update lxmf user icon name in config
|
|
if "lxmf_user_icon_name" in data:
|
|
self.config.lxmf_user_icon_name.set(data["lxmf_user_icon_name"])
|
|
|
|
# update lxmf user icon foreground colour in config
|
|
if "lxmf_user_icon_foreground_colour" in data:
|
|
self.config.lxmf_user_icon_foreground_colour.set(
|
|
data["lxmf_user_icon_foreground_colour"],
|
|
)
|
|
|
|
# update lxmf user icon background colour in config
|
|
if "lxmf_user_icon_background_colour" in data:
|
|
self.config.lxmf_user_icon_background_colour.set(
|
|
data["lxmf_user_icon_background_colour"],
|
|
)
|
|
|
|
# send config to websocket clients
|
|
await self.send_config_to_websocket_clients()
|
|
|
|
# converts nomadnetwork page variables from a string to a map
|
|
# converts: "field1=123|field2=456"
|
|
# to the following map:
|
|
# - var_field1: 123
|
|
# - var_field2: 456
|
|
@staticmethod
|
|
def convert_nomadnet_string_data_to_map(path_data: str | None):
|
|
data = {}
|
|
if path_data is not None:
|
|
for field in path_data.split("|"):
|
|
if "=" in field:
|
|
variable_name, variable_value = field.split("=")
|
|
data[f"var_{variable_name}"] = variable_value
|
|
else:
|
|
print(f"unhandled field: {field}")
|
|
return data
|
|
|
|
@staticmethod
|
|
def convert_nomadnet_field_data_to_map(field_data):
|
|
data = {}
|
|
if field_data is not None or "{}":
|
|
try:
|
|
json_data = field_data
|
|
if isinstance(json_data, dict):
|
|
# add the prefixed keys to the result dictionary
|
|
data = {f"field_{key}": value for key, value in json_data.items()}
|
|
else:
|
|
return None
|
|
except Exception as e:
|
|
print(f"skipping invalid field data: {e}")
|
|
|
|
return data
|
|
|
|
# handle data received from websocket client
|
|
async def on_websocket_data_received(self, client, data):
|
|
# get type from client data
|
|
_type = data["type"]
|
|
|
|
# handle ping
|
|
if _type == "ping":
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "pong",
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# handle updating config
|
|
elif _type == "config.set":
|
|
# get config from websocket
|
|
config = data["config"]
|
|
|
|
# update config
|
|
await self.update_config(config)
|
|
|
|
# handle canceling a download
|
|
elif _type == "nomadnet.download.cancel":
|
|
# get data from websocket client
|
|
download_id = data["download_id"]
|
|
|
|
# cancel the download
|
|
if download_id in self.active_downloads:
|
|
downloader = self.active_downloads[download_id]
|
|
downloader.cancel()
|
|
del self.active_downloads[download_id]
|
|
|
|
# notify client
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.download.cancelled",
|
|
"download_id": download_id,
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# handle downloading a file from a nomadnet node
|
|
elif _type == "nomadnet.file.download":
|
|
# get data from websocket client
|
|
destination_hash = data["nomadnet_file_download"]["destination_hash"]
|
|
file_path = data["nomadnet_file_download"]["file_path"]
|
|
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# generate download id
|
|
self.download_id_counter += 1
|
|
download_id = self.download_id_counter
|
|
|
|
# handle successful file download
|
|
def on_file_download_success(file_name, file_bytes):
|
|
# remove from active downloads
|
|
if download_id in self.active_downloads:
|
|
del self.active_downloads[download_id]
|
|
|
|
# Track download speed
|
|
download_size = len(file_bytes)
|
|
if hasattr(downloader, "start_time") and downloader.start_time:
|
|
download_duration = time.time() - downloader.start_time
|
|
if download_duration > 0:
|
|
self.download_speeds.append((download_size, download_duration))
|
|
# Keep only last 100 downloads for average calculation
|
|
if len(self.download_speeds) > 100:
|
|
self.download_speeds.pop(0)
|
|
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.file.download",
|
|
"download_id": download_id,
|
|
"nomadnet_file_download": {
|
|
"status": "success",
|
|
"destination_hash": destination_hash.hex(),
|
|
"file_path": file_path,
|
|
"file_name": file_name,
|
|
"file_bytes": base64.b64encode(file_bytes).decode(
|
|
"utf-8",
|
|
),
|
|
},
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# handle file download failure
|
|
def on_file_download_failure(failure_reason):
|
|
# remove from active downloads
|
|
if download_id in self.active_downloads:
|
|
del self.active_downloads[download_id]
|
|
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.file.download",
|
|
"download_id": download_id,
|
|
"nomadnet_file_download": {
|
|
"status": "failure",
|
|
"failure_reason": failure_reason,
|
|
"destination_hash": destination_hash.hex(),
|
|
"file_path": file_path,
|
|
},
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# handle file download progress
|
|
def on_file_download_progress(progress):
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.file.download",
|
|
"download_id": download_id,
|
|
"nomadnet_file_download": {
|
|
"status": "progress",
|
|
"progress": progress,
|
|
"destination_hash": destination_hash.hex(),
|
|
"file_path": file_path,
|
|
},
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# download the file
|
|
downloader = NomadnetFileDownloader(
|
|
destination_hash,
|
|
file_path,
|
|
on_file_download_success,
|
|
on_file_download_failure,
|
|
on_file_download_progress,
|
|
)
|
|
downloader.start_time = time.time()
|
|
self.active_downloads[download_id] = downloader
|
|
|
|
# notify client download started
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.file.download",
|
|
"download_id": download_id,
|
|
"nomadnet_file_download": {
|
|
"status": "started",
|
|
"destination_hash": destination_hash.hex(),
|
|
"file_path": file_path,
|
|
},
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
AsyncUtils.run_async(downloader.download())
|
|
|
|
# handle downloading a page from a nomadnet node
|
|
elif _type == "nomadnet.page.download":
|
|
# get data from websocket client
|
|
destination_hash = data["nomadnet_page_download"]["destination_hash"]
|
|
page_path = data["nomadnet_page_download"]["page_path"]
|
|
field_data = data["nomadnet_page_download"]["field_data"]
|
|
|
|
# generate download id
|
|
self.download_id_counter += 1
|
|
download_id = self.download_id_counter
|
|
|
|
combined_data = {}
|
|
# parse data from page path
|
|
# example: hash:/page/index.mu`field1=123|field2=456
|
|
page_data = None
|
|
page_path_to_download = page_path
|
|
if "`" in page_path:
|
|
page_path_parts = page_path.split("`")
|
|
page_path_to_download = page_path_parts[0]
|
|
page_data = self.convert_nomadnet_string_data_to_map(page_path_parts[1])
|
|
|
|
# Field data
|
|
field_data = self.convert_nomadnet_field_data_to_map(field_data)
|
|
|
|
# Combine page data and field data
|
|
if page_data is not None:
|
|
combined_data.update(page_data)
|
|
if field_data is not None:
|
|
combined_data.update(field_data)
|
|
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# handle successful page download
|
|
def on_page_download_success(page_content):
|
|
# remove from active downloads
|
|
if download_id in self.active_downloads:
|
|
del self.active_downloads[download_id]
|
|
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.page.download",
|
|
"download_id": download_id,
|
|
"nomadnet_page_download": {
|
|
"status": "success",
|
|
"destination_hash": destination_hash.hex(),
|
|
"page_path": page_path,
|
|
"page_content": page_content,
|
|
},
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# handle page download failure
|
|
def on_page_download_failure(failure_reason):
|
|
# remove from active downloads
|
|
if download_id in self.active_downloads:
|
|
del self.active_downloads[download_id]
|
|
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.page.download",
|
|
"download_id": download_id,
|
|
"nomadnet_page_download": {
|
|
"status": "failure",
|
|
"failure_reason": failure_reason,
|
|
"destination_hash": destination_hash.hex(),
|
|
"page_path": page_path,
|
|
},
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# handle page download progress
|
|
def on_page_download_progress(progress):
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.page.download",
|
|
"download_id": download_id,
|
|
"nomadnet_page_download": {
|
|
"status": "progress",
|
|
"progress": progress,
|
|
"destination_hash": destination_hash.hex(),
|
|
"page_path": page_path,
|
|
},
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# download the page
|
|
downloader = NomadnetPageDownloader(
|
|
destination_hash,
|
|
page_path_to_download,
|
|
combined_data,
|
|
on_page_download_success,
|
|
on_page_download_failure,
|
|
on_page_download_progress,
|
|
)
|
|
self.active_downloads[download_id] = downloader
|
|
|
|
# notify client download started
|
|
AsyncUtils.run_async(
|
|
client.send_str(
|
|
json.dumps(
|
|
{
|
|
"type": "nomadnet.page.download",
|
|
"download_id": download_id,
|
|
"nomadnet_page_download": {
|
|
"status": "started",
|
|
"destination_hash": destination_hash.hex(),
|
|
"page_path": page_path,
|
|
},
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
AsyncUtils.run_async(downloader.download())
|
|
|
|
# unhandled type
|
|
else:
|
|
print("unhandled client message type: " + _type)
|
|
|
|
# broadcast provided data to all connected websocket clients
|
|
async def websocket_broadcast(self, data):
|
|
for websocket_client in self.websocket_clients:
|
|
try:
|
|
await websocket_client.send_str(data)
|
|
except Exception:
|
|
# do nothing if failed to broadcast to a specific websocket client
|
|
pass
|
|
|
|
# broadcasts config to all websocket clients
|
|
async def send_config_to_websocket_clients(self):
|
|
await self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "config",
|
|
"config": self.get_config_dict(),
|
|
},
|
|
),
|
|
)
|
|
|
|
# broadcasts to all websocket clients that we just announced
|
|
async def send_announced_to_websocket_clients(self):
|
|
await self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "announced",
|
|
},
|
|
),
|
|
)
|
|
|
|
# returns a dictionary of config
|
|
def get_config_dict(self):
|
|
return {
|
|
"display_name": self.config.display_name.get(),
|
|
"identity_hash": self.identity.hexhash,
|
|
"lxmf_address_hash": self.local_lxmf_destination.hexhash,
|
|
"audio_call_address_hash": self.audio_call_manager.audio_call_receiver.destination.hexhash,
|
|
"is_transport_enabled": self.reticulum.transport_enabled(),
|
|
"auto_announce_enabled": self.config.auto_announce_enabled.get(),
|
|
"auto_announce_interval_seconds": self.config.auto_announce_interval_seconds.get(),
|
|
"last_announced_at": self.config.last_announced_at.get(),
|
|
"theme": self.config.theme.get(),
|
|
"auto_resend_failed_messages_when_announce_received": self.config.auto_resend_failed_messages_when_announce_received.get(),
|
|
"allow_auto_resending_failed_messages_with_attachments": self.config.allow_auto_resending_failed_messages_with_attachments.get(),
|
|
"auto_send_failed_messages_to_propagation_node": self.config.auto_send_failed_messages_to_propagation_node.get(),
|
|
"show_suggested_community_interfaces": self.config.show_suggested_community_interfaces.get(),
|
|
"lxmf_local_propagation_node_enabled": self.config.lxmf_local_propagation_node_enabled.get(),
|
|
"lxmf_local_propagation_node_address_hash": self.message_router.propagation_destination.hexhash,
|
|
"lxmf_preferred_propagation_node_destination_hash": self.config.lxmf_preferred_propagation_node_destination_hash.get(),
|
|
"lxmf_preferred_propagation_node_auto_sync_interval_seconds": self.config.lxmf_preferred_propagation_node_auto_sync_interval_seconds.get(),
|
|
"lxmf_preferred_propagation_node_last_synced_at": self.config.lxmf_preferred_propagation_node_last_synced_at.get(),
|
|
"lxmf_user_icon_name": self.config.lxmf_user_icon_name.get(),
|
|
"lxmf_user_icon_foreground_colour": self.config.lxmf_user_icon_foreground_colour.get(),
|
|
"lxmf_user_icon_background_colour": self.config.lxmf_user_icon_background_colour.get(),
|
|
"lxmf_inbound_stamp_cost": self.config.lxmf_inbound_stamp_cost.get(),
|
|
"lxmf_propagation_node_stamp_cost": self.config.lxmf_propagation_node_stamp_cost.get(),
|
|
}
|
|
|
|
# convert audio call to dict
|
|
def convert_audio_call_to_dict(self, audio_call: AudioCall):
|
|
# get remote identity hash
|
|
remote_identity_hash = None
|
|
remote_identity = audio_call.get_remote_identity()
|
|
if remote_identity is not None:
|
|
remote_identity_hash = remote_identity.hash.hex()
|
|
|
|
# get remote destination hash
|
|
# we need to know the remote identity to determine their destination hash
|
|
remote_destination_hash = None
|
|
remote_destination_hash_hex = None
|
|
if remote_identity is not None:
|
|
remote_destination_hash = RNS.Destination.hash(
|
|
remote_identity,
|
|
"call",
|
|
"audio",
|
|
)
|
|
remote_destination_hash_hex = remote_destination_hash.hex()
|
|
|
|
# determine path to remote destination
|
|
path = None
|
|
if remote_destination_hash is not None:
|
|
# determine next hop and hop count
|
|
hops = RNS.Transport.hops_to(remote_destination_hash)
|
|
next_hop_bytes = self.reticulum.get_next_hop(remote_destination_hash)
|
|
|
|
# ensure next hop provided
|
|
if next_hop_bytes is not None:
|
|
next_hop = next_hop_bytes.hex()
|
|
next_hop_interface = self.reticulum.get_next_hop_if_name(
|
|
remote_destination_hash,
|
|
)
|
|
path = {
|
|
"hops": hops,
|
|
"next_hop": next_hop,
|
|
"next_hop_interface": next_hop_interface,
|
|
}
|
|
|
|
return {
|
|
"hash": audio_call.link.hash.hex(),
|
|
"remote_destination_hash": remote_destination_hash_hex,
|
|
"remote_identity_hash": remote_identity_hash,
|
|
"is_active": audio_call.is_active(),
|
|
"is_outbound": audio_call.is_outbound,
|
|
"path": path,
|
|
}
|
|
|
|
# convert an lxmf message to a dictionary, for sending over websocket
|
|
def convert_lxmf_message_to_dict(self, lxmf_message: LXMF.LXMessage):
|
|
# handle fields
|
|
fields = {}
|
|
message_fields = lxmf_message.get_fields()
|
|
for field_type in message_fields:
|
|
value = message_fields[field_type]
|
|
|
|
# handle file attachments field
|
|
if field_type == LXMF.FIELD_FILE_ATTACHMENTS:
|
|
# process file attachments
|
|
file_attachments = []
|
|
for file_attachment in value:
|
|
file_name = file_attachment[0]
|
|
file_bytes = base64.b64encode(file_attachment[1]).decode("utf-8")
|
|
file_attachments.append(
|
|
{
|
|
"file_name": file_name,
|
|
"file_bytes": file_bytes,
|
|
},
|
|
)
|
|
|
|
# add to fields
|
|
fields["file_attachments"] = file_attachments
|
|
|
|
# handle image field
|
|
if field_type == LXMF.FIELD_IMAGE:
|
|
image_type = value[0]
|
|
image_bytes = base64.b64encode(value[1]).decode("utf-8")
|
|
fields["image"] = {
|
|
"image_type": image_type,
|
|
"image_bytes": image_bytes,
|
|
}
|
|
|
|
# handle audio field
|
|
if field_type == LXMF.FIELD_AUDIO:
|
|
audio_mode = value[0]
|
|
audio_bytes = base64.b64encode(value[1]).decode("utf-8")
|
|
fields["audio"] = {
|
|
"audio_mode": audio_mode,
|
|
"audio_bytes": audio_bytes,
|
|
}
|
|
|
|
# convert 0.0-1.0 progress to 0.00-100 percentage
|
|
progress_percentage = round(lxmf_message.progress * 100, 2)
|
|
|
|
# get rssi
|
|
rssi = lxmf_message.rssi
|
|
if rssi is None:
|
|
rssi = self.reticulum.get_packet_rssi(lxmf_message.hash)
|
|
|
|
# get snr
|
|
snr = lxmf_message.snr
|
|
if snr is None:
|
|
snr = self.reticulum.get_packet_snr(lxmf_message.hash)
|
|
|
|
# get quality
|
|
quality = lxmf_message.q
|
|
if quality is None:
|
|
quality = self.reticulum.get_packet_q(lxmf_message.hash)
|
|
|
|
return {
|
|
"hash": lxmf_message.hash.hex(),
|
|
"source_hash": lxmf_message.source_hash.hex(),
|
|
"destination_hash": lxmf_message.destination_hash.hex(),
|
|
"is_incoming": lxmf_message.incoming,
|
|
"state": self.convert_lxmf_state_to_string(lxmf_message),
|
|
"progress": progress_percentage,
|
|
"method": self.convert_lxmf_method_to_string(lxmf_message),
|
|
"delivery_attempts": lxmf_message.delivery_attempts,
|
|
"next_delivery_attempt_at": getattr(
|
|
lxmf_message,
|
|
"next_delivery_attempt",
|
|
None,
|
|
), # attribute may not exist yet
|
|
"title": lxmf_message.title.decode("utf-8"),
|
|
"content": lxmf_message.content.decode("utf-8"),
|
|
"fields": fields,
|
|
"timestamp": lxmf_message.timestamp,
|
|
"rssi": rssi,
|
|
"snr": snr,
|
|
"quality": quality,
|
|
}
|
|
|
|
# convert lxmf state to a human friendly string
|
|
@staticmethod
|
|
def convert_lxmf_state_to_string(lxmf_message: LXMF.LXMessage):
|
|
# convert state to string
|
|
lxmf_message_state = "unknown"
|
|
if lxmf_message.state == LXMF.LXMessage.GENERATING:
|
|
lxmf_message_state = "generating"
|
|
elif lxmf_message.state == LXMF.LXMessage.OUTBOUND:
|
|
lxmf_message_state = "outbound"
|
|
elif lxmf_message.state == LXMF.LXMessage.SENDING:
|
|
lxmf_message_state = "sending"
|
|
elif lxmf_message.state == LXMF.LXMessage.SENT:
|
|
lxmf_message_state = "sent"
|
|
elif lxmf_message.state == LXMF.LXMessage.DELIVERED:
|
|
lxmf_message_state = "delivered"
|
|
elif lxmf_message.state == LXMF.LXMessage.REJECTED:
|
|
lxmf_message_state = "rejected"
|
|
elif lxmf_message.state == LXMF.LXMessage.CANCELLED:
|
|
lxmf_message_state = "cancelled"
|
|
elif lxmf_message.state == LXMF.LXMessage.FAILED:
|
|
lxmf_message_state = "failed"
|
|
|
|
return lxmf_message_state
|
|
|
|
# convert lxmf method to a human friendly string
|
|
@staticmethod
|
|
def convert_lxmf_method_to_string(lxmf_message: LXMF.LXMessage):
|
|
# convert method to string
|
|
lxmf_message_method = "unknown"
|
|
if lxmf_message.method == LXMF.LXMessage.OPPORTUNISTIC:
|
|
lxmf_message_method = "opportunistic"
|
|
elif lxmf_message.method == LXMF.LXMessage.DIRECT:
|
|
lxmf_message_method = "direct"
|
|
elif lxmf_message.method == LXMF.LXMessage.PROPAGATED:
|
|
lxmf_message_method = "propagated"
|
|
elif lxmf_message.method == LXMF.LXMessage.PAPER:
|
|
lxmf_message_method = "paper"
|
|
|
|
return lxmf_message_method
|
|
|
|
@staticmethod
|
|
def convert_propagation_node_state_to_string(state):
|
|
# map states to strings
|
|
state_map = {
|
|
LXMRouter.PR_IDLE: "idle",
|
|
LXMRouter.PR_PATH_REQUESTED: "path_requested",
|
|
LXMRouter.PR_LINK_ESTABLISHING: "link_establishing",
|
|
LXMRouter.PR_LINK_ESTABLISHED: "link_established",
|
|
LXMRouter.PR_REQUEST_SENT: "request_sent",
|
|
LXMRouter.PR_RECEIVING: "receiving",
|
|
LXMRouter.PR_RESPONSE_RECEIVED: "response_received",
|
|
LXMRouter.PR_COMPLETE: "complete",
|
|
LXMRouter.PR_NO_PATH: "no_path",
|
|
LXMRouter.PR_LINK_FAILED: "link_failed",
|
|
LXMRouter.PR_TRANSFER_FAILED: "transfer_failed",
|
|
LXMRouter.PR_NO_IDENTITY_RCVD: "no_identity_received",
|
|
LXMRouter.PR_NO_ACCESS: "no_access",
|
|
LXMRouter.PR_FAILED: "failed",
|
|
}
|
|
|
|
# return string for state, or fallback to unknown
|
|
if state in state_map:
|
|
return state_map[state]
|
|
return "unknown"
|
|
|
|
# convert database announce to a dictionary
|
|
def convert_db_announce_to_dict(self, announce: database.Announce):
|
|
# parse display name from announce
|
|
display_name = None
|
|
if announce.aspect == "lxmf.delivery":
|
|
display_name = self.parse_lxmf_display_name(announce.app_data)
|
|
elif announce.aspect == "nomadnetwork.node":
|
|
display_name = ReticulumMeshChat.parse_nomadnetwork_node_display_name(
|
|
announce.app_data,
|
|
)
|
|
|
|
# find lxmf user icon from database
|
|
lxmf_user_icon = None
|
|
db_lxmf_user_icon = database.LxmfUserIcon.get_or_none(
|
|
database.LxmfUserIcon.destination_hash == announce.destination_hash,
|
|
)
|
|
if db_lxmf_user_icon is not None:
|
|
lxmf_user_icon = {
|
|
"icon_name": db_lxmf_user_icon.icon_name,
|
|
"foreground_colour": db_lxmf_user_icon.foreground_colour,
|
|
"background_colour": db_lxmf_user_icon.background_colour,
|
|
}
|
|
|
|
# get current hops away
|
|
hops = RNS.Transport.hops_to(bytes.fromhex(announce.destination_hash))
|
|
|
|
return {
|
|
"id": announce.id,
|
|
"destination_hash": announce.destination_hash,
|
|
"aspect": announce.aspect,
|
|
"identity_hash": announce.identity_hash,
|
|
"identity_public_key": announce.identity_public_key,
|
|
"app_data": announce.app_data,
|
|
"hops": hops,
|
|
"rssi": announce.rssi,
|
|
"snr": announce.snr,
|
|
"quality": announce.quality,
|
|
"display_name": display_name,
|
|
"custom_display_name": self.get_custom_destination_display_name(
|
|
announce.destination_hash,
|
|
),
|
|
"lxmf_user_icon": lxmf_user_icon,
|
|
"created_at": announce.created_at,
|
|
"updated_at": announce.updated_at,
|
|
}
|
|
|
|
# convert database favourite to a dictionary
|
|
@staticmethod
|
|
def convert_db_favourite_to_dict(favourite: database.FavouriteDestination):
|
|
return {
|
|
"id": favourite.id,
|
|
"destination_hash": favourite.destination_hash,
|
|
"display_name": favourite.display_name,
|
|
"aspect": favourite.aspect,
|
|
"created_at": favourite.created_at,
|
|
"updated_at": favourite.updated_at,
|
|
}
|
|
|
|
# convert database lxmf message to a dictionary
|
|
@staticmethod
|
|
def convert_db_lxmf_message_to_dict(db_lxmf_message: database.LxmfMessage):
|
|
return {
|
|
"id": db_lxmf_message.id,
|
|
"hash": db_lxmf_message.hash,
|
|
"source_hash": db_lxmf_message.source_hash,
|
|
"destination_hash": db_lxmf_message.destination_hash,
|
|
"is_incoming": db_lxmf_message.is_incoming,
|
|
"state": db_lxmf_message.state,
|
|
"progress": db_lxmf_message.progress,
|
|
"method": db_lxmf_message.method,
|
|
"delivery_attempts": db_lxmf_message.delivery_attempts,
|
|
"next_delivery_attempt_at": db_lxmf_message.next_delivery_attempt_at,
|
|
"title": db_lxmf_message.title,
|
|
"content": db_lxmf_message.content,
|
|
"fields": json.loads(db_lxmf_message.fields),
|
|
"timestamp": db_lxmf_message.timestamp,
|
|
"rssi": db_lxmf_message.rssi,
|
|
"snr": db_lxmf_message.snr,
|
|
"quality": db_lxmf_message.quality,
|
|
"is_spam": db_lxmf_message.is_spam,
|
|
"created_at": db_lxmf_message.created_at,
|
|
"updated_at": db_lxmf_message.updated_at,
|
|
}
|
|
|
|
# updates the lxmf user icon for the provided destination hash
|
|
@staticmethod
|
|
def update_lxmf_user_icon(
|
|
destination_hash: str,
|
|
icon_name: str,
|
|
foreground_colour: str,
|
|
background_colour: str,
|
|
):
|
|
# log
|
|
print(
|
|
f"updating lxmf user icon for {destination_hash} to icon_name={icon_name}, foreground_colour={foreground_colour}, background_colour={background_colour}",
|
|
)
|
|
|
|
# prepare data to insert or update
|
|
data = {
|
|
"destination_hash": destination_hash,
|
|
"icon_name": icon_name,
|
|
"foreground_colour": foreground_colour,
|
|
"background_colour": background_colour,
|
|
"updated_at": datetime.now(UTC),
|
|
}
|
|
|
|
# upsert to database
|
|
query = database.LxmfUserIcon.insert(data)
|
|
query = query.on_conflict(
|
|
conflict_target=[database.LxmfUserIcon.destination_hash],
|
|
update=data,
|
|
)
|
|
query.execute()
|
|
|
|
# check if a destination is blocked
|
|
@staticmethod
|
|
def is_destination_blocked(destination_hash: str) -> bool:
|
|
try:
|
|
blocked = database.BlockedDestination.get_or_none(
|
|
database.BlockedDestination.destination_hash == destination_hash,
|
|
)
|
|
return blocked is not None
|
|
except Exception:
|
|
return False
|
|
|
|
# check if message content matches spam keywords
|
|
@staticmethod
|
|
def check_spam_keywords(title: str, content: str) -> bool:
|
|
try:
|
|
spam_keywords = database.SpamKeyword.select()
|
|
search_text = (title + " " + content).lower()
|
|
for keyword in spam_keywords:
|
|
if keyword.keyword.lower() in search_text:
|
|
return True
|
|
return False
|
|
except Exception:
|
|
return False
|
|
|
|
# check if message has attachments and should be rejected
|
|
@staticmethod
|
|
def has_attachments(lxmf_fields: dict) -> bool:
|
|
try:
|
|
if LXMF.FIELD_FILE_ATTACHMENTS in lxmf_fields:
|
|
return len(lxmf_fields[LXMF.FIELD_FILE_ATTACHMENTS]) > 0
|
|
if LXMF.FIELD_IMAGE in lxmf_fields:
|
|
return True
|
|
if LXMF.FIELD_AUDIO in lxmf_fields:
|
|
return True
|
|
return False
|
|
except Exception:
|
|
return False
|
|
|
|
# handle an lxmf delivery from reticulum
|
|
# NOTE: cant be async, as Reticulum doesn't await it
|
|
def on_lxmf_delivery(self, lxmf_message: LXMF.LXMessage):
|
|
try:
|
|
source_hash = lxmf_message.source_hash.hex()
|
|
|
|
# check if source is blocked - reject immediately
|
|
if self.is_destination_blocked(source_hash):
|
|
print(f"Rejecting LXMF message from blocked source: {source_hash}")
|
|
return
|
|
|
|
# check if this lxmf message contains a telemetry request command from sideband
|
|
is_sideband_telemetry_request = False
|
|
lxmf_fields = lxmf_message.get_fields()
|
|
if LXMF.FIELD_COMMANDS in lxmf_fields:
|
|
for command in lxmf_fields[LXMF.FIELD_COMMANDS]:
|
|
if SidebandCommands.TELEMETRY_REQUEST in command:
|
|
is_sideband_telemetry_request = True
|
|
|
|
# ignore telemetry requests from sideband
|
|
if is_sideband_telemetry_request:
|
|
print(
|
|
"Ignoring received LXMF message as it is a telemetry request command",
|
|
)
|
|
return
|
|
|
|
# check for spam keywords
|
|
is_spam = False
|
|
message_title = lxmf_message.title if hasattr(lxmf_message, "title") else ""
|
|
message_content = (
|
|
lxmf_message.content if hasattr(lxmf_message, "content") else ""
|
|
)
|
|
|
|
# check spam keywords
|
|
if self.check_spam_keywords(message_title, message_content):
|
|
is_spam = True
|
|
print(
|
|
f"Marking LXMF message as spam due to keyword match: {source_hash}",
|
|
)
|
|
|
|
# reject attachments from blocked sources (already checked above, but double-check)
|
|
if self.has_attachments(lxmf_fields):
|
|
if self.is_destination_blocked(source_hash):
|
|
print(
|
|
f"Rejecting LXMF message with attachments from blocked source: {source_hash}",
|
|
)
|
|
return
|
|
# reject attachments from spam sources
|
|
if is_spam:
|
|
print(
|
|
f"Rejecting LXMF message with attachments from spam source: {source_hash}",
|
|
)
|
|
return
|
|
|
|
# upsert lxmf message to database with spam flag
|
|
self.db_upsert_lxmf_message(lxmf_message, is_spam=is_spam)
|
|
|
|
# update lxmf user icon if icon appearance field is available
|
|
try:
|
|
message_fields = lxmf_message.get_fields()
|
|
if LXMF.FIELD_ICON_APPEARANCE in message_fields:
|
|
icon_appearance = message_fields[LXMF.FIELD_ICON_APPEARANCE]
|
|
icon_name = icon_appearance[0]
|
|
foreground_colour = "#" + icon_appearance[1].hex()
|
|
background_colour = "#" + icon_appearance[2].hex()
|
|
self.update_lxmf_user_icon(
|
|
lxmf_message.source_hash.hex(),
|
|
icon_name,
|
|
foreground_colour,
|
|
background_colour,
|
|
)
|
|
except Exception as e:
|
|
print("failed to update lxmf user icon from lxmf message")
|
|
print(e)
|
|
|
|
# find message from database
|
|
db_lxmf_message = database.LxmfMessage.get_or_none(
|
|
database.LxmfMessage.hash == lxmf_message.hash.hex(),
|
|
)
|
|
if db_lxmf_message is None:
|
|
return
|
|
|
|
# send received lxmf message data to all websocket clients
|
|
AsyncUtils.run_async(
|
|
self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "lxmf.delivery",
|
|
"lxmf_message": self.convert_db_lxmf_message_to_dict(
|
|
db_lxmf_message,
|
|
),
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
except Exception as e:
|
|
# do nothing on error
|
|
print(f"lxmf_delivery error: {e}")
|
|
|
|
# handle delivery status update for an outbound lxmf message
|
|
def on_lxmf_sending_state_updated(self, lxmf_message):
|
|
# upsert lxmf message to database
|
|
self.db_upsert_lxmf_message(lxmf_message)
|
|
|
|
# send lxmf message state to all websocket clients
|
|
AsyncUtils.run_async(
|
|
self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "lxmf_message_state_updated",
|
|
"lxmf_message": self.convert_lxmf_message_to_dict(lxmf_message),
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# handle delivery failed for an outbound lxmf message
|
|
def on_lxmf_sending_failed(self, lxmf_message):
|
|
# check if this failed message should fall back to sending via a propagation node
|
|
if (
|
|
lxmf_message.state == LXMF.LXMessage.FAILED
|
|
and hasattr(lxmf_message, "try_propagation_on_fail")
|
|
and lxmf_message.try_propagation_on_fail
|
|
):
|
|
self.send_failed_message_via_propagation_node(lxmf_message)
|
|
|
|
# update state
|
|
self.on_lxmf_sending_state_updated(lxmf_message)
|
|
|
|
# sends a previously failed message via a propagation node
|
|
def send_failed_message_via_propagation_node(self, lxmf_message: LXMF.LXMessage):
|
|
# reset internal message state
|
|
lxmf_message.packed = None
|
|
lxmf_message.delivery_attempts = 0
|
|
if hasattr(lxmf_message, "next_delivery_attempt"):
|
|
del lxmf_message.next_delivery_attempt
|
|
|
|
# this message should now be sent via a propagation node
|
|
lxmf_message.desired_method = LXMF.LXMessage.PROPAGATED
|
|
lxmf_message.try_propagation_on_fail = False
|
|
|
|
# resend message
|
|
self.message_router.handle_outbound(lxmf_message)
|
|
|
|
# upserts the provided lxmf message to the database
|
|
def db_upsert_lxmf_message(
|
|
self,
|
|
lxmf_message: LXMF.LXMessage,
|
|
is_spam: bool = False,
|
|
):
|
|
# convert lxmf message to dict
|
|
lxmf_message_dict = self.convert_lxmf_message_to_dict(lxmf_message)
|
|
|
|
# prepare data to insert or update
|
|
data = {
|
|
"hash": lxmf_message_dict["hash"],
|
|
"source_hash": lxmf_message_dict["source_hash"],
|
|
"destination_hash": lxmf_message_dict["destination_hash"],
|
|
"is_incoming": lxmf_message_dict["is_incoming"],
|
|
"state": lxmf_message_dict["state"],
|
|
"progress": lxmf_message_dict["progress"],
|
|
"method": lxmf_message_dict["method"],
|
|
"delivery_attempts": lxmf_message_dict["delivery_attempts"],
|
|
"next_delivery_attempt_at": lxmf_message_dict["next_delivery_attempt_at"],
|
|
"title": lxmf_message_dict["title"],
|
|
"content": lxmf_message_dict["content"],
|
|
"fields": json.dumps(lxmf_message_dict["fields"]),
|
|
"timestamp": lxmf_message_dict["timestamp"],
|
|
"rssi": lxmf_message_dict["rssi"],
|
|
"snr": lxmf_message_dict["snr"],
|
|
"quality": lxmf_message_dict["quality"],
|
|
"is_spam": is_spam,
|
|
"updated_at": datetime.now(UTC),
|
|
}
|
|
|
|
# upsert to database
|
|
query = database.LxmfMessage.insert(data)
|
|
query = query.on_conflict(
|
|
conflict_target=[database.LxmfMessage.hash],
|
|
update=data,
|
|
)
|
|
query.execute()
|
|
|
|
# upserts the provided announce to the database
|
|
def db_upsert_announce(
|
|
self,
|
|
identity: RNS.Identity,
|
|
destination_hash: bytes,
|
|
aspect: str,
|
|
app_data: bytes,
|
|
announce_packet_hash: bytes,
|
|
):
|
|
# get rssi, snr and signal quality if available
|
|
rssi = self.reticulum.get_packet_rssi(announce_packet_hash)
|
|
snr = self.reticulum.get_packet_snr(announce_packet_hash)
|
|
quality = self.reticulum.get_packet_q(announce_packet_hash)
|
|
|
|
# prepare data to insert or update
|
|
data = {
|
|
"destination_hash": destination_hash.hex(),
|
|
"aspect": aspect,
|
|
"identity_hash": identity.hash.hex(),
|
|
"identity_public_key": base64.b64encode(identity.get_public_key()).decode(
|
|
"utf-8",
|
|
),
|
|
"rssi": rssi,
|
|
"snr": snr,
|
|
"quality": quality,
|
|
"updated_at": datetime.now(UTC),
|
|
}
|
|
|
|
# only set app data if provided, as we don't want to wipe existing data when we request keys from the network
|
|
if app_data is not None:
|
|
# save app data as base64 string
|
|
data["app_data"] = base64.b64encode(app_data).decode("utf-8")
|
|
|
|
# upsert to database
|
|
query = database.Announce.insert(data)
|
|
query = query.on_conflict(
|
|
conflict_target=[database.Announce.destination_hash],
|
|
update=data,
|
|
)
|
|
query.execute()
|
|
|
|
# upserts a custom destination display name to the database
|
|
@staticmethod
|
|
def db_upsert_custom_destination_display_name(
|
|
destination_hash: str,
|
|
display_name: str,
|
|
):
|
|
# prepare data to insert or update
|
|
data = {
|
|
"destination_hash": destination_hash,
|
|
"display_name": display_name,
|
|
"updated_at": datetime.now(UTC),
|
|
}
|
|
|
|
# upsert to database
|
|
query = database.CustomDestinationDisplayName.insert(data)
|
|
query = query.on_conflict(
|
|
conflict_target=[database.CustomDestinationDisplayName.destination_hash],
|
|
update=data,
|
|
)
|
|
query.execute()
|
|
|
|
# upserts a custom destination display name to the database
|
|
@staticmethod
|
|
def db_upsert_favourite(
|
|
destination_hash: str,
|
|
display_name: str,
|
|
aspect: str,
|
|
):
|
|
# prepare data to insert or update
|
|
data = {
|
|
"destination_hash": destination_hash,
|
|
"display_name": display_name,
|
|
"aspect": aspect,
|
|
"updated_at": datetime.now(UTC),
|
|
}
|
|
|
|
# upsert to database
|
|
query = database.FavouriteDestination.insert(data)
|
|
query = query.on_conflict(
|
|
conflict_target=[database.FavouriteDestination.destination_hash],
|
|
update=data,
|
|
)
|
|
query.execute()
|
|
|
|
# upserts lxmf conversation read state to the database
|
|
@staticmethod
|
|
def db_mark_lxmf_conversation_as_read(destination_hash: str):
|
|
# prepare data to insert or update
|
|
data = {
|
|
"destination_hash": destination_hash,
|
|
"last_read_at": datetime.now(UTC),
|
|
"updated_at": datetime.now(UTC),
|
|
}
|
|
|
|
# upsert to database
|
|
query = database.LxmfConversationReadState.insert(data)
|
|
query = query.on_conflict(
|
|
conflict_target=[database.LxmfConversationReadState.destination_hash],
|
|
update=data,
|
|
)
|
|
query.execute()
|
|
|
|
# handle sending an lxmf message to reticulum
|
|
async def send_message(
|
|
self,
|
|
destination_hash: str,
|
|
content: str,
|
|
image_field: LxmfImageField = None,
|
|
audio_field: LxmfAudioField = None,
|
|
file_attachments_field: LxmfFileAttachmentsField = None,
|
|
delivery_method: str = None,
|
|
) -> LXMF.LXMessage:
|
|
# convert destination hash to bytes
|
|
destination_hash = bytes.fromhex(destination_hash)
|
|
|
|
# determine when to timeout finding path
|
|
timeout_after_seconds = time.time() + 10
|
|
|
|
# check if we have a path to the destination
|
|
if not RNS.Transport.has_path(destination_hash):
|
|
# we don't have a path, so we need to request it
|
|
RNS.Transport.request_path(destination_hash)
|
|
|
|
# wait until we have a path, or give up after the configured timeout
|
|
while (
|
|
not RNS.Transport.has_path(destination_hash)
|
|
and time.time() < timeout_after_seconds
|
|
):
|
|
await asyncio.sleep(0.1)
|
|
|
|
# find destination identity from hash
|
|
destination_identity = RNS.Identity.recall(destination_hash)
|
|
if destination_identity is None:
|
|
# we have to bail out of sending, since we don't have the identity/path yet
|
|
raise Exception("Could not find path to destination. Try again later.")
|
|
|
|
# create destination for recipients lxmf delivery address
|
|
lxmf_destination = RNS.Destination(
|
|
destination_identity,
|
|
RNS.Destination.OUT,
|
|
RNS.Destination.SINGLE,
|
|
"lxmf",
|
|
"delivery",
|
|
)
|
|
|
|
# determine how the user wants to send the message
|
|
desired_delivery_method = None
|
|
if delivery_method == "direct":
|
|
desired_delivery_method = LXMF.LXMessage.DIRECT
|
|
elif delivery_method == "opportunistic":
|
|
desired_delivery_method = LXMF.LXMessage.OPPORTUNISTIC
|
|
elif delivery_method == "propagated":
|
|
desired_delivery_method = LXMF.LXMessage.PROPAGATED
|
|
|
|
# determine how to send the message if the user didn't provide a method
|
|
if desired_delivery_method is None:
|
|
# send messages over a direct link by default
|
|
desired_delivery_method = LXMF.LXMessage.DIRECT
|
|
if (
|
|
not self.message_router.delivery_link_available(destination_hash)
|
|
and RNS.Identity.current_ratchet_id(destination_hash) is not None
|
|
):
|
|
# since there's no link established to the destination, it's faster to send opportunistically
|
|
# this is because it takes several packets to establish a link, and then we still have to send the message over it
|
|
# oppotunistic mode will send the message in a single packet (if the message is small enough, otherwise it falls back to a direct link)
|
|
# we will only do this if an encryption ratchet is available, so single packet delivery is more secure
|
|
desired_delivery_method = LXMF.LXMessage.OPPORTUNISTIC
|
|
|
|
# create lxmf message
|
|
lxmf_message = LXMF.LXMessage(
|
|
lxmf_destination,
|
|
self.local_lxmf_destination,
|
|
content,
|
|
desired_method=desired_delivery_method,
|
|
)
|
|
lxmf_message.try_propagation_on_fail = (
|
|
self.config.auto_send_failed_messages_to_propagation_node.get()
|
|
)
|
|
|
|
lxmf_message.fields = {}
|
|
|
|
# add file attachments field
|
|
if file_attachments_field is not None:
|
|
# create array of [[file_name, file_bytes], [file_name, file_bytes], ...]
|
|
file_attachments = []
|
|
for file_attachment in file_attachments_field.file_attachments:
|
|
file_attachments.append(
|
|
[file_attachment.file_name, file_attachment.file_bytes],
|
|
)
|
|
|
|
# set field attachments field
|
|
lxmf_message.fields[LXMF.FIELD_FILE_ATTACHMENTS] = file_attachments
|
|
|
|
# add image field
|
|
if image_field is not None:
|
|
lxmf_message.fields[LXMF.FIELD_IMAGE] = [
|
|
image_field.image_type,
|
|
image_field.image_bytes,
|
|
]
|
|
|
|
# add audio field
|
|
if audio_field is not None:
|
|
lxmf_message.fields[LXMF.FIELD_AUDIO] = [
|
|
audio_field.audio_mode,
|
|
audio_field.audio_bytes,
|
|
]
|
|
|
|
# add icon appearance if configured
|
|
# fixme: we could save a tiny amount of bandwidth here, but this requires more effort...
|
|
# we could keep track of when the icon appearance was last sent to this destination, and when it last changed
|
|
# we could save 6 bytes for the 2x colours, and also however long the icon name is, but not today!
|
|
lxmf_user_icon_name = self.config.lxmf_user_icon_name.get()
|
|
lxmf_user_icon_foreground_colour = (
|
|
self.config.lxmf_user_icon_foreground_colour.get()
|
|
)
|
|
lxmf_user_icon_background_colour = (
|
|
self.config.lxmf_user_icon_background_colour.get()
|
|
)
|
|
if (
|
|
lxmf_user_icon_name is not None
|
|
and lxmf_user_icon_foreground_colour is not None
|
|
and lxmf_user_icon_background_colour is not None
|
|
):
|
|
lxmf_message.fields[LXMF.FIELD_ICON_APPEARANCE] = [
|
|
lxmf_user_icon_name,
|
|
ColourUtils.hex_colour_to_byte_array(lxmf_user_icon_foreground_colour),
|
|
ColourUtils.hex_colour_to_byte_array(lxmf_user_icon_background_colour),
|
|
]
|
|
|
|
# register delivery callbacks
|
|
lxmf_message.register_delivery_callback(self.on_lxmf_sending_state_updated)
|
|
lxmf_message.register_failed_callback(self.on_lxmf_sending_failed)
|
|
|
|
# send lxmf message to be routed to destination
|
|
self.message_router.handle_outbound(lxmf_message)
|
|
|
|
# upsert lxmf message to database
|
|
self.db_upsert_lxmf_message(lxmf_message)
|
|
|
|
# tell all websocket clients that old failed message was deleted so it can remove from ui
|
|
await self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "lxmf_message_created",
|
|
"lxmf_message": self.convert_lxmf_message_to_dict(lxmf_message),
|
|
},
|
|
),
|
|
)
|
|
|
|
# handle lxmf message progress loop without blocking or awaiting
|
|
# otherwise other incoming websocket packets will not be processed until sending is complete
|
|
# which results in the next message not showing up until the first message is finished
|
|
AsyncUtils.run_async(self.handle_lxmf_message_progress(lxmf_message))
|
|
|
|
return lxmf_message
|
|
|
|
# updates lxmf message in database and broadcasts to websocket until it's delivered, or it fails
|
|
async def handle_lxmf_message_progress(self, lxmf_message):
|
|
# FIXME: there's no register_progress_callback on the lxmf message, so manually send progress until delivered, propagated or failed
|
|
# we also can't use on_lxmf_sending_state_updated method to do this, because of async/await issues...
|
|
should_update_message = True
|
|
while should_update_message:
|
|
# wait 1 second between sending updates
|
|
await asyncio.sleep(1)
|
|
|
|
# upsert lxmf message to database (as we want to update the progress in database too)
|
|
self.db_upsert_lxmf_message(lxmf_message)
|
|
|
|
# send update to websocket clients
|
|
await self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "lxmf_message_state_updated",
|
|
"lxmf_message": self.convert_lxmf_message_to_dict(lxmf_message),
|
|
},
|
|
),
|
|
)
|
|
|
|
# check message state
|
|
has_delivered = lxmf_message.state == LXMF.LXMessage.DELIVERED
|
|
has_propagated = (
|
|
lxmf_message.state == LXMF.LXMessage.SENT
|
|
and lxmf_message.method == LXMF.LXMessage.PROPAGATED
|
|
)
|
|
has_failed = lxmf_message.state == LXMF.LXMessage.FAILED
|
|
is_cancelled = lxmf_message.state == LXMF.LXMessage.CANCELLED
|
|
|
|
# check if we should stop updating
|
|
if has_delivered or has_propagated or has_failed or is_cancelled:
|
|
should_update_message = False
|
|
|
|
# handle an announce received from reticulum, for an audio call address
|
|
# NOTE: cant be async, as Reticulum doesn't await it
|
|
def on_audio_call_announce_received(
|
|
self,
|
|
aspect,
|
|
destination_hash,
|
|
announced_identity,
|
|
app_data,
|
|
announce_packet_hash,
|
|
):
|
|
# check if source is blocked - drop announce and path if blocked
|
|
identity_hash = announced_identity.hash.hex()
|
|
if self.is_destination_blocked(identity_hash):
|
|
print(f"Dropping audio call announce from blocked source: {identity_hash}")
|
|
RNS.Transport.drop_path(destination_hash)
|
|
return
|
|
|
|
# log received announce
|
|
print(
|
|
"Received an announce from "
|
|
+ RNS.prettyhexrep(destination_hash)
|
|
+ " for [call.audio]",
|
|
)
|
|
|
|
# track announce timestamp
|
|
self.announce_timestamps.append(time.time())
|
|
|
|
# upsert announce to database
|
|
self.db_upsert_announce(
|
|
announced_identity,
|
|
destination_hash,
|
|
aspect,
|
|
app_data,
|
|
announce_packet_hash,
|
|
)
|
|
|
|
# find announce from database
|
|
announce = database.Announce.get_or_none(
|
|
database.Announce.destination_hash == destination_hash.hex(),
|
|
)
|
|
if announce is None:
|
|
return
|
|
|
|
# send database announce to all websocket clients
|
|
AsyncUtils.run_async(
|
|
self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "announce",
|
|
"announce": self.convert_db_announce_to_dict(announce),
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# handle an announce received from reticulum, for an lxmf address
|
|
# NOTE: cant be async, as Reticulum doesn't await it
|
|
def on_lxmf_announce_received(
|
|
self,
|
|
aspect,
|
|
destination_hash,
|
|
announced_identity,
|
|
app_data,
|
|
announce_packet_hash,
|
|
):
|
|
# check if source is blocked - drop announce and path if blocked
|
|
identity_hash = announced_identity.hash.hex()
|
|
if self.is_destination_blocked(identity_hash):
|
|
print(f"Dropping announce from blocked source: {identity_hash}")
|
|
RNS.Transport.drop_path(destination_hash)
|
|
return
|
|
|
|
# log received announce
|
|
print(
|
|
"Received an announce from "
|
|
+ RNS.prettyhexrep(destination_hash)
|
|
+ " for [lxmf.delivery]",
|
|
)
|
|
|
|
# track announce timestamp
|
|
self.announce_timestamps.append(time.time())
|
|
|
|
# upsert announce to database
|
|
self.db_upsert_announce(
|
|
announced_identity,
|
|
destination_hash,
|
|
aspect,
|
|
app_data,
|
|
announce_packet_hash,
|
|
)
|
|
|
|
# find announce from database
|
|
announce = database.Announce.get_or_none(
|
|
database.Announce.destination_hash == destination_hash.hex(),
|
|
)
|
|
if announce is None:
|
|
return
|
|
|
|
# send database announce to all websocket clients
|
|
AsyncUtils.run_async(
|
|
self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "announce",
|
|
"announce": self.convert_db_announce_to_dict(announce),
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# resend all failed messages that were intended for this destination
|
|
if self.config.auto_resend_failed_messages_when_announce_received.get():
|
|
AsyncUtils.run_async(
|
|
self.resend_failed_messages_for_destination(destination_hash.hex()),
|
|
)
|
|
|
|
# handle an announce received from reticulum, for an lxmf propagation node address
|
|
# NOTE: cant be async, as Reticulum doesn't await it
|
|
def on_lxmf_propagation_announce_received(
|
|
self,
|
|
aspect,
|
|
destination_hash,
|
|
announced_identity,
|
|
app_data,
|
|
announce_packet_hash,
|
|
):
|
|
# log received announce
|
|
print(
|
|
"Received an announce from "
|
|
+ RNS.prettyhexrep(destination_hash)
|
|
+ " for [lxmf.propagation]",
|
|
)
|
|
|
|
# track announce timestamp
|
|
self.announce_timestamps.append(time.time())
|
|
|
|
# upsert announce to database
|
|
self.db_upsert_announce(
|
|
announced_identity,
|
|
destination_hash,
|
|
aspect,
|
|
app_data,
|
|
announce_packet_hash,
|
|
)
|
|
|
|
# find announce from database
|
|
announce = database.Announce.get_or_none(
|
|
database.Announce.destination_hash == destination_hash.hex(),
|
|
)
|
|
if announce is None:
|
|
return
|
|
|
|
# send database announce to all websocket clients
|
|
AsyncUtils.run_async(
|
|
self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "announce",
|
|
"announce": self.convert_db_announce_to_dict(announce),
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# resends all messages that previously failed to send to the provided destination hash
|
|
async def resend_failed_messages_for_destination(self, destination_hash: str):
|
|
# get messages that failed to send to this destination
|
|
failed_messages = (
|
|
database.LxmfMessage.select()
|
|
.where(database.LxmfMessage.state == "failed")
|
|
.where(database.LxmfMessage.destination_hash == destination_hash)
|
|
.order_by(database.LxmfMessage.id.asc())
|
|
)
|
|
|
|
# resend failed messages
|
|
for failed_message in failed_messages:
|
|
try:
|
|
# parse fields as json
|
|
fields = json.loads(failed_message.fields)
|
|
|
|
# parse image field
|
|
image_field = None
|
|
if "image" in fields:
|
|
image_field = LxmfImageField(
|
|
fields["image"]["image_type"],
|
|
base64.b64decode(fields["image"]["image_bytes"]),
|
|
)
|
|
|
|
# parse audio field
|
|
audio_field = None
|
|
if "audio" in fields:
|
|
audio_field = LxmfAudioField(
|
|
fields["audio"]["audio_mode"],
|
|
base64.b64decode(fields["audio"]["audio_bytes"]),
|
|
)
|
|
|
|
# parse file attachments field
|
|
file_attachments_field = None
|
|
if "file_attachments" in fields:
|
|
file_attachments = []
|
|
for file_attachment in fields["file_attachments"]:
|
|
file_attachments.append(
|
|
LxmfFileAttachment(
|
|
file_attachment["file_name"],
|
|
base64.b64decode(file_attachment["file_bytes"]),
|
|
),
|
|
)
|
|
file_attachments_field = LxmfFileAttachmentsField(file_attachments)
|
|
|
|
# don't resend message with attachments if not allowed
|
|
if not self.config.allow_auto_resending_failed_messages_with_attachments.get():
|
|
if (
|
|
image_field is not None
|
|
or audio_field is not None
|
|
or file_attachments_field is not None
|
|
):
|
|
print(
|
|
"Not resending failed message with attachments, as setting is disabled",
|
|
)
|
|
continue
|
|
|
|
# send new message with failed message content
|
|
await self.send_message(
|
|
failed_message.destination_hash,
|
|
failed_message.content,
|
|
image_field,
|
|
audio_field,
|
|
file_attachments_field,
|
|
)
|
|
|
|
# remove original failed message from database
|
|
database.LxmfMessage.delete().where(
|
|
database.LxmfMessage.hash == failed_message.hash,
|
|
).execute()
|
|
|
|
# tell all websocket clients that old failed message was deleted so it can remove from ui
|
|
await self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "lxmf_message_deleted",
|
|
"hash": failed_message.hash,
|
|
},
|
|
),
|
|
)
|
|
|
|
except Exception as e:
|
|
print("Error resending failed message: " + str(e))
|
|
|
|
# handle an announce received from reticulum, for a nomadnet node
|
|
# NOTE: cant be async, as Reticulum doesn't await it
|
|
def on_nomadnet_node_announce_received(
|
|
self,
|
|
aspect,
|
|
destination_hash,
|
|
announced_identity,
|
|
app_data,
|
|
announce_packet_hash,
|
|
):
|
|
# log received announce
|
|
print(
|
|
"Received an announce from "
|
|
+ RNS.prettyhexrep(destination_hash)
|
|
+ " for [nomadnetwork.node]",
|
|
)
|
|
|
|
# track announce timestamp
|
|
self.announce_timestamps.append(time.time())
|
|
|
|
# upsert announce to database
|
|
self.db_upsert_announce(
|
|
announced_identity,
|
|
destination_hash,
|
|
aspect,
|
|
app_data,
|
|
announce_packet_hash,
|
|
)
|
|
|
|
# find announce from database
|
|
announce = database.Announce.get_or_none(
|
|
database.Announce.destination_hash == destination_hash.hex(),
|
|
)
|
|
if announce is None:
|
|
return
|
|
|
|
# send database announce to all websocket clients
|
|
AsyncUtils.run_async(
|
|
self.websocket_broadcast(
|
|
json.dumps(
|
|
{
|
|
"type": "announce",
|
|
"announce": self.convert_db_announce_to_dict(announce),
|
|
},
|
|
),
|
|
),
|
|
)
|
|
|
|
# gets the custom display name a user has set for the provided destination hash
|
|
@staticmethod
|
|
def get_custom_destination_display_name(destination_hash: str):
|
|
# get display name from database
|
|
db_destination_display_name = database.CustomDestinationDisplayName.get_or_none(
|
|
database.CustomDestinationDisplayName.destination_hash == destination_hash,
|
|
)
|
|
if db_destination_display_name is not None:
|
|
return db_destination_display_name.display_name
|
|
|
|
return None
|
|
|
|
# get name to show for an lxmf conversation
|
|
# currently, this will use the app data from the most recent announce
|
|
# TODO: we should fetch this from our contacts database, when it gets implemented, and if not found, fallback to app data
|
|
@staticmethod
|
|
def get_lxmf_conversation_name(destination_hash):
|
|
# get lxmf.delivery announce from database for the provided destination hash
|
|
lxmf_announce = (
|
|
database.Announce.select()
|
|
.where(database.Announce.aspect == "lxmf.delivery")
|
|
.where(database.Announce.destination_hash == destination_hash)
|
|
.get_or_none()
|
|
)
|
|
|
|
# if app data is available in database, it should be base64 encoded text that was announced
|
|
# we will return the parsed lxmf display name as the conversation name
|
|
if lxmf_announce is not None and lxmf_announce.app_data is not None:
|
|
return ReticulumMeshChat.parse_lxmf_display_name(
|
|
app_data_base64=lxmf_announce.app_data,
|
|
)
|
|
|
|
# announce did not have app data, so provide a fallback name
|
|
return "Anonymous Peer"
|
|
|
|
# reads the lxmf display name from the provided base64 app data
|
|
@staticmethod
|
|
def parse_lxmf_display_name(
|
|
app_data_base64: str,
|
|
default_value: str | None = "Anonymous Peer",
|
|
):
|
|
try:
|
|
app_data_bytes = base64.b64decode(app_data_base64)
|
|
display_name = LXMF.display_name_from_app_data(app_data_bytes)
|
|
if display_name is not None:
|
|
return display_name
|
|
except Exception:
|
|
pass
|
|
|
|
return default_value
|
|
|
|
# reads the lxmf stamp cost from the provided base64 app data
|
|
@staticmethod
|
|
def parse_lxmf_stamp_cost(app_data_base64: str):
|
|
try:
|
|
app_data_bytes = base64.b64decode(app_data_base64)
|
|
return LXMF.stamp_cost_from_app_data(app_data_bytes)
|
|
except Exception:
|
|
return None
|
|
|
|
# reads the nomadnetwork node display name from the provided base64 app data
|
|
@staticmethod
|
|
def parse_nomadnetwork_node_display_name(
|
|
app_data_base64: str,
|
|
default_value: str | None = "Anonymous Node",
|
|
):
|
|
try:
|
|
app_data_bytes = base64.b64decode(app_data_base64)
|
|
return app_data_bytes.decode("utf-8")
|
|
except Exception:
|
|
return default_value
|
|
|
|
# parses lxmf propagation node app data
|
|
@staticmethod
|
|
def parse_lxmf_propagation_node_app_data(app_data_base64: str):
|
|
try:
|
|
app_data_bytes = base64.b64decode(app_data_base64)
|
|
data = msgpack.unpackb(app_data_bytes)
|
|
return {
|
|
"enabled": bool(data[2]),
|
|
"timebase": int(data[1]),
|
|
"per_transfer_limit": int(data[3]),
|
|
}
|
|
except Exception:
|
|
return None
|
|
|
|
# returns true if the conversation has messages newer than the last read at timestamp
|
|
@staticmethod
|
|
def is_lxmf_conversation_unread(destination_hash):
|
|
# get lxmf conversation read state from database for the provided destination hash
|
|
lxmf_conversation_read_state = (
|
|
database.LxmfConversationReadState.select()
|
|
.where(
|
|
database.LxmfConversationReadState.destination_hash == destination_hash,
|
|
)
|
|
.get_or_none()
|
|
)
|
|
|
|
# get most recent incoming message from destination hash
|
|
latest_incoming_lxmf_message = (
|
|
database.LxmfMessage.select()
|
|
.where(database.LxmfMessage.source_hash == destination_hash)
|
|
.order_by(database.LxmfMessage.created_at.desc())
|
|
.get_or_none()
|
|
)
|
|
|
|
# there's no incoming message, so it can't be unread
|
|
if latest_incoming_lxmf_message is None:
|
|
return False
|
|
|
|
# user has never read this conversation, so it's unread
|
|
if lxmf_conversation_read_state is None:
|
|
return True
|
|
|
|
# conversation is unread if last read at is before the latest incoming message creation date
|
|
conversation_last_read_at = datetime.strptime(
|
|
lxmf_conversation_read_state.last_read_at,
|
|
"%Y-%m-%d %H:%M:%S.%f%z",
|
|
)
|
|
conversation_latest_message_at = datetime.strptime(
|
|
latest_incoming_lxmf_message.created_at,
|
|
"%Y-%m-%d %H:%M:%S.%f%z",
|
|
)
|
|
return conversation_last_read_at < conversation_latest_message_at
|
|
|
|
# returns number of messages that failed to send in a conversation
|
|
@staticmethod
|
|
def lxmf_conversation_failed_messages_count(destination_hash: str):
|
|
return (
|
|
database.LxmfMessage.select()
|
|
.where(database.LxmfMessage.state == "failed")
|
|
.where(database.LxmfMessage.destination_hash == destination_hash)
|
|
.count()
|
|
)
|
|
|
|
# find an interface by name
|
|
@staticmethod
|
|
def find_interface_by_name(name: str):
|
|
for interface in RNS.Transport.interfaces:
|
|
interface_name = str(interface)
|
|
if name == interface_name:
|
|
return interface
|
|
|
|
return None
|
|
|
|
|
|
# class to manage config stored in database
|
|
class Config:
|
|
@staticmethod
|
|
def get(key: str, default_value=None) -> str | None:
|
|
# get config value from database
|
|
config_item = database.Config.get_or_none(database.Config.key == key)
|
|
|
|
# return value if available
|
|
if config_item is not None:
|
|
return config_item.value
|
|
|
|
# fallback to returning default value
|
|
return default_value
|
|
|
|
@staticmethod
|
|
def set(key: str, value: str | None):
|
|
# if none, delete the config entry
|
|
if value is None:
|
|
database.Config.delete().where(database.Config.key == key).execute()
|
|
return
|
|
|
|
# prepare data to insert or update
|
|
data = {
|
|
"key": key,
|
|
"value": value,
|
|
"updated_at": datetime.now(UTC),
|
|
}
|
|
|
|
# upsert to database
|
|
query = database.Config.insert(data)
|
|
query = query.on_conflict(conflict_target=[database.Config.key], update=data)
|
|
query.execute()
|
|
|
|
# handle config values that should be strings
|
|
class StringConfig:
|
|
def __init__(self, key: str, default_value: str | None = None):
|
|
self.key = key
|
|
self.default_value = default_value
|
|
|
|
def get(self, default_value: str = None) -> str | None:
|
|
_default_value = default_value or self.default_value
|
|
return Config.get(self.key, default_value=_default_value)
|
|
|
|
def set(self, value: str | None):
|
|
Config.set(self.key, value)
|
|
|
|
# handle config values that should be bools
|
|
class BoolConfig:
|
|
def __init__(self, key: str, default_value: bool = False):
|
|
self.key = key
|
|
self.default_value = default_value
|
|
|
|
def get(self) -> bool:
|
|
# get string value, or return default
|
|
config_value = Config.get(self.key, default_value=None)
|
|
if config_value is None:
|
|
return self.default_value
|
|
|
|
return config_value == "true"
|
|
|
|
def set(self, value: bool):
|
|
# determine string value for bool
|
|
if value is True:
|
|
config_value = "true"
|
|
else:
|
|
config_value = "false"
|
|
|
|
Config.set(self.key, config_value)
|
|
|
|
# handle config values that should be integers
|
|
class IntConfig:
|
|
def __init__(self, key: str, default_value: int | None = 0):
|
|
self.key = key
|
|
self.default_value = default_value
|
|
|
|
def get(self) -> int | None:
|
|
# get string value, or return default
|
|
config_value = Config.get(self.key, default_value=None)
|
|
if config_value is None:
|
|
return self.default_value
|
|
|
|
return int(config_value)
|
|
|
|
def set(self, value: int):
|
|
Config.set(self.key, str(value))
|
|
|
|
# all possible config items
|
|
database_version = IntConfig("database_version", None)
|
|
display_name = StringConfig("display_name", "Anonymous Peer")
|
|
auto_announce_enabled = BoolConfig("auto_announce_enabled", False)
|
|
auto_announce_interval_seconds = IntConfig("auto_announce_interval_seconds", 0)
|
|
last_announced_at = IntConfig("last_announced_at", None)
|
|
theme = StringConfig("theme", "light")
|
|
auto_resend_failed_messages_when_announce_received = BoolConfig(
|
|
"auto_resend_failed_messages_when_announce_received",
|
|
True,
|
|
)
|
|
allow_auto_resending_failed_messages_with_attachments = BoolConfig(
|
|
"allow_auto_resending_failed_messages_with_attachments",
|
|
False,
|
|
)
|
|
auto_send_failed_messages_to_propagation_node = BoolConfig(
|
|
"auto_send_failed_messages_to_propagation_node",
|
|
False,
|
|
)
|
|
show_suggested_community_interfaces = BoolConfig(
|
|
"show_suggested_community_interfaces",
|
|
True,
|
|
)
|
|
lxmf_delivery_transfer_limit_in_bytes = IntConfig(
|
|
"lxmf_delivery_transfer_limit_in_bytes",
|
|
1000 * 1000 * 10,
|
|
) # 10MB
|
|
lxmf_preferred_propagation_node_destination_hash = StringConfig(
|
|
"lxmf_preferred_propagation_node_destination_hash",
|
|
None,
|
|
)
|
|
lxmf_preferred_propagation_node_auto_sync_interval_seconds = IntConfig(
|
|
"lxmf_preferred_propagation_node_auto_sync_interval_seconds",
|
|
0,
|
|
)
|
|
lxmf_preferred_propagation_node_last_synced_at = IntConfig(
|
|
"lxmf_preferred_propagation_node_last_synced_at",
|
|
None,
|
|
)
|
|
lxmf_local_propagation_node_enabled = BoolConfig(
|
|
"lxmf_local_propagation_node_enabled",
|
|
False,
|
|
)
|
|
lxmf_user_icon_name = StringConfig("lxmf_user_icon_name", None)
|
|
lxmf_user_icon_foreground_colour = StringConfig(
|
|
"lxmf_user_icon_foreground_colour",
|
|
None,
|
|
)
|
|
lxmf_user_icon_background_colour = StringConfig(
|
|
"lxmf_user_icon_background_colour",
|
|
None,
|
|
)
|
|
lxmf_inbound_stamp_cost = IntConfig(
|
|
"lxmf_inbound_stamp_cost",
|
|
8,
|
|
) # for direct delivery messages
|
|
lxmf_propagation_node_stamp_cost = IntConfig(
|
|
"lxmf_propagation_node_stamp_cost",
|
|
16,
|
|
) # for propagation node messages
|
|
|
|
|
|
# FIXME: we should probably set this as an instance variable of ReticulumMeshChat so it has a proper home, and pass it in to the constructor?
|
|
nomadnet_cached_links = {}
|
|
|
|
|
|
class NomadnetDownloader:
|
|
def __init__(
|
|
self,
|
|
destination_hash: bytes,
|
|
path: str,
|
|
data: str | None,
|
|
on_download_success: Callable[[RNS.RequestReceipt], None],
|
|
on_download_failure: Callable[[str], None],
|
|
on_progress_update: Callable[[float], None],
|
|
timeout: int | None = None,
|
|
):
|
|
self.app_name = "nomadnetwork"
|
|
self.aspects = "node"
|
|
self.destination_hash = destination_hash
|
|
self.path = path
|
|
self.data = data
|
|
self.timeout = timeout
|
|
self._download_success_callback = on_download_success
|
|
self._download_failure_callback = on_download_failure
|
|
self.on_progress_update = on_progress_update
|
|
self.request_receipt = None
|
|
self.is_cancelled = False
|
|
self.link = None
|
|
|
|
# cancel the download
|
|
def cancel(self):
|
|
self.is_cancelled = True
|
|
|
|
# cancel the request if it exists
|
|
if self.request_receipt is not None:
|
|
try:
|
|
self.request_receipt.cancel()
|
|
except Exception:
|
|
pass
|
|
|
|
# clean up the link if we created it
|
|
if self.link is not None:
|
|
try:
|
|
self.link.teardown()
|
|
except Exception:
|
|
pass
|
|
|
|
# notify that download was cancelled
|
|
self._download_failure_callback("cancelled")
|
|
|
|
# setup link to destination and request download
|
|
async def download(
|
|
self,
|
|
path_lookup_timeout: int = 15,
|
|
link_establishment_timeout: int = 15,
|
|
):
|
|
# check if cancelled before starting
|
|
if self.is_cancelled:
|
|
return
|
|
|
|
# use existing established link if it's active
|
|
if self.destination_hash in nomadnet_cached_links:
|
|
link = nomadnet_cached_links[self.destination_hash]
|
|
if link.status is RNS.Link.ACTIVE:
|
|
print("[NomadnetDownloader] using existing link for request")
|
|
self.link_established(link)
|
|
return
|
|
|
|
# determine when to timeout
|
|
timeout_after_seconds = time.time() + path_lookup_timeout
|
|
|
|
# check if we have a path to the destination
|
|
if not RNS.Transport.has_path(self.destination_hash):
|
|
# we don't have a path, so we need to request it
|
|
RNS.Transport.request_path(self.destination_hash)
|
|
|
|
# wait until we have a path, or give up after the configured timeout
|
|
while (
|
|
not RNS.Transport.has_path(self.destination_hash)
|
|
and time.time() < timeout_after_seconds
|
|
):
|
|
# check if cancelled during path lookup
|
|
if self.is_cancelled:
|
|
return
|
|
await asyncio.sleep(0.1)
|
|
|
|
# if we still don't have a path, we can't establish a link, so bail out
|
|
if not RNS.Transport.has_path(self.destination_hash):
|
|
self._download_failure_callback("Could not find path to destination.")
|
|
return
|
|
|
|
# check if cancelled before establishing link
|
|
if self.is_cancelled:
|
|
return
|
|
|
|
# create destination to nomadnet node
|
|
identity = RNS.Identity.recall(self.destination_hash)
|
|
destination = RNS.Destination(
|
|
identity,
|
|
RNS.Destination.OUT,
|
|
RNS.Destination.SINGLE,
|
|
self.app_name,
|
|
self.aspects,
|
|
)
|
|
|
|
# create link to destination
|
|
print("[NomadnetDownloader] establishing new link for request")
|
|
link = RNS.Link(destination, established_callback=self.link_established)
|
|
self.link = link
|
|
|
|
# determine when to timeout
|
|
timeout_after_seconds = time.time() + link_establishment_timeout
|
|
|
|
# wait until we have established a link, or give up after the configured timeout
|
|
while (
|
|
link.status is not RNS.Link.ACTIVE and time.time() < timeout_after_seconds
|
|
):
|
|
# check if cancelled during link establishment
|
|
if self.is_cancelled:
|
|
return
|
|
await asyncio.sleep(0.1)
|
|
|
|
# if we still haven't established a link, bail out
|
|
if link.status is not RNS.Link.ACTIVE:
|
|
self._download_failure_callback("Could not establish link to destination.")
|
|
|
|
# link to destination was established, we should now request the download
|
|
def link_established(self, link):
|
|
# check if cancelled before requesting
|
|
if self.is_cancelled:
|
|
return
|
|
|
|
# cache link for using in future requests
|
|
nomadnet_cached_links[self.destination_hash] = link
|
|
|
|
# request download over link
|
|
self.request_receipt = link.request(
|
|
self.path,
|
|
data=self.data,
|
|
response_callback=self.on_response,
|
|
failed_callback=self.on_failed,
|
|
progress_callback=self.on_progress,
|
|
timeout=self.timeout,
|
|
)
|
|
|
|
# handle successful download
|
|
def on_response(self, request_receipt: RNS.RequestReceipt):
|
|
self._download_success_callback(request_receipt)
|
|
|
|
# handle failure
|
|
def on_failed(self, request_receipt=None):
|
|
self._download_failure_callback("request_failed")
|
|
|
|
# handle download progress
|
|
def on_progress(self, request_receipt):
|
|
self.on_progress_update(request_receipt.progress)
|
|
|
|
|
|
class NomadnetPageDownloader(NomadnetDownloader):
|
|
def __init__(
|
|
self,
|
|
destination_hash: bytes,
|
|
page_path: str,
|
|
data: str | None,
|
|
on_page_download_success: Callable[[str], None],
|
|
on_page_download_failure: Callable[[str], None],
|
|
on_progress_update: Callable[[float], None],
|
|
timeout: int | None = None,
|
|
):
|
|
self.on_page_download_success = on_page_download_success
|
|
self.on_page_download_failure = on_page_download_failure
|
|
super().__init__(
|
|
destination_hash,
|
|
page_path,
|
|
data,
|
|
self.on_download_success,
|
|
self.on_download_failure,
|
|
on_progress_update,
|
|
timeout,
|
|
)
|
|
|
|
# page download was successful, decode the response and send to provided callback
|
|
def on_download_success(self, request_receipt: RNS.RequestReceipt):
|
|
micron_markup_response = request_receipt.response.decode("utf-8")
|
|
self.on_page_download_success(micron_markup_response)
|
|
|
|
# page download failed, send error to provided callback
|
|
def on_download_failure(self, failure_reason):
|
|
self.on_page_download_failure(failure_reason)
|
|
|
|
|
|
class NomadnetFileDownloader(NomadnetDownloader):
|
|
def __init__(
|
|
self,
|
|
destination_hash: bytes,
|
|
page_path: str,
|
|
on_file_download_success: Callable[[str, bytes], None],
|
|
on_file_download_failure: Callable[[str], None],
|
|
on_progress_update: Callable[[float], None],
|
|
timeout: int | None = None,
|
|
):
|
|
self.on_file_download_success = on_file_download_success
|
|
self.on_file_download_failure = on_file_download_failure
|
|
super().__init__(
|
|
destination_hash,
|
|
page_path,
|
|
None,
|
|
self.on_download_success,
|
|
self.on_download_failure,
|
|
on_progress_update,
|
|
timeout,
|
|
)
|
|
|
|
# file download was successful, decode the response and send to provided callback
|
|
def on_download_success(self, request_receipt: RNS.RequestReceipt):
|
|
# get response
|
|
response = request_receipt.response
|
|
|
|
# handle buffered reader response
|
|
if isinstance(response, io.BufferedReader):
|
|
# get file name from metadata
|
|
file_name = "downloaded_file"
|
|
metadata = request_receipt.metadata
|
|
if metadata is not None and "name" in metadata:
|
|
file_path = metadata["name"].decode("utf-8")
|
|
file_name = os.path.basename(file_path)
|
|
|
|
# get file data
|
|
file_data: bytes = response.read()
|
|
|
|
self.on_file_download_success(file_name, file_data)
|
|
return
|
|
|
|
# check for list response with bytes in position 0, and metadata dict in position 1
|
|
# e.g: [file_bytes, {name: "filename.ext"}]
|
|
if isinstance(response, list) and isinstance(response[1], dict):
|
|
file_data: bytes = response[0]
|
|
metadata: dict = response[1]
|
|
|
|
# get file name from metadata
|
|
file_name = "downloaded_file"
|
|
if metadata is not None and "name" in metadata:
|
|
file_path = metadata["name"].decode("utf-8")
|
|
file_name = os.path.basename(file_path)
|
|
|
|
self.on_file_download_success(file_name, file_data)
|
|
return
|
|
|
|
# try using original response format
|
|
# unsure if this is actually used anymore now that a buffered reader is provided
|
|
# have left here just in case...
|
|
try:
|
|
file_name: str = response[0]
|
|
file_data: bytes = response[1]
|
|
self.on_file_download_success(file_name, file_data)
|
|
except Exception:
|
|
self.on_download_failure("unsupported_response")
|
|
|
|
# page download failed, send error to provided callback
|
|
def on_download_failure(self, failure_reason):
|
|
self.on_file_download_failure(failure_reason)
|
|
|
|
|
|
def main():
|
|
# parse command line args
|
|
parser = argparse.ArgumentParser(description="ReticulumMeshChat")
|
|
parser.add_argument(
|
|
"--host",
|
|
nargs="?",
|
|
default="127.0.0.1",
|
|
type=str,
|
|
help="The address the web server should listen on.",
|
|
)
|
|
parser.add_argument(
|
|
"--port",
|
|
nargs="?",
|
|
default="8000",
|
|
type=int,
|
|
help="The port the web server should listen on.",
|
|
)
|
|
parser.add_argument(
|
|
"--headless",
|
|
action="store_true",
|
|
help="Web browser will not automatically launch when this flag is passed.",
|
|
)
|
|
parser.add_argument(
|
|
"--identity-file",
|
|
type=str,
|
|
help="Path to a Reticulum Identity file to use as your LXMF address.",
|
|
)
|
|
parser.add_argument(
|
|
"--identity-base64",
|
|
type=str,
|
|
help="A base64 encoded Reticulum Identity to use as your LXMF address.",
|
|
)
|
|
parser.add_argument(
|
|
"--generate-identity-file",
|
|
type=str,
|
|
help="Generates and saves a new Reticulum Identity to the provided file path and then exits.",
|
|
)
|
|
parser.add_argument(
|
|
"--generate-identity-base64",
|
|
action="store_true",
|
|
help="Outputs a randomly generated Reticulum Identity as base64 and then exits.",
|
|
)
|
|
parser.add_argument(
|
|
"--reticulum-config-dir",
|
|
type=str,
|
|
help="Path to a Reticulum config directory for the RNS stack to use (e.g: ~/.reticulum)",
|
|
)
|
|
parser.add_argument(
|
|
"--storage-dir",
|
|
type=str,
|
|
help="Path to a directory for storing databases and config files (default: ./storage)",
|
|
)
|
|
parser.add_argument(
|
|
"--test-exception-message",
|
|
type=str,
|
|
help="Throws an exception. Used for testing the electron error dialog",
|
|
)
|
|
parser.add_argument(
|
|
"args",
|
|
nargs=argparse.REMAINDER,
|
|
) # allow unknown command line args
|
|
args = parser.parse_args()
|
|
|
|
# check if we want to test exception messages
|
|
if args.test_exception_message is not None:
|
|
raise Exception(args.test_exception_message)
|
|
|
|
# util to generate reticulum identity and save to file without using rnid
|
|
if args.generate_identity_file is not None:
|
|
# do not overwrite existing files, otherwise user could lose existing keys
|
|
if os.path.exists(args.generate_identity_file):
|
|
print(
|
|
"DANGER: the provided identity file path already exists, not overwriting!",
|
|
)
|
|
return
|
|
|
|
# generate a new identity and save to provided file path
|
|
identity = RNS.Identity(create_keys=True)
|
|
with open(args.generate_identity_file, "wb") as file:
|
|
file.write(identity.get_private_key())
|
|
|
|
print(
|
|
f"A new Reticulum Identity has been saved to: {args.generate_identity_file}",
|
|
)
|
|
return
|
|
|
|
# util to generate reticulum identity as base64 without using rnid
|
|
if args.generate_identity_base64 is True:
|
|
identity = RNS.Identity(create_keys=True)
|
|
print(base64.b64encode(identity.get_private_key()).decode("utf-8"))
|
|
return
|
|
|
|
# use provided identity, or fallback to a random one
|
|
if args.identity_file is not None:
|
|
identity = RNS.Identity(create_keys=False)
|
|
identity.load(args.identity_file)
|
|
print(
|
|
f"Reticulum Identity <{identity.hash.hex()}> has been loaded from file {args.identity_file}.",
|
|
)
|
|
elif args.identity_base64 is not None:
|
|
identity = RNS.Identity(create_keys=False)
|
|
identity.load_private_key(base64.b64decode(args.identity_base64))
|
|
print(
|
|
f"Reticulum Identity <{identity.hash.hex()}> has been loaded from base64.",
|
|
)
|
|
else:
|
|
# ensure provided storage dir exists, or the default storage dir exists
|
|
base_storage_dir = args.storage_dir or os.path.join("storage")
|
|
os.makedirs(base_storage_dir, exist_ok=True)
|
|
|
|
# configure path to default identity file
|
|
default_identity_file = os.path.join(base_storage_dir, "identity")
|
|
|
|
# if default identity file does not exist, generate a new identity and save it
|
|
if not os.path.exists(default_identity_file):
|
|
identity = RNS.Identity(create_keys=True)
|
|
with open(default_identity_file, "wb") as file:
|
|
file.write(identity.get_private_key())
|
|
print(
|
|
f"Reticulum Identity <{identity.hash.hex()}> has been randomly generated and saved to {default_identity_file}.",
|
|
)
|
|
|
|
# default identity file exists, load it
|
|
identity = RNS.Identity(create_keys=False)
|
|
identity.load(default_identity_file)
|
|
print(
|
|
f"Reticulum Identity <{identity.hash.hex()}> has been loaded from file {default_identity_file}.",
|
|
)
|
|
|
|
# init app
|
|
reticulum_meshchat = ReticulumMeshChat(
|
|
identity,
|
|
args.storage_dir,
|
|
args.reticulum_config_dir,
|
|
)
|
|
reticulum_meshchat.run(args.host, args.port, launch_browser=args.headless is False)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
main()
|