2023-09-13 Alarm Watchdog Documentation

 

Drone operation

Drone operation is based on ROS-1 architecture. It consist of topics, like speed, battery, odometry that are constantly emitted by drone and ROS actions like startMapping, flyToPOI, land etc. which have continous control in between nodes. “Exit node” currently is rosbridge that exposes ROS infrastructure to websocket for use in TypeScript based react-native app.

image-20230913122646670

 

Current App architecture

Drone
ROS-1
ROS Node ADC control
ROS Node mapping
RosBridge WebSocket API
RTSP stream
React-Native App
AI map backend processing API

Green nodes - ASYA team Red nodes - Alarm Watchdog team Blue nodes - Alarm app team

Proposed app architecture - A

Drone
vpn
vpn
ROS-1
ROS Node ADC control
ROS Node mapping
Scheduling service, Python
Custom control API, Python FastAPI
Fusion Streaming Service
Alarm App Server
Alarm Native Apps
AI map backend processing API

Green nodes - ASYA team Red nodes - Alarm Watchdog team Blue nodes - Alarm app team

Need to resolve interrupts for schefulling service - wake on LAN, time, sound

 

Proposed app architecture - B

Advantage: AI processing client-side, more control over drone, better battery life Disadvantage: More cost to client-side

Drone
Companion Server
vpn
ROS-1
ROS Node ADC control
ROS Node mapping
Custom control API, Python FastAPI
Fusion Streaming Service
Scheduling service, Python
AI map backend processing API
Alarm App Server
Alarm Native Apps

 

Green nodes - ASYA team Red nodes - Alarm Watchdog team Blue nodes - Alarm app team

 

Github / Figma

React-Native mobile app https://github.com/asya-ai/alarm_watchdog_react_native_mobile_app

AI backend processing server (FastAPI Swagger UI REST API) https://github.com/asya-ai/alarm_flight_model

Figma app designs (whole history of UX experiments and iterations) https://www.figma.com/file/NQicYun9EL9JCDF2yTO4rg/Alarm?type=design&node-id=1479-47358&mode=design&t=bIzK3oygncZTw1JX-0

Backend map processing

LIDAR on drone and SLAM produces very low quality data, often with large missing pieces and in arbritary rotation and scaling. AI based generative models allow to fix point cloud and transformation, it also produces enhanced drone path, walls as vectors, rooms as polygons, transformation matrices Notice significant differences between map orientation and detail between offline and online version

Offline mode / No transformationsBackend mode / With transformations
image-20230913101910777image-20230913102108235
image-20230913105320716image-20230913105355980

 

Offline mode

Offline mode can be set in Settings, it allows to use app directly with drone without any internet connection. This setting most likely would not be used for end-user experience, but just as backup for testing. When in offline mode map will not be processed in backend server and will just operate with raw point cloud. Beaware that whole point cloud should not be drawn on screen as it contains too many points and will cause rendering to be slow and app to be unresponsive.

 

App operation - Mapping

image-20230913124618182

Sepecific procedure to walk tgrough the rooms to collect point-cloud data.

Afterwards sent to AI backend server for processing.

image-20230913124726240

Must be able to store multiple maps and different missions for each map.

image-20230913125620912

After processing it should be possible to distinguish rooms in each map as polygons

image-20230913125353495

 

App operation - Control

Functions to control drone: Take off, Land in base, Emergency land, Rotate in flight, as well as add POIs and fly to POIs. See camera feed in real time and other features.

image-20230913124850306

App operation - Missions

Functions to drag and drop POIs in sequence, set different settings like sleep in each POI, loops etc. To automatically execute flight starting from flight to landing and collect photos from stream.

image-20230913125427535

App operation - Scheduled Missions

Not yet decided on implementation. Will need service on drone or comapnion server.

Testing without drone

We have deployed in-cloud ROS node for testing purposes to simulate drone environment. In order to run it you must be sure that rosbridge screen is open and running ( ⚠️ do not close any other screens because on the backend map processing nodes are running )

image-20230913104802723

SSH connection h: 144.24.250.72 u: ubuntu key: [will be sent by Evalds]

 

Running bag file (for mapping or control screen simulation)

Choose bag recording from

Run bag recording in replay mode

^ Beaware that bags files contain only part of ROS topics and usually they are incomplete, it might be necessary to run different bag files to test all functions

Troubleshooting

Sometimes if bag file data is not reaching app it might be that rosbridge node is not working properly then you need to restart it. In order to do it you must first open screen, kill node and then restart it

Connect to screen (if you are not familiar how to work with screens, read carefully below section “Linux Screens”)

Kill rosbridge

Restart

 

Testing with drone

To test with drone, first connect to local router where drone is connected then in Settings enter ROS IP to local IP address of drone. Local router should also have access to internet and should work with Backend processing server which is deployed on ROS testing node. If no internet connectivity then Offline mode should be used.

 

Linux Screen

create with name screen -S newname

List screens: screen -ls

Attach to existing screen -rd newname

Detach from screen ctrl+a => d

Kill screen ctrl+a => k

 

Scroll mode

image-20211222182726250

Screen shortcuts

458e4480-d0da-44e5-9656-981be1074ffa

Send commands to screen: