Compare commits
59 Commits
lab5/koval
...
project/sh
| Author | SHA1 | Date | |
|---|---|---|---|
| 119547d288 | |||
| b58167f0de | |||
| 121bd007b3 | |||
| db63eb6d79 | |||
| 77d6968297 | |||
| 0c2392dc0b | |||
|
|
65f767d38e | ||
| 0695e3d092 | |||
|
|
d6e094e6c0 | ||
|
|
2167eb2960 | ||
|
|
38374a6723 | ||
| c08612f71a | |||
| bde51ca5e1 | |||
| a204bb1676 | |||
| 764fb77f27 | |||
|
|
a55fc17711 | ||
| b34e385128 | |||
| a8a0ef5e15 | |||
| 00b037a243 | |||
| d1b6c0eed1 | |||
| 5e890d4f03 | |||
| a8e50d0386 | |||
| 1b42be264d | |||
|
|
b12bdc334c | ||
|
|
e8ff1c6cbd | ||
|
|
ad70519f47 | ||
|
|
b10aec1020 | ||
|
|
c085a49c8c | ||
| 0b8d2eb18b | |||
| 2846130e4e | |||
| 30af132033 | |||
| 60a846d8b8 | |||
| fe6bb6ab3a | |||
|
|
30f81ec1ae | ||
| 1b6f47fa0d | |||
| b1e6ad7c94 | |||
|
|
1eddfd966b | ||
| 8af68d6dd9 | |||
| 63aca15824 | |||
| ee509f72a4 | |||
| da9fe69d4e | |||
| 1c856dca0e | |||
|
|
17738d07fe | ||
|
|
6b5831ff1b | ||
|
|
54505db70e | ||
|
|
6f4b3b0ea6 | ||
|
|
948a936a1f | ||
|
|
87facff668 | ||
|
|
294ed5958e | ||
|
|
cbdf81c028 | ||
|
|
a98223cbb0 | ||
|
|
0bd3eaa91d | ||
|
|
9bdb98c19b | ||
|
|
c3b71daaeb | ||
|
|
31c760f4a8 | ||
|
|
24aeb1a19f | ||
|
|
ceffcfeac2 | ||
|
|
312177e087 | ||
|
|
f96930a259 |
23
.dockerignore
Normal file
23
.dockerignore
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
# IDEs
|
||||||
|
.idea/
|
||||||
|
.vscode/
|
||||||
|
.git/
|
||||||
|
.gitignore
|
||||||
|
.dockerignore
|
||||||
|
.DS_Store
|
||||||
|
Thumbs.db
|
||||||
|
|
||||||
|
# Python
|
||||||
|
**/__pycache__/
|
||||||
|
**/*.py[cod]
|
||||||
|
**/*.pyo
|
||||||
|
**/*.pyd
|
||||||
|
venv/
|
||||||
|
.env
|
||||||
|
|
||||||
|
# Logs & Database & Broker data
|
||||||
|
*.log
|
||||||
|
**/mosquitto/data/
|
||||||
|
**/mosquitto/log/
|
||||||
|
**/postgres_data/
|
||||||
|
**/pgadmin-data/
|
||||||
16
.gitea/workflows/reset-docker.yaml
Normal file
16
.gitea/workflows/reset-docker.yaml
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
name: Reset docker state
|
||||||
|
on: workflow_dispatch
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
reset:
|
||||||
|
runs-on: host-arch-x86_64
|
||||||
|
name: Reset docker state
|
||||||
|
steps:
|
||||||
|
- name: Stop all containers
|
||||||
|
run: docker stop $(docker ps -a | cut -d " " -f 1 | tail -n +2)
|
||||||
|
|
||||||
|
- name: Remove all containers
|
||||||
|
run: docker rm $(docker ps -a | cut -d " " -f 1 | tail -n +2)
|
||||||
|
|
||||||
|
- name: Remove extra volumes
|
||||||
|
run: docker volume rm road_vision_postgres_data road_vision_pgadmin-data
|
||||||
71
.gitea/workflows/tests.yaml
Normal file
71
.gitea/workflows/tests.yaml
Normal file
@@ -0,0 +1,71 @@
|
|||||||
|
name: Component testing
|
||||||
|
on: [push, workflow_dispatch]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
hub-test:
|
||||||
|
name: Hub testing
|
||||||
|
runs-on: host-arch-x86_64
|
||||||
|
steps:
|
||||||
|
- name: Clone repository
|
||||||
|
run: git clone --revision ${{ gitea.sha }} --depth 1 ${{ gitea.server_url }}/${{ gitea.repository }}
|
||||||
|
|
||||||
|
- name: Build Hub testing container
|
||||||
|
working-directory: IoT-Systems
|
||||||
|
run: docker build -t local/hub/${{gitea.sha}} -f hub/Dockerfile-test .
|
||||||
|
|
||||||
|
- name: Run Hub tests
|
||||||
|
working-directory: IoT-Systems
|
||||||
|
run: docker run --rm -it local/hub/${{gitea.sha}}
|
||||||
|
|
||||||
|
- name: Clean up containers
|
||||||
|
if: ${{always()}}
|
||||||
|
run: docker image rm local/hub/${{gitea.sha}}
|
||||||
|
|
||||||
|
store-test:
|
||||||
|
name: Store testing
|
||||||
|
runs-on: host-arch-x86_64
|
||||||
|
steps:
|
||||||
|
- name: Clone repository
|
||||||
|
run: git clone --revision ${{ gitea.sha }} --depth 1 ${{ gitea.server_url }}/${{ gitea.repository }}
|
||||||
|
|
||||||
|
- name: Build Store testing container
|
||||||
|
working-directory: IoT-Systems
|
||||||
|
run: docker build -t local/store/${{gitea.sha}} -f store/Dockerfile-test .
|
||||||
|
|
||||||
|
- name: Run Store tests
|
||||||
|
working-directory: IoT-Systems
|
||||||
|
run: docker run --rm -it local/store/${{gitea.sha}}
|
||||||
|
|
||||||
|
- name: Clean up containers
|
||||||
|
if: ${{always()}}
|
||||||
|
run: docker image rm local/store/${{gitea.sha}}
|
||||||
|
|
||||||
|
integration-smoke-test:
|
||||||
|
name: Integration smoke testing
|
||||||
|
runs-on: host-arch-x86_64
|
||||||
|
needs:
|
||||||
|
- hub-test
|
||||||
|
- store-test
|
||||||
|
steps:
|
||||||
|
- name: Clone repository
|
||||||
|
run: git clone --revision ${{ gitea.sha }} --depth 1 ${{ gitea.server_url }}/${{ gitea.repository }}
|
||||||
|
|
||||||
|
- name: Build all production containers
|
||||||
|
working-directory: IoT-Systems
|
||||||
|
run: docker-compose build
|
||||||
|
|
||||||
|
- name: Start all production containers
|
||||||
|
working-directory: IoT-Systems
|
||||||
|
run: docker-compose up -d
|
||||||
|
|
||||||
|
- name: Wait for crashes to happen
|
||||||
|
run: sleep 30
|
||||||
|
|
||||||
|
- name: Check for dead containers
|
||||||
|
working-directory: IoT-Systems
|
||||||
|
run: docker ps -a | python3 utils/check-up.py
|
||||||
|
|
||||||
|
- name: Clean up
|
||||||
|
if: ${{always()}}
|
||||||
|
working-directory: IoT-Systems
|
||||||
|
run: docker-compose down -v
|
||||||
27
.gitignore
vendored
27
.gitignore
vendored
@@ -1,4 +1,25 @@
|
|||||||
agent/docker/mosquitto/data/
|
# IDEs
|
||||||
agent/docker/mosquitto/log/
|
.idea/
|
||||||
|
.vscode/
|
||||||
|
*.swp
|
||||||
|
*.swo
|
||||||
|
|
||||||
.idea/
|
# Python
|
||||||
|
venv/
|
||||||
|
__pycache__/
|
||||||
|
*.py[cod]
|
||||||
|
*$py.class
|
||||||
|
.env
|
||||||
|
|
||||||
|
# Logs
|
||||||
|
*.log
|
||||||
|
app.log
|
||||||
|
|
||||||
|
# Database & Broker data
|
||||||
|
**/mosquitto/data/
|
||||||
|
**/mosquitto/log/
|
||||||
|
**/postgres_data/
|
||||||
|
**/pgadmin-data/
|
||||||
|
|
||||||
|
# OS specific
|
||||||
|
.DS_Store
|
||||||
3
.idea/.gitignore
generated
vendored
3
.idea/.gitignore
generated
vendored
@@ -1,3 +0,0 @@
|
|||||||
# Default ignored files
|
|
||||||
/shelf/
|
|
||||||
/workspace.xml
|
|
||||||
15
.idea/IoT-Systems.iml
generated
15
.idea/IoT-Systems.iml
generated
@@ -1,15 +0,0 @@
|
|||||||
<?xml version="1.0" encoding="UTF-8"?>
|
|
||||||
<module type="PYTHON_MODULE" version="4">
|
|
||||||
<component name="NewModuleRootManager">
|
|
||||||
<content url="file://$MODULE_DIR$">
|
|
||||||
<sourceFolder url="file://$MODULE_DIR$/store" isTestSource="false" />
|
|
||||||
<excludeFolder url="file://$MODULE_DIR$/.venv" />
|
|
||||||
</content>
|
|
||||||
<orderEntry type="jdk" jdkName="Python 3.12 (IoT-Systems)" jdkType="Python SDK" />
|
|
||||||
<orderEntry type="sourceFolder" forTests="false" />
|
|
||||||
</component>
|
|
||||||
<component name="PyDocumentationSettings">
|
|
||||||
<option name="format" value="PLAIN" />
|
|
||||||
<option name="myDocStringFormat" value="Plain" />
|
|
||||||
</component>
|
|
||||||
</module>
|
|
||||||
425
.idea/dbnavigator.xml
generated
425
.idea/dbnavigator.xml
generated
@@ -1,425 +0,0 @@
|
|||||||
<?xml version="1.0" encoding="UTF-8"?>
|
|
||||||
<project version="4">
|
|
||||||
<component name="DBNavigator.Project.DDLFileAttachmentManager">
|
|
||||||
<mappings />
|
|
||||||
<preferences />
|
|
||||||
</component>
|
|
||||||
<component name="DBNavigator.Project.DatabaseAssistantManager">
|
|
||||||
<assistants />
|
|
||||||
</component>
|
|
||||||
<component name="DBNavigator.Project.DatabaseFileManager">
|
|
||||||
<open-files />
|
|
||||||
</component>
|
|
||||||
<component name="DBNavigator.Project.Settings">
|
|
||||||
<connections />
|
|
||||||
<browser-settings>
|
|
||||||
<general>
|
|
||||||
<display-mode value="TABBED" />
|
|
||||||
<navigation-history-size value="100" />
|
|
||||||
<show-object-details value="false" />
|
|
||||||
<enable-sticky-paths value="true" />
|
|
||||||
</general>
|
|
||||||
<filters>
|
|
||||||
<object-type-filter>
|
|
||||||
<object-type name="SCHEMA" enabled="true" />
|
|
||||||
<object-type name="USER" enabled="true" />
|
|
||||||
<object-type name="ROLE" enabled="true" />
|
|
||||||
<object-type name="PRIVILEGE" enabled="true" />
|
|
||||||
<object-type name="CHARSET" enabled="true" />
|
|
||||||
<object-type name="TABLE" enabled="true" />
|
|
||||||
<object-type name="VIEW" enabled="true" />
|
|
||||||
<object-type name="MATERIALIZED_VIEW" enabled="true" />
|
|
||||||
<object-type name="NESTED_TABLE" enabled="true" />
|
|
||||||
<object-type name="COLUMN" enabled="true" />
|
|
||||||
<object-type name="INDEX" enabled="true" />
|
|
||||||
<object-type name="CONSTRAINT" enabled="true" />
|
|
||||||
<object-type name="DATASET_TRIGGER" enabled="true" />
|
|
||||||
<object-type name="DATABASE_TRIGGER" enabled="true" />
|
|
||||||
<object-type name="SYNONYM" enabled="true" />
|
|
||||||
<object-type name="SEQUENCE" enabled="true" />
|
|
||||||
<object-type name="PROCEDURE" enabled="true" />
|
|
||||||
<object-type name="FUNCTION" enabled="true" />
|
|
||||||
<object-type name="PACKAGE" enabled="true" />
|
|
||||||
<object-type name="TYPE" enabled="true" />
|
|
||||||
<object-type name="TYPE_ATTRIBUTE" enabled="true" />
|
|
||||||
<object-type name="ARGUMENT" enabled="true" />
|
|
||||||
<object-type name="JAVA_CLASS" enabled="true" />
|
|
||||||
<object-type name="JAVA_INNER_CLASS" enabled="true" />
|
|
||||||
<object-type name="JAVA_FIELD" enabled="true" />
|
|
||||||
<object-type name="JAVA_METHOD" enabled="true" />
|
|
||||||
<object-type name="DIMENSION" enabled="true" />
|
|
||||||
<object-type name="CLUSTER" enabled="true" />
|
|
||||||
<object-type name="DBLINK" enabled="true" />
|
|
||||||
<object-type name="CREDENTIAL" enabled="true" />
|
|
||||||
<object-type name="AI_PROFILE" enabled="true" />
|
|
||||||
</object-type-filter>
|
|
||||||
</filters>
|
|
||||||
<sorting>
|
|
||||||
<object-type name="COLUMN" sorting-type="NAME" />
|
|
||||||
<object-type name="FUNCTION" sorting-type="NAME" />
|
|
||||||
<object-type name="PROCEDURE" sorting-type="NAME" />
|
|
||||||
<object-type name="ARGUMENT" sorting-type="POSITION" />
|
|
||||||
<object-type name="TYPE ATTRIBUTE" sorting-type="POSITION" />
|
|
||||||
</sorting>
|
|
||||||
<default-editors>
|
|
||||||
<object-type name="VIEW" editor-type="SELECTION" />
|
|
||||||
<object-type name="PACKAGE" editor-type="SELECTION" />
|
|
||||||
<object-type name="TYPE" editor-type="SELECTION" />
|
|
||||||
</default-editors>
|
|
||||||
</browser-settings>
|
|
||||||
<navigation-settings>
|
|
||||||
<lookup-filters>
|
|
||||||
<lookup-objects>
|
|
||||||
<object-type name="SCHEMA" enabled="true" />
|
|
||||||
<object-type name="USER" enabled="false" />
|
|
||||||
<object-type name="ROLE" enabled="false" />
|
|
||||||
<object-type name="PRIVILEGE" enabled="false" />
|
|
||||||
<object-type name="CHARSET" enabled="false" />
|
|
||||||
<object-type name="TABLE" enabled="true" />
|
|
||||||
<object-type name="VIEW" enabled="true" />
|
|
||||||
<object-type name="MATERIALIZED VIEW" enabled="true" />
|
|
||||||
<object-type name="INDEX" enabled="true" />
|
|
||||||
<object-type name="CONSTRAINT" enabled="true" />
|
|
||||||
<object-type name="DATASET TRIGGER" enabled="true" />
|
|
||||||
<object-type name="DATABASE TRIGGER" enabled="true" />
|
|
||||||
<object-type name="SYNONYM" enabled="false" />
|
|
||||||
<object-type name="SEQUENCE" enabled="true" />
|
|
||||||
<object-type name="PROCEDURE" enabled="true" />
|
|
||||||
<object-type name="FUNCTION" enabled="true" />
|
|
||||||
<object-type name="PACKAGE" enabled="true" />
|
|
||||||
<object-type name="TYPE" enabled="true" />
|
|
||||||
<object-type name="JAVA CLASS" enabled="true" />
|
|
||||||
<object-type name="INNER CLASS" enabled="true" />
|
|
||||||
<object-type name="JAVA FIELD" enabled="true" />
|
|
||||||
<object-type name="JAVA METHOD" enabled="true" />
|
|
||||||
<object-type name="JAVA PARAMETER" enabled="true" />
|
|
||||||
<object-type name="DIMENSION" enabled="false" />
|
|
||||||
<object-type name="CLUSTER" enabled="false" />
|
|
||||||
<object-type name="DBLINK" enabled="false" />
|
|
||||||
<object-type name="CREDENTIAL" enabled="false" />
|
|
||||||
</lookup-objects>
|
|
||||||
<force-database-load value="false" />
|
|
||||||
<prompt-connection-selection value="true" />
|
|
||||||
<prompt-schema-selection value="true" />
|
|
||||||
</lookup-filters>
|
|
||||||
</navigation-settings>
|
|
||||||
<dataset-grid-settings>
|
|
||||||
<general>
|
|
||||||
<enable-zooming value="true" />
|
|
||||||
<enable-column-tooltip value="true" />
|
|
||||||
</general>
|
|
||||||
<sorting>
|
|
||||||
<nulls-first value="true" />
|
|
||||||
<max-sorting-columns value="4" />
|
|
||||||
</sorting>
|
|
||||||
<audit-columns>
|
|
||||||
<column-names value="" />
|
|
||||||
<visible value="true" />
|
|
||||||
<editable value="false" />
|
|
||||||
</audit-columns>
|
|
||||||
</dataset-grid-settings>
|
|
||||||
<dataset-editor-settings>
|
|
||||||
<text-editor-popup>
|
|
||||||
<active value="false" />
|
|
||||||
<active-if-empty value="false" />
|
|
||||||
<data-length-threshold value="100" />
|
|
||||||
<popup-delay value="1000" />
|
|
||||||
</text-editor-popup>
|
|
||||||
<values-actions-popup>
|
|
||||||
<show-popup-button value="true" />
|
|
||||||
<element-count-threshold value="1000" />
|
|
||||||
<data-length-threshold value="250" />
|
|
||||||
</values-actions-popup>
|
|
||||||
<general>
|
|
||||||
<fetch-block-size value="100" />
|
|
||||||
<fetch-timeout value="30" />
|
|
||||||
<trim-whitespaces value="true" />
|
|
||||||
<convert-empty-strings-to-null value="true" />
|
|
||||||
<select-content-on-cell-edit value="true" />
|
|
||||||
<large-value-preview-active value="true" />
|
|
||||||
</general>
|
|
||||||
<filters>
|
|
||||||
<prompt-filter-dialog value="true" />
|
|
||||||
<default-filter-type value="BASIC" />
|
|
||||||
</filters>
|
|
||||||
<qualified-text-editor text-length-threshold="300">
|
|
||||||
<content-types>
|
|
||||||
<content-type name="Text" enabled="true" />
|
|
||||||
<content-type name="Properties" enabled="true" />
|
|
||||||
<content-type name="XML" enabled="true" />
|
|
||||||
<content-type name="DTD" enabled="true" />
|
|
||||||
<content-type name="HTML" enabled="true" />
|
|
||||||
<content-type name="XHTML" enabled="true" />
|
|
||||||
<content-type name="SQL" enabled="true" />
|
|
||||||
<content-type name="PL/SQL" enabled="true" />
|
|
||||||
<content-type name="JSON" enabled="true" />
|
|
||||||
<content-type name="JSON5" enabled="true" />
|
|
||||||
<content-type name="YAML" enabled="true" />
|
|
||||||
</content-types>
|
|
||||||
</qualified-text-editor>
|
|
||||||
<record-navigation>
|
|
||||||
<navigation-target value="VIEWER" />
|
|
||||||
</record-navigation>
|
|
||||||
</dataset-editor-settings>
|
|
||||||
<code-editor-settings>
|
|
||||||
<general>
|
|
||||||
<show-object-navigation-gutter value="false" />
|
|
||||||
<show-spec-declaration-navigation-gutter value="true" />
|
|
||||||
<enable-spellchecking value="true" />
|
|
||||||
<enable-reference-spellchecking value="false" />
|
|
||||||
</general>
|
|
||||||
<confirmations>
|
|
||||||
<save-changes value="false" />
|
|
||||||
<revert-changes value="true" />
|
|
||||||
<exit-on-changes value="ASK" />
|
|
||||||
</confirmations>
|
|
||||||
</code-editor-settings>
|
|
||||||
<code-completion-settings>
|
|
||||||
<filters>
|
|
||||||
<basic-filter>
|
|
||||||
<filter-element type="RESERVED_WORD" id="keyword" selected="true" />
|
|
||||||
<filter-element type="RESERVED_WORD" id="function" selected="true" />
|
|
||||||
<filter-element type="RESERVED_WORD" id="parameter" selected="true" />
|
|
||||||
<filter-element type="RESERVED_WORD" id="datatype" selected="true" />
|
|
||||||
<filter-element type="RESERVED_WORD" id="exception" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="schema" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="role" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="user" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="privilege" selected="true" />
|
|
||||||
<user-schema>
|
|
||||||
<filter-element type="OBJECT" id="table" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="materialized view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="index" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="constraint" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="trigger" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="synonym" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="sequence" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="procedure" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="function" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="package" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="type" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dimension" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="cluster" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dblink" selected="true" />
|
|
||||||
</user-schema>
|
|
||||||
<public-schema>
|
|
||||||
<filter-element type="OBJECT" id="table" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="view" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="materialized view" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="index" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="constraint" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="trigger" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="synonym" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="sequence" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="procedure" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="function" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="package" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="type" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="dimension" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="cluster" selected="false" />
|
|
||||||
<filter-element type="OBJECT" id="dblink" selected="false" />
|
|
||||||
</public-schema>
|
|
||||||
<any-schema>
|
|
||||||
<filter-element type="OBJECT" id="table" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="materialized view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="index" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="constraint" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="trigger" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="synonym" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="sequence" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="procedure" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="function" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="package" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="type" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dimension" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="cluster" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dblink" selected="true" />
|
|
||||||
</any-schema>
|
|
||||||
</basic-filter>
|
|
||||||
<extended-filter>
|
|
||||||
<filter-element type="RESERVED_WORD" id="keyword" selected="true" />
|
|
||||||
<filter-element type="RESERVED_WORD" id="function" selected="true" />
|
|
||||||
<filter-element type="RESERVED_WORD" id="parameter" selected="true" />
|
|
||||||
<filter-element type="RESERVED_WORD" id="datatype" selected="true" />
|
|
||||||
<filter-element type="RESERVED_WORD" id="exception" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="schema" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="user" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="role" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="privilege" selected="true" />
|
|
||||||
<user-schema>
|
|
||||||
<filter-element type="OBJECT" id="table" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="materialized view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="index" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="constraint" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="trigger" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="synonym" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="sequence" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="procedure" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="function" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="package" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="type" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dimension" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="cluster" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dblink" selected="true" />
|
|
||||||
</user-schema>
|
|
||||||
<public-schema>
|
|
||||||
<filter-element type="OBJECT" id="table" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="materialized view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="index" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="constraint" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="trigger" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="synonym" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="sequence" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="procedure" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="function" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="package" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="type" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dimension" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="cluster" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dblink" selected="true" />
|
|
||||||
</public-schema>
|
|
||||||
<any-schema>
|
|
||||||
<filter-element type="OBJECT" id="table" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="materialized view" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="index" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="constraint" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="trigger" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="synonym" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="sequence" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="procedure" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="function" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="package" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="type" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dimension" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="cluster" selected="true" />
|
|
||||||
<filter-element type="OBJECT" id="dblink" selected="true" />
|
|
||||||
</any-schema>
|
|
||||||
</extended-filter>
|
|
||||||
</filters>
|
|
||||||
<sorting enabled="true">
|
|
||||||
<sorting-element type="RESERVED_WORD" id="keyword" />
|
|
||||||
<sorting-element type="RESERVED_WORD" id="datatype" />
|
|
||||||
<sorting-element type="OBJECT" id="column" />
|
|
||||||
<sorting-element type="OBJECT" id="table" />
|
|
||||||
<sorting-element type="OBJECT" id="view" />
|
|
||||||
<sorting-element type="OBJECT" id="materialized view" />
|
|
||||||
<sorting-element type="OBJECT" id="index" />
|
|
||||||
<sorting-element type="OBJECT" id="constraint" />
|
|
||||||
<sorting-element type="OBJECT" id="trigger" />
|
|
||||||
<sorting-element type="OBJECT" id="synonym" />
|
|
||||||
<sorting-element type="OBJECT" id="sequence" />
|
|
||||||
<sorting-element type="OBJECT" id="procedure" />
|
|
||||||
<sorting-element type="OBJECT" id="function" />
|
|
||||||
<sorting-element type="OBJECT" id="package" />
|
|
||||||
<sorting-element type="OBJECT" id="type" />
|
|
||||||
<sorting-element type="OBJECT" id="dimension" />
|
|
||||||
<sorting-element type="OBJECT" id="cluster" />
|
|
||||||
<sorting-element type="OBJECT" id="dblink" />
|
|
||||||
<sorting-element type="OBJECT" id="schema" />
|
|
||||||
<sorting-element type="OBJECT" id="role" />
|
|
||||||
<sorting-element type="OBJECT" id="user" />
|
|
||||||
<sorting-element type="RESERVED_WORD" id="function" />
|
|
||||||
<sorting-element type="RESERVED_WORD" id="parameter" />
|
|
||||||
</sorting>
|
|
||||||
<format>
|
|
||||||
<enforce-code-style-case value="true" />
|
|
||||||
</format>
|
|
||||||
</code-completion-settings>
|
|
||||||
<execution-engine-settings>
|
|
||||||
<statement-execution>
|
|
||||||
<fetch-block-size value="100" />
|
|
||||||
<execution-timeout value="20" />
|
|
||||||
<debug-execution-timeout value="600" />
|
|
||||||
<focus-result value="false" />
|
|
||||||
<prompt-execution value="false" />
|
|
||||||
</statement-execution>
|
|
||||||
<script-execution>
|
|
||||||
<command-line-interfaces />
|
|
||||||
<execution-timeout value="300" />
|
|
||||||
</script-execution>
|
|
||||||
<method-execution>
|
|
||||||
<execution-timeout value="30" />
|
|
||||||
<debug-execution-timeout value="600" />
|
|
||||||
<parameter-history-size value="10" />
|
|
||||||
</method-execution>
|
|
||||||
</execution-engine-settings>
|
|
||||||
<operation-settings>
|
|
||||||
<transactions>
|
|
||||||
<uncommitted-changes>
|
|
||||||
<on-project-close value="ASK" />
|
|
||||||
<on-disconnect value="ASK" />
|
|
||||||
<on-autocommit-toggle value="ASK" />
|
|
||||||
</uncommitted-changes>
|
|
||||||
<multiple-uncommitted-changes>
|
|
||||||
<on-commit value="ASK" />
|
|
||||||
<on-rollback value="ASK" />
|
|
||||||
</multiple-uncommitted-changes>
|
|
||||||
</transactions>
|
|
||||||
<session-browser>
|
|
||||||
<disconnect-session value="ASK" />
|
|
||||||
<kill-session value="ASK" />
|
|
||||||
<reload-on-filter-change value="false" />
|
|
||||||
</session-browser>
|
|
||||||
<compiler>
|
|
||||||
<compile-type value="KEEP" />
|
|
||||||
<compile-dependencies value="ASK" />
|
|
||||||
<always-show-controls value="false" />
|
|
||||||
</compiler>
|
|
||||||
</operation-settings>
|
|
||||||
<ddl-file-settings>
|
|
||||||
<extensions>
|
|
||||||
<mapping file-type-id="VIEW" extensions="vw" />
|
|
||||||
<mapping file-type-id="TRIGGER" extensions="trg" />
|
|
||||||
<mapping file-type-id="PROCEDURE" extensions="prc" />
|
|
||||||
<mapping file-type-id="FUNCTION" extensions="fnc" />
|
|
||||||
<mapping file-type-id="PACKAGE" extensions="pkg" />
|
|
||||||
<mapping file-type-id="PACKAGE_SPEC" extensions="pks" />
|
|
||||||
<mapping file-type-id="PACKAGE_BODY" extensions="pkb" />
|
|
||||||
<mapping file-type-id="TYPE" extensions="tpe" />
|
|
||||||
<mapping file-type-id="TYPE_SPEC" extensions="tps" />
|
|
||||||
<mapping file-type-id="TYPE_BODY" extensions="tpb" />
|
|
||||||
<mapping file-type-id="JAVA_SOURCE" extensions="sql" />
|
|
||||||
</extensions>
|
|
||||||
<general>
|
|
||||||
<lookup-ddl-files value="true" />
|
|
||||||
<create-ddl-files value="false" />
|
|
||||||
<synchronize-ddl-files value="true" />
|
|
||||||
<use-qualified-names value="false" />
|
|
||||||
<make-scripts-rerunnable value="true" />
|
|
||||||
</general>
|
|
||||||
</ddl-file-settings>
|
|
||||||
<assistant-settings>
|
|
||||||
<credential-settings>
|
|
||||||
<credentials />
|
|
||||||
</credential-settings>
|
|
||||||
</assistant-settings>
|
|
||||||
<general-settings>
|
|
||||||
<regional-settings>
|
|
||||||
<date-format value="MEDIUM" />
|
|
||||||
<number-format value="UNGROUPED" />
|
|
||||||
<locale value="SYSTEM_DEFAULT" />
|
|
||||||
<use-custom-formats value="false" />
|
|
||||||
</regional-settings>
|
|
||||||
<environment>
|
|
||||||
<environment-types>
|
|
||||||
<environment-type id="development" name="Development" description="Development environment" color="-2430209/-12296320" readonly-code="false" readonly-data="false" />
|
|
||||||
<environment-type id="integration" name="Integration" description="Integration environment" color="-2621494/-12163514" readonly-code="true" readonly-data="false" />
|
|
||||||
<environment-type id="production" name="Production" description="Productive environment" color="-11574/-10271420" readonly-code="true" readonly-data="true" />
|
|
||||||
<environment-type id="other" name="Other" description="" color="-1576/-10724543" readonly-code="false" readonly-data="false" />
|
|
||||||
</environment-types>
|
|
||||||
<visibility-settings>
|
|
||||||
<connection-tabs value="true" />
|
|
||||||
<dialog-headers value="true" />
|
|
||||||
<object-editor-tabs value="true" />
|
|
||||||
<script-editor-tabs value="false" />
|
|
||||||
<execution-result-tabs value="true" />
|
|
||||||
</visibility-settings>
|
|
||||||
</environment>
|
|
||||||
</general-settings>
|
|
||||||
</component>
|
|
||||||
</project>
|
|
||||||
16
.idea/inspectionProfiles/Project_Default.xml
generated
16
.idea/inspectionProfiles/Project_Default.xml
generated
@@ -1,16 +0,0 @@
|
|||||||
<component name="InspectionProjectProfileManager">
|
|
||||||
<profile version="1.0">
|
|
||||||
<option name="myName" value="Project Default" />
|
|
||||||
<inspection_tool class="PyInterpreterInspection" enabled="false" level="WARNING" enabled_by_default="false" />
|
|
||||||
<inspection_tool class="PyPackageRequirementsInspection" enabled="true" level="WARNING" enabled_by_default="true">
|
|
||||||
<option name="ignoredPackages">
|
|
||||||
<value>
|
|
||||||
<list size="2">
|
|
||||||
<item index="0" class="java.lang.String" itemvalue="requests" />
|
|
||||||
<item index="1" class="java.lang.String" itemvalue="python-telegram-bot" />
|
|
||||||
</list>
|
|
||||||
</value>
|
|
||||||
</option>
|
|
||||||
</inspection_tool>
|
|
||||||
</profile>
|
|
||||||
</component>
|
|
||||||
6
.idea/inspectionProfiles/profiles_settings.xml
generated
6
.idea/inspectionProfiles/profiles_settings.xml
generated
@@ -1,6 +0,0 @@
|
|||||||
<component name="InspectionProjectProfileManager">
|
|
||||||
<settings>
|
|
||||||
<option name="USE_PROJECT_PROFILE" value="false" />
|
|
||||||
<version value="1.0" />
|
|
||||||
</settings>
|
|
||||||
</component>
|
|
||||||
7
.idea/misc.xml
generated
7
.idea/misc.xml
generated
@@ -1,7 +0,0 @@
|
|||||||
<?xml version="1.0" encoding="UTF-8"?>
|
|
||||||
<project version="4">
|
|
||||||
<component name="Black">
|
|
||||||
<option name="sdkName" value="Python 3.12 (IoT-Systems)" />
|
|
||||||
</component>
|
|
||||||
<component name="ProjectRootManager" version="2" project-jdk-name="Python 3.12 (IoT-Systems)" project-jdk-type="Python SDK" />
|
|
||||||
</project>
|
|
||||||
8
.idea/modules.xml
generated
8
.idea/modules.xml
generated
@@ -1,8 +0,0 @@
|
|||||||
<?xml version="1.0" encoding="UTF-8"?>
|
|
||||||
<project version="4">
|
|
||||||
<component name="ProjectModuleManager">
|
|
||||||
<modules>
|
|
||||||
<module fileurl="file://$PROJECT_DIR$/.idea/IoT-Systems.iml" filepath="$PROJECT_DIR$/.idea/IoT-Systems.iml" />
|
|
||||||
</modules>
|
|
||||||
</component>
|
|
||||||
</project>
|
|
||||||
6
.idea/vcs.xml
generated
6
.idea/vcs.xml
generated
@@ -1,6 +0,0 @@
|
|||||||
<?xml version="1.0" encoding="UTF-8"?>
|
|
||||||
<project version="4">
|
|
||||||
<component name="VcsDirectoryMappings">
|
|
||||||
<mapping directory="$PROJECT_DIR$" vcs="Git" />
|
|
||||||
</component>
|
|
||||||
</project>
|
|
||||||
3
MapView/.gitignore
vendored
3
MapView/.gitignore
vendored
@@ -1,3 +0,0 @@
|
|||||||
.idea
|
|
||||||
venv
|
|
||||||
__pycache__
|
|
||||||
@@ -2,3 +2,5 @@ import os
|
|||||||
|
|
||||||
STORE_HOST = os.environ.get("STORE_HOST") or "localhost"
|
STORE_HOST = os.environ.get("STORE_HOST") or "localhost"
|
||||||
STORE_PORT = os.environ.get("STORE_PORT") or 8000
|
STORE_PORT = os.environ.get("STORE_PORT") or 8000
|
||||||
|
|
||||||
|
TRACK_ID = int(os.environ.get("TID") or '1')
|
||||||
|
|||||||
@@ -72,9 +72,10 @@ class Datasource:
|
|||||||
)
|
)
|
||||||
new_points = [
|
new_points = [
|
||||||
(
|
(
|
||||||
processed_agent_data.latitude,
|
|
||||||
processed_agent_data.longitude,
|
processed_agent_data.longitude,
|
||||||
|
processed_agent_data.latitude,
|
||||||
processed_agent_data.road_state,
|
processed_agent_data.road_state,
|
||||||
|
processed_agent_data.user_id
|
||||||
)
|
)
|
||||||
for processed_agent_data in processed_agent_data_list
|
for processed_agent_data in processed_agent_data_list
|
||||||
]
|
]
|
||||||
|
|||||||
@@ -2,6 +2,6 @@ from dataclasses import dataclass
|
|||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
class Accelerometer:
|
class Accelerometer:
|
||||||
x: int
|
x: float
|
||||||
y: int
|
y: float
|
||||||
z: int
|
z: float
|
||||||
|
|||||||
117
MapView/main.py
117
MapView/main.py
@@ -4,48 +4,143 @@ from kivy_garden.mapview import MapMarker, MapView
|
|||||||
from kivy.clock import Clock
|
from kivy.clock import Clock
|
||||||
from lineMapLayer import LineMapLayer
|
from lineMapLayer import LineMapLayer
|
||||||
from datasource import Datasource
|
from datasource import Datasource
|
||||||
|
import config
|
||||||
|
|
||||||
|
line_layer_colors = [
|
||||||
|
[1, 0, 0, 1],
|
||||||
|
[1, 0.5, 0, 1],
|
||||||
|
[0, 1, 0, 1],
|
||||||
|
[0, 1, 1, 1],
|
||||||
|
[0, 0, 1, 1],
|
||||||
|
[1, 0, 1, 1],
|
||||||
|
]
|
||||||
|
|
||||||
class MapViewApp(App):
|
class MapViewApp(App):
|
||||||
def __init__(self, **kwargs):
|
def __init__(self, **kwargs):
|
||||||
super().__init__()
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
|
self.mapview = None
|
||||||
|
self.datasource = Datasource(user_id=1)
|
||||||
|
self.line_layers = dict()
|
||||||
|
self.car_markers = dict()
|
||||||
|
|
||||||
# додати необхідні змінні
|
# додати необхідні змінні
|
||||||
|
self.bump_markers = []
|
||||||
|
self.pothole_markers = []
|
||||||
|
|
||||||
def on_start(self):
|
def on_start(self):
|
||||||
"""
|
"""
|
||||||
Встановлює необхідні маркери, викликає функцію для оновлення мапи
|
Встановлює необхідні маркери, викликає функцію для оновлення мапи
|
||||||
"""
|
"""
|
||||||
|
self.update()
|
||||||
|
Clock.schedule_interval(self.update, 0.1)
|
||||||
|
|
||||||
def update(self, *args):
|
def update(self, *args):
|
||||||
"""
|
"""
|
||||||
Викликається регулярно для оновлення мапи
|
Викликається регулярно для оновлення мапи
|
||||||
"""
|
"""
|
||||||
|
new_points = self.datasource.get_new_points()
|
||||||
|
|
||||||
|
if not new_points:
|
||||||
|
return
|
||||||
|
|
||||||
def update_car_marker(self, point):
|
for point in new_points:
|
||||||
|
|
||||||
|
lat, lon, road_state, user_id = point
|
||||||
|
|
||||||
|
# Оновлює лінію маршрута
|
||||||
|
if user_id not in self.line_layers:
|
||||||
|
self.line_layers[user_id] = LineMapLayer(color = line_layer_colors[user_id % len(line_layer_colors)])
|
||||||
|
self.mapview.add_layer(self.line_layers[user_id])
|
||||||
|
|
||||||
|
self.line_layers[user_id].add_point((lat, lon))
|
||||||
|
|
||||||
|
# Оновлює маркер маниши
|
||||||
|
self.update_car_marker(lat, lon, user_id)
|
||||||
|
|
||||||
|
# Перевіряємо стан дороги
|
||||||
|
self.check_road_quality(point)
|
||||||
|
|
||||||
|
def check_road_quality(self, point):
|
||||||
|
"""
|
||||||
|
Аналізує дані акселерометра для подальшого визначення
|
||||||
|
та відображення ям та лежачих поліцейських
|
||||||
|
"""
|
||||||
|
if len(point) < 3:
|
||||||
|
return
|
||||||
|
|
||||||
|
lat, lon, road_state, user_id = point
|
||||||
|
|
||||||
|
if road_state == "pothole":
|
||||||
|
self.set_pothole_marker((lat, lon))
|
||||||
|
elif road_state == "bump":
|
||||||
|
self.set_bump_marker((lat, lon))
|
||||||
|
|
||||||
|
def update_car_marker(self, lat, lon, user_id):
|
||||||
"""
|
"""
|
||||||
Оновлює відображення маркера машини на мапі
|
Оновлює відображення маркера машини на мапі
|
||||||
:param point: GPS координати
|
:param point: GPS координати
|
||||||
"""
|
"""
|
||||||
|
if user_id not in self.car_markers:
|
||||||
|
self.car_markers[user_id] = MapMarker(lat=lat, lon=lon, source='./images/car.png')
|
||||||
|
self.mapview.add_marker(self.car_markers[user_id])
|
||||||
|
else:
|
||||||
|
self.car_markers[user_id].lat = lat
|
||||||
|
self.car_markers[user_id].lon = lon
|
||||||
|
|
||||||
|
if user_id == config.TRACK_ID:
|
||||||
|
self.mapview.center_on(lat, lon)
|
||||||
|
|
||||||
def set_pothole_marker(self, point):
|
def set_pothole_marker(self, point):
|
||||||
"""
|
if isinstance(point, dict):
|
||||||
Встановлює маркер для ями
|
lat = point.get("lat")
|
||||||
:param point: GPS координати
|
lon = point.get("lon")
|
||||||
"""
|
else:
|
||||||
|
lat, lon = point
|
||||||
|
|
||||||
|
if lat is None or lon is None:
|
||||||
|
return
|
||||||
|
|
||||||
|
marker = MapMarker(
|
||||||
|
lat=lat,
|
||||||
|
lon=lon,
|
||||||
|
source="images/pothole.png"
|
||||||
|
)
|
||||||
|
|
||||||
|
self.mapview.add_marker(marker)
|
||||||
|
self.pothole_markers.append(marker)
|
||||||
|
|
||||||
def set_bump_marker(self, point):
|
def set_bump_marker(self, point):
|
||||||
"""
|
if isinstance(point, dict):
|
||||||
Встановлює маркер для лежачого поліцейського
|
lat = point.get("lat")
|
||||||
:param point: GPS координати
|
lon = point.get("lon")
|
||||||
"""
|
else:
|
||||||
|
lat, lon = point
|
||||||
|
|
||||||
|
if lat is None or lon is None:
|
||||||
|
return
|
||||||
|
|
||||||
|
marker = MapMarker(
|
||||||
|
lat=lat,
|
||||||
|
lon=lon,
|
||||||
|
source="images/bump.png"
|
||||||
|
)
|
||||||
|
|
||||||
|
self.mapview.add_marker(marker)
|
||||||
|
self.bump_markers.append(marker)
|
||||||
|
|
||||||
|
|
||||||
def build(self):
|
def build(self):
|
||||||
"""
|
"""
|
||||||
Ініціалізує мапу MapView(zoom, lat, lon)
|
Ініціалізує мапу MapView(zoom, lat, lon)
|
||||||
:return: мапу
|
:return: мапу
|
||||||
"""
|
"""
|
||||||
self.mapview = MapView()
|
self.mapview = MapView(
|
||||||
|
zoom=15,
|
||||||
|
lat=50.4501,
|
||||||
|
lon=30.5234
|
||||||
|
)
|
||||||
|
|
||||||
return self.mapview
|
return self.mapview
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
2
agent/.gitignore
vendored
2
agent/.gitignore
vendored
@@ -1,2 +0,0 @@
|
|||||||
venv
|
|
||||||
__pycache__
|
|
||||||
@@ -3,10 +3,10 @@ FROM python:latest
|
|||||||
# set the working directory in the container
|
# set the working directory in the container
|
||||||
WORKDIR /usr/agent
|
WORKDIR /usr/agent
|
||||||
# copy the dependencies file to the working directory
|
# copy the dependencies file to the working directory
|
||||||
COPY requirements.txt .
|
COPY agent/requirements.txt .
|
||||||
# install dependencies
|
# install dependencies
|
||||||
RUN pip install -r requirements.txt
|
RUN pip install --no-cache-dir -r requirements.txt
|
||||||
# copy the content of the local src directory to the working directory
|
# copy the content of the local src directory to the working directory
|
||||||
COPY src/ .
|
COPY agent/src/ .
|
||||||
# command to run on container start
|
# command to run on container start
|
||||||
CMD ["python", "main.py"]
|
CMD ["python", "main.py"]
|
||||||
@@ -1,32 +0,0 @@
|
|||||||
name: "road_vision"
|
|
||||||
services:
|
|
||||||
mqtt:
|
|
||||||
image: eclipse-mosquitto
|
|
||||||
container_name: mqtt
|
|
||||||
volumes:
|
|
||||||
- ./mosquitto:/mosquitto
|
|
||||||
- ./mosquitto/data:/mosquitto/data
|
|
||||||
- ./mosquitto/log:/mosquitto/log
|
|
||||||
ports:
|
|
||||||
- 1883:1883
|
|
||||||
- 9001:9001
|
|
||||||
networks:
|
|
||||||
mqtt_network:
|
|
||||||
|
|
||||||
|
|
||||||
fake_agent:
|
|
||||||
container_name: agent
|
|
||||||
build: ../
|
|
||||||
depends_on:
|
|
||||||
- mqtt
|
|
||||||
environment:
|
|
||||||
MQTT_BROKER_HOST: "mqtt"
|
|
||||||
MQTT_BROKER_PORT: 1883
|
|
||||||
MQTT_TOPIC: "agent_data_topic"
|
|
||||||
DELAY: 0.1
|
|
||||||
networks:
|
|
||||||
mqtt_network:
|
|
||||||
|
|
||||||
|
|
||||||
networks:
|
|
||||||
mqtt_network:
|
|
||||||
@@ -8,7 +8,7 @@ def try_parse(type, value: str):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
USER_ID = 1
|
USER_ID = try_parse(int, os.environ.get("USER_ID")) or 1
|
||||||
# MQTT config
|
# MQTT config
|
||||||
MQTT_BROKER_HOST = os.environ.get("MQTT_BROKER_HOST") or "mqtt"
|
MQTT_BROKER_HOST = os.environ.get("MQTT_BROKER_HOST") or "mqtt"
|
||||||
MQTT_BROKER_PORT = try_parse(int, os.environ.get("MQTT_BROKER_PORT")) or 1883
|
MQTT_BROKER_PORT = try_parse(int, os.environ.get("MQTT_BROKER_PORT")) or 1883
|
||||||
|
|||||||
@@ -3,6 +3,6 @@ from dataclasses import dataclass
|
|||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
class Accelerometer:
|
class Accelerometer:
|
||||||
x: int
|
x: float
|
||||||
y: int
|
y: float
|
||||||
z: int
|
z: float
|
||||||
|
|||||||
@@ -15,6 +15,7 @@ class FileDatasource:
|
|||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
|
acc_divisor: float,
|
||||||
accelerometer_filename: str,
|
accelerometer_filename: str,
|
||||||
gps_filename: str,
|
gps_filename: str,
|
||||||
park_filename: str,
|
park_filename: str,
|
||||||
@@ -34,6 +35,8 @@ class FileDatasource:
|
|||||||
|
|
||||||
self._started = False
|
self._started = False
|
||||||
|
|
||||||
|
self.acc_divisor = acc_divisor
|
||||||
|
|
||||||
def startReading(self, *args, **kwargs):
|
def startReading(self, *args, **kwargs):
|
||||||
"""Must be called before read()"""
|
"""Must be called before read()"""
|
||||||
if self._started:
|
if self._started:
|
||||||
@@ -160,15 +163,14 @@ class FileDatasource:
|
|||||||
|
|
||||||
return row
|
return row
|
||||||
|
|
||||||
@staticmethod
|
def _parse_acc(self, row: List[str]) -> Accelerometer:
|
||||||
def _parse_acc(row: List[str]) -> Accelerometer:
|
|
||||||
if len(row) < 3:
|
if len(row) < 3:
|
||||||
raise ValueError(f"Accelerometer row must have 3 values (x,y,z). Got: {row}")
|
raise ValueError(f"Accelerometer row must have 3 values (x,y,z). Got: {row}")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
x = int(row[0])
|
x = int(row[0]) / self.acc_divisor
|
||||||
y = int(row[1])
|
y = int(row[1]) / self.acc_divisor
|
||||||
z = int(row[2])
|
z = int(row[2]) / self.acc_divisor
|
||||||
except ValueError as e:
|
except ValueError as e:
|
||||||
raise ValueError(f"Invalid accelerometer values (expected integers): {row}") from e
|
raise ValueError(f"Invalid accelerometer values (expected integers): {row}") from e
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
from paho.mqtt import client as mqtt_client
|
from paho.mqtt import client as mqtt_client
|
||||||
from schema.aggregated_data_schema import AggregatedDataSchema
|
from schema.aggregated_data_schema import AggregatedDataSchema
|
||||||
from file_datasource import FileDatasource
|
from file_datasource import FileDatasource
|
||||||
|
import logging
|
||||||
import config
|
import config
|
||||||
|
|
||||||
|
|
||||||
@@ -15,6 +16,8 @@ def connect_mqtt(broker, port):
|
|||||||
print("Failed to connect {broker}:{port}, return code %d\n", rc)
|
print("Failed to connect {broker}:{port}, return code %d\n", rc)
|
||||||
exit(rc) # Stop execution
|
exit(rc) # Stop execution
|
||||||
|
|
||||||
|
logging.info(f"Acting as USER_ID = {config.USER_ID}")
|
||||||
|
|
||||||
client = mqtt_client.Client()
|
client = mqtt_client.Client()
|
||||||
client.on_connect = on_connect
|
client.on_connect = on_connect
|
||||||
client.connect(broker, port)
|
client.connect(broker, port)
|
||||||
@@ -28,16 +31,18 @@ def publish(client, topic, datasource):
|
|||||||
data = datasource.read()
|
data = datasource.read()
|
||||||
msg = AggregatedDataSchema().dumps(data)
|
msg = AggregatedDataSchema().dumps(data)
|
||||||
result = client.publish(topic, msg)
|
result = client.publish(topic, msg)
|
||||||
|
logging.debug(f"Published to {topic}: {msg[:50]}...")
|
||||||
status = result[0]
|
status = result[0]
|
||||||
if status != 0:
|
if status != 0:
|
||||||
print(f"Failed to send message to topic {topic}")
|
logging.error(f"Failed to send message to topic {topic}")
|
||||||
|
|
||||||
|
|
||||||
def run():
|
def run():
|
||||||
|
logging.basicConfig(level = logging.INFO)
|
||||||
# Prepare mqtt client
|
# Prepare mqtt client
|
||||||
client = connect_mqtt(config.MQTT_BROKER_HOST, config.MQTT_BROKER_PORT)
|
client = connect_mqtt(config.MQTT_BROKER_HOST, config.MQTT_BROKER_PORT)
|
||||||
# Prepare datasource
|
# Prepare datasource
|
||||||
datasource = FileDatasource("data/accelerometer.csv", "data/gps.csv", "data/parking.csv")
|
datasource = FileDatasource(16384.0, "data/accelerometer.csv", "data/gps.csv", "data/parking.csv")
|
||||||
# Infinity publish data
|
# Infinity publish data
|
||||||
publish(client, config.MQTT_TOPIC, datasource)
|
publish(client, config.MQTT_TOPIC, datasource)
|
||||||
|
|
||||||
|
|||||||
@@ -2,6 +2,6 @@ from marshmallow import Schema, fields
|
|||||||
|
|
||||||
|
|
||||||
class AccelerometerSchema(Schema):
|
class AccelerometerSchema(Schema):
|
||||||
x = fields.Int()
|
x = fields.Float()
|
||||||
y = fields.Int()
|
y = fields.Float()
|
||||||
z = fields.Int()
|
z = fields.Float()
|
||||||
|
|||||||
@@ -1,13 +1,12 @@
|
|||||||
version: "3.9"
|
name: "road_vision"
|
||||||
name: "road_vision__hub"
|
|
||||||
services:
|
services:
|
||||||
mqtt:
|
mqtt:
|
||||||
image: eclipse-mosquitto
|
image: eclipse-mosquitto
|
||||||
container_name: mqtt
|
container_name: mqtt
|
||||||
volumes:
|
volumes:
|
||||||
- ./mosquitto:/mosquitto
|
- ./agent/docker/mosquitto:/mosquitto
|
||||||
- ./mosquitto/data:/mosquitto/data
|
- ./agent/docker/mosquitto/data:/mosquitto/data
|
||||||
- ./mosquitto/log:/mosquitto/log
|
- ./agent/docker/mosquitto/log:/mosquitto/log
|
||||||
ports:
|
ports:
|
||||||
- 1883:1883
|
- 1883:1883
|
||||||
- 9001:9001
|
- 9001:9001
|
||||||
@@ -15,8 +14,45 @@ services:
|
|||||||
mqtt_network:
|
mqtt_network:
|
||||||
|
|
||||||
|
|
||||||
|
fake_agent:
|
||||||
|
container_name: agent
|
||||||
|
build:
|
||||||
|
context: .
|
||||||
|
dockerfile: agent/Dockerfile
|
||||||
|
depends_on:
|
||||||
|
- mqtt
|
||||||
|
environment:
|
||||||
|
PYTHONUNBUFFERED: 1
|
||||||
|
MQTT_BROKER_HOST: "mqtt"
|
||||||
|
MQTT_BROKER_PORT: 1883
|
||||||
|
MQTT_TOPIC: "agent_data_topic"
|
||||||
|
DELAY: 0.1
|
||||||
|
networks:
|
||||||
|
mqtt_network:
|
||||||
|
|
||||||
|
edge:
|
||||||
|
container_name: edge
|
||||||
|
build:
|
||||||
|
context: .
|
||||||
|
dockerfile: edge/Dockerfile
|
||||||
|
depends_on:
|
||||||
|
- mqtt
|
||||||
|
environment:
|
||||||
|
MQTT_BROKER_HOST: "mqtt"
|
||||||
|
MQTT_BROKER_PORT: 1883
|
||||||
|
MQTT_TOPIC: "agent_data_topic"
|
||||||
|
HUB_HOST: "hub"
|
||||||
|
HUB_PORT: 8000
|
||||||
|
HUB_CONNECTION_TYPE: "http"
|
||||||
|
HUB_MQTT_BROKER_HOST: "mqtt"
|
||||||
|
HUB_MQTT_BROKER_PORT: 1883
|
||||||
|
HUB_MQTT_TOPIC: "processed_data_topic"
|
||||||
|
networks:
|
||||||
|
mqtt_network:
|
||||||
|
edge_hub:
|
||||||
|
|
||||||
postgres_db:
|
postgres_db:
|
||||||
image: postgres:latest
|
image: postgres:17
|
||||||
container_name: postgres_db
|
container_name: postgres_db
|
||||||
restart: always
|
restart: always
|
||||||
environment:
|
environment:
|
||||||
@@ -25,13 +61,12 @@ services:
|
|||||||
POSTGRES_DB: test_db
|
POSTGRES_DB: test_db
|
||||||
volumes:
|
volumes:
|
||||||
- postgres_data:/var/lib/postgresql/data
|
- postgres_data:/var/lib/postgresql/data
|
||||||
- ./db/structure.sql:/docker-entrypoint-initdb.d/structure.sql
|
- ./store/docker/db/structure.sql:/docker-entrypoint-initdb.d/structure.sql
|
||||||
ports:
|
ports:
|
||||||
- "5432:5432"
|
- "5432:5432"
|
||||||
networks:
|
networks:
|
||||||
db_network:
|
db_network:
|
||||||
|
|
||||||
|
|
||||||
pgadmin:
|
pgadmin:
|
||||||
container_name: pgadmin4
|
container_name: pgadmin4
|
||||||
image: dpage/pgadmin4
|
image: dpage/pgadmin4
|
||||||
@@ -49,11 +84,14 @@ services:
|
|||||||
|
|
||||||
store:
|
store:
|
||||||
container_name: store
|
container_name: store
|
||||||
build: ../../store
|
build:
|
||||||
|
context: .
|
||||||
|
dockerfile: store/Dockerfile
|
||||||
depends_on:
|
depends_on:
|
||||||
- postgres_db
|
- postgres_db
|
||||||
restart: always
|
restart: always
|
||||||
environment:
|
environment:
|
||||||
|
PYTHONUNBUFFERED: 1
|
||||||
POSTGRES_USER: user
|
POSTGRES_USER: user
|
||||||
POSTGRES_PASSWORD: pass
|
POSTGRES_PASSWORD: pass
|
||||||
POSTGRES_DB: test_db
|
POSTGRES_DB: test_db
|
||||||
@@ -77,12 +115,15 @@ services:
|
|||||||
|
|
||||||
hub:
|
hub:
|
||||||
container_name: hub
|
container_name: hub
|
||||||
build: ../
|
build:
|
||||||
|
context: .
|
||||||
|
dockerfile: hub/Dockerfile
|
||||||
depends_on:
|
depends_on:
|
||||||
- mqtt
|
- mqtt
|
||||||
- redis
|
- redis
|
||||||
- store
|
- store
|
||||||
environment:
|
environment:
|
||||||
|
PYTHONUNBUFFERED: 1
|
||||||
STORE_API_HOST: "store"
|
STORE_API_HOST: "store"
|
||||||
STORE_API_PORT: 8000
|
STORE_API_PORT: 8000
|
||||||
REDIS_HOST: "redis"
|
REDIS_HOST: "redis"
|
||||||
@@ -90,7 +131,7 @@ services:
|
|||||||
MQTT_BROKER_HOST: "mqtt"
|
MQTT_BROKER_HOST: "mqtt"
|
||||||
MQTT_BROKER_PORT: 1883
|
MQTT_BROKER_PORT: 1883
|
||||||
MQTT_TOPIC: "processed_data_topic"
|
MQTT_TOPIC: "processed_data_topic"
|
||||||
BATCH_SIZE: 1
|
BATCH_SIZE: 20
|
||||||
ports:
|
ports:
|
||||||
- "9000:8000"
|
- "9000:8000"
|
||||||
networks:
|
networks:
|
||||||
@@ -98,10 +139,11 @@ services:
|
|||||||
hub_store:
|
hub_store:
|
||||||
hub_redis:
|
hub_redis:
|
||||||
|
|
||||||
|
|
||||||
networks:
|
networks:
|
||||||
mqtt_network:
|
mqtt_network:
|
||||||
db_network:
|
db_network:
|
||||||
|
edge_hub:
|
||||||
|
hub:
|
||||||
hub_store:
|
hub_store:
|
||||||
hub_redis:
|
hub_redis:
|
||||||
|
|
||||||
2
edge/.gitignore
vendored
2
edge/.gitignore
vendored
@@ -1,2 +0,0 @@
|
|||||||
venv
|
|
||||||
app.log
|
|
||||||
@@ -3,9 +3,9 @@ FROM python:3.9-slim
|
|||||||
# Set the working directory inside the container
|
# Set the working directory inside the container
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
# Copy the requirements.txt file and install dependencies
|
# Copy the requirements.txt file and install dependencies
|
||||||
COPY requirements.txt .
|
COPY edge/requirements.txt .
|
||||||
RUN pip install --no-cache-dir -r requirements.txt
|
RUN pip install --no-cache-dir -r requirements.txt
|
||||||
# Copy the entire application into the container
|
# Copy the entire application into the container
|
||||||
COPY . .
|
COPY edge/. .
|
||||||
# Run the main.py script inside the container when it starts
|
# Run the main.py script inside the container when it starts
|
||||||
CMD ["python", "main.py"]
|
CMD ["python", "main.py"]
|
||||||
@@ -13,9 +13,7 @@ class AgentMQTTAdapter(AgentGateway):
|
|||||||
broker_port,
|
broker_port,
|
||||||
topic,
|
topic,
|
||||||
hub_gateway: HubGateway,
|
hub_gateway: HubGateway,
|
||||||
batch_size=10,
|
|
||||||
):
|
):
|
||||||
self.batch_size = batch_size
|
|
||||||
# MQTT
|
# MQTT
|
||||||
self.broker_host = broker_host
|
self.broker_host = broker_host
|
||||||
self.broker_port = broker_port
|
self.broker_port = broker_port
|
||||||
@@ -35,42 +33,21 @@ class AgentMQTTAdapter(AgentGateway):
|
|||||||
"""Processing agent data and sent it to hub gateway"""
|
"""Processing agent data and sent it to hub gateway"""
|
||||||
try:
|
try:
|
||||||
payload: str = msg.payload.decode("utf-8")
|
payload: str = msg.payload.decode("utf-8")
|
||||||
# Create AgentData instance with the received data
|
|
||||||
agent_data = AgentData.model_validate_json(payload, strict=True)
|
agent_data = AgentData.model_validate_json(payload, strict=True)
|
||||||
# Process the received data (you can call a use case here if needed)
|
|
||||||
processed_data = process_agent_data(agent_data)
|
processed_data = process_agent_data(agent_data)
|
||||||
# Store the agent_data in the database (you can send it to the data processing module)
|
|
||||||
if not self.hub_gateway.save_data(processed_data):
|
if self.hub_gateway.save_data(processed_data):
|
||||||
logging.error("Hub is not available")
|
logging.info("Processed data successfully forwarded to the Hub.")
|
||||||
|
else:
|
||||||
|
logging.error("Failed to send data: Hub gateway is unavailable.")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.info(f"Error processing MQTT message: {e}")
|
logging.error(f"Error processing MQTT message: {e}")
|
||||||
|
|
||||||
def connect(self):
|
def connect(self):
|
||||||
self.client.on_connect = self.on_connect
|
self.client.on_connect = self.on_connect
|
||||||
self.client.on_message = self.on_message
|
self.client.on_message = self.on_message
|
||||||
self.client.connect(self.broker_host, self.broker_port, 60)
|
self.client.connect(self.broker_host, self.broker_port, 60)
|
||||||
|
|
||||||
def start(self):
|
def loop_forever(self):
|
||||||
self.client.loop_start()
|
self.client.loop_forever()
|
||||||
|
|
||||||
def stop(self):
|
|
||||||
self.client.loop_stop()
|
|
||||||
|
|
||||||
|
|
||||||
# Usage example:
|
|
||||||
if __name__ == "__main__":
|
|
||||||
broker_host = "localhost"
|
|
||||||
broker_port = 1883
|
|
||||||
topic = "agent_data_topic"
|
|
||||||
# Assuming you have implemented the StoreGateway and passed it to the adapter
|
|
||||||
store_gateway = HubGateway()
|
|
||||||
adapter = AgentMQTTAdapter(broker_host, broker_port, topic, store_gateway)
|
|
||||||
adapter.connect()
|
|
||||||
adapter.start()
|
|
||||||
try:
|
|
||||||
# Keep the adapter running in the background
|
|
||||||
while True:
|
|
||||||
pass
|
|
||||||
except KeyboardInterrupt:
|
|
||||||
adapter.stop()
|
|
||||||
logging.info("Adapter stopped.")
|
|
||||||
|
|||||||
@@ -14,6 +14,7 @@ class GpsData(BaseModel):
|
|||||||
|
|
||||||
|
|
||||||
class AgentData(BaseModel):
|
class AgentData(BaseModel):
|
||||||
|
user_id: int
|
||||||
accelerometer: AccelerometerData
|
accelerometer: AccelerometerData
|
||||||
gps: GpsData
|
gps: GpsData
|
||||||
timestamp: datetime
|
timestamp: datetime
|
||||||
|
|||||||
@@ -26,15 +26,8 @@ class AgentGateway(ABC):
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
@abstractmethod
|
@abstractmethod
|
||||||
def start(self):
|
def loop_forever(self):
|
||||||
"""
|
"""
|
||||||
Method to start listening for messages from the agent.
|
Method to await for new messages.
|
||||||
"""
|
|
||||||
pass
|
|
||||||
|
|
||||||
@abstractmethod
|
|
||||||
def stop(self):
|
|
||||||
"""
|
|
||||||
Method to stop the agent gateway and clean up resources.
|
|
||||||
"""
|
"""
|
||||||
pass
|
pass
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
from app.entities.agent_data import AgentData
|
from app.entities.agent_data import AgentData
|
||||||
from app.entities.processed_agent_data import ProcessedAgentData
|
from app.entities.processed_agent_data import ProcessedAgentData
|
||||||
|
|
||||||
|
_last_detection_state = {}
|
||||||
|
|
||||||
def process_agent_data(
|
def process_agent_data(
|
||||||
agent_data: AgentData,
|
agent_data: AgentData,
|
||||||
@@ -12,4 +13,24 @@ def process_agent_data(
|
|||||||
Returns:
|
Returns:
|
||||||
processed_data_batch (ProcessedAgentData): Processed data containing the classified state of the road surface and agent data.
|
processed_data_batch (ProcessedAgentData): Processed data containing the classified state of the road surface and agent data.
|
||||||
"""
|
"""
|
||||||
# Implement it
|
user_id = agent_data.user_id
|
||||||
|
road_state = "normal"
|
||||||
|
|
||||||
|
last_detection_state = _last_detection_state.get(user_id, False)
|
||||||
|
|
||||||
|
if (agent_data.accelerometer.z < 0.6):
|
||||||
|
road_state = "pothole"
|
||||||
|
elif (agent_data.accelerometer.z > 1.2):
|
||||||
|
road_state = "bump"
|
||||||
|
|
||||||
|
detection_happened = road_state != "normal"
|
||||||
|
|
||||||
|
if not (not last_detection_state and detection_happened):
|
||||||
|
road_state = "normal"
|
||||||
|
|
||||||
|
_last_detection_state[user_id] = detection_happened
|
||||||
|
|
||||||
|
return ProcessedAgentData(
|
||||||
|
road_state=road_state,
|
||||||
|
agent_data=agent_data
|
||||||
|
)
|
||||||
|
|||||||
@@ -16,9 +16,12 @@ MQTT_TOPIC = os.environ.get("MQTT_TOPIC") or "agent_data_topic"
|
|||||||
# Configuration for hub MQTT
|
# Configuration for hub MQTT
|
||||||
HUB_MQTT_BROKER_HOST = os.environ.get("HUB_MQTT_BROKER_HOST") or "localhost"
|
HUB_MQTT_BROKER_HOST = os.environ.get("HUB_MQTT_BROKER_HOST") or "localhost"
|
||||||
HUB_MQTT_BROKER_PORT = try_parse_int(os.environ.get("HUB_MQTT_BROKER_PORT")) or 1883
|
HUB_MQTT_BROKER_PORT = try_parse_int(os.environ.get("HUB_MQTT_BROKER_PORT")) or 1883
|
||||||
HUB_MQTT_TOPIC = os.environ.get("HUB_MQTT_TOPIC") or "processed_agent_data_topic"
|
HUB_MQTT_TOPIC = os.environ.get("HUB_MQTT_TOPIC") or "processed_data_topic"
|
||||||
|
|
||||||
# Configuration for the Hub
|
# Configuration for the Hub
|
||||||
HUB_HOST = os.environ.get("HUB_HOST") or "localhost"
|
HUB_HOST = os.environ.get("HUB_HOST") or "localhost"
|
||||||
HUB_PORT = try_parse_int(os.environ.get("HUB_PORT")) or 12000
|
HUB_PORT = try_parse_int(os.environ.get("HUB_PORT")) or 8000
|
||||||
HUB_URL = f"http://{HUB_HOST}:{HUB_PORT}"
|
HUB_URL = f"http://{HUB_HOST}:{HUB_PORT}"
|
||||||
|
|
||||||
|
# For choosing type of connection
|
||||||
|
HUB_CONNECTION_TYPE = os.environ.get("HUB_CONNECTION_TYPE") or "mqtt"
|
||||||
@@ -1,48 +0,0 @@
|
|||||||
version: "3.9"
|
|
||||||
# name: "road_vision"
|
|
||||||
services:
|
|
||||||
mqtt:
|
|
||||||
image: eclipse-mosquitto
|
|
||||||
container_name: mqtt
|
|
||||||
volumes:
|
|
||||||
- ./mosquitto:/mosquitto
|
|
||||||
- ./mosquitto/data:/mosquitto/data
|
|
||||||
- ./mosquitto/log:/mosquitto/log
|
|
||||||
ports:
|
|
||||||
- 1883:1883
|
|
||||||
- 19001:9001
|
|
||||||
networks:
|
|
||||||
mqtt_network:
|
|
||||||
|
|
||||||
|
|
||||||
edge:
|
|
||||||
container_name: edge
|
|
||||||
build: ../
|
|
||||||
depends_on:
|
|
||||||
- mqtt
|
|
||||||
environment:
|
|
||||||
MQTT_BROKER_HOST: "mqtt"
|
|
||||||
MQTT_BROKER_PORT: 1883
|
|
||||||
MQTT_TOPIC: " "
|
|
||||||
HUB_HOST: "store"
|
|
||||||
HUB_PORT: 8000
|
|
||||||
HUB_MQTT_BROKER_HOST: "mqtt"
|
|
||||||
HUB_MQTT_BROKER_PORT: 1883
|
|
||||||
HUB_MQTT_TOPIC: "processed_data_topic"
|
|
||||||
networks:
|
|
||||||
mqtt_network:
|
|
||||||
edge_hub:
|
|
||||||
|
|
||||||
|
|
||||||
networks:
|
|
||||||
mqtt_network:
|
|
||||||
db_network:
|
|
||||||
edge_hub:
|
|
||||||
hub:
|
|
||||||
hub_store:
|
|
||||||
hub_redis:
|
|
||||||
|
|
||||||
|
|
||||||
volumes:
|
|
||||||
postgres_data:
|
|
||||||
pgadmin-data:
|
|
||||||
51
edge/main.py
51
edge/main.py
@@ -10,42 +10,51 @@ from config import (
|
|||||||
HUB_MQTT_BROKER_HOST,
|
HUB_MQTT_BROKER_HOST,
|
||||||
HUB_MQTT_BROKER_PORT,
|
HUB_MQTT_BROKER_PORT,
|
||||||
HUB_MQTT_TOPIC,
|
HUB_MQTT_TOPIC,
|
||||||
|
HUB_CONNECTION_TYPE,
|
||||||
)
|
)
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
# Configure logging settings
|
# Configure logging settings
|
||||||
logging.basicConfig(
|
logging.basicConfig(
|
||||||
level=logging.INFO, # Set the log level to INFO (you can use logging.DEBUG for more detailed logs)
|
level=logging.INFO,
|
||||||
format="[%(asctime)s] [%(levelname)s] [%(module)s] %(message)s",
|
format="[%(asctime)s] [%(levelname)s] [%(module)s] %(message)s",
|
||||||
handlers=[
|
handlers=[
|
||||||
logging.StreamHandler(), # Output log messages to the console
|
logging.StreamHandler(),
|
||||||
logging.FileHandler("app.log"), # Save log messages to a file
|
logging.FileHandler("app.log"),
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
# Create an instance of the StoreApiAdapter using the configuration
|
|
||||||
# hub_adapter = HubHttpAdapter(
|
# Logic to select the adapter based on configuration (SCRUM-93 & SCRUM-94)
|
||||||
# api_base_url=HUB_URL,
|
# This allows easy switching between HTTP and MQTT protocols
|
||||||
# )
|
if HUB_CONNECTION_TYPE.lower() == "http":
|
||||||
hub_adapter = HubMqttAdapter(
|
logging.info("Initializing HubHttpAdapter (SCRUM-93 integration)")
|
||||||
broker=HUB_MQTT_BROKER_HOST,
|
hub_adapter = HubHttpAdapter(
|
||||||
port=HUB_MQTT_BROKER_PORT,
|
api_base_url=HUB_URL,
|
||||||
topic=HUB_MQTT_TOPIC,
|
)
|
||||||
)
|
else:
|
||||||
# Create an instance of the AgentMQTTAdapter using the configuration
|
logging.info("Initializing HubMqttAdapter (SCRUM-94 integration)")
|
||||||
|
hub_adapter = HubMqttAdapter(
|
||||||
|
broker=HUB_MQTT_BROKER_HOST,
|
||||||
|
port=HUB_MQTT_BROKER_PORT,
|
||||||
|
topic=HUB_MQTT_TOPIC,
|
||||||
|
)
|
||||||
|
|
||||||
|
# Create an instance of the AgentMQTTAdapter using the selected hub adapter
|
||||||
|
# This adapter acts as a bridge between the Agent and the Hub
|
||||||
agent_adapter = AgentMQTTAdapter(
|
agent_adapter = AgentMQTTAdapter(
|
||||||
broker_host=MQTT_BROKER_HOST,
|
broker_host=MQTT_BROKER_HOST,
|
||||||
broker_port=MQTT_BROKER_PORT,
|
broker_port=MQTT_BROKER_PORT,
|
||||||
topic=MQTT_TOPIC,
|
topic=MQTT_TOPIC,
|
||||||
hub_gateway=hub_adapter,
|
hub_gateway=hub_adapter,
|
||||||
)
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Connect to the MQTT broker and start listening for messages
|
logging.info(f"Connecting to MQTT broker at {MQTT_BROKER_HOST}:{MQTT_BROKER_PORT}")
|
||||||
agent_adapter.connect()
|
agent_adapter.connect()
|
||||||
agent_adapter.start()
|
|
||||||
# Keep the system running indefinitely (you can add other logic as needed)
|
logging.info("Broker connection success. Waiting for data...")
|
||||||
while True:
|
agent_adapter.loop_forever()
|
||||||
pass
|
|
||||||
except KeyboardInterrupt:
|
except KeyboardInterrupt:
|
||||||
# Stop the MQTT adapter and exit gracefully if interrupted by the user
|
logging.info("Interrupt signal received. Shutting down...")
|
||||||
agent_adapter.stop()
|
agent_adapter.disconnect()
|
||||||
logging.info("System stopped.")
|
logging.info("Disconnected from MQTT broker.")
|
||||||
|
|||||||
2
hub/.gitignore
vendored
2
hub/.gitignore
vendored
@@ -1,2 +0,0 @@
|
|||||||
venv
|
|
||||||
__pycache__
|
|
||||||
@@ -3,9 +3,10 @@ FROM python:3.9-slim
|
|||||||
# Set the working directory inside the container
|
# Set the working directory inside the container
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
# Copy the requirements.txt file and install dependencies
|
# Copy the requirements.txt file and install dependencies
|
||||||
COPY requirements.txt .
|
COPY hub/requirements.txt .
|
||||||
|
|
||||||
RUN pip install --no-cache-dir -r requirements.txt
|
RUN pip install --no-cache-dir -r requirements.txt
|
||||||
# Copy the entire application into the container
|
# Copy the entire application into the container
|
||||||
COPY . .
|
COPY hub/. .
|
||||||
# Run the main.py script inside the container when it starts
|
# Run the main.py script inside the container when it starts
|
||||||
CMD ["uvicorn", "main:app", "--host", "0.0.0.0"]
|
CMD ["uvicorn", "main:app", "--host", "0.0.0.0"]
|
||||||
12
hub/Dockerfile-test
Normal file
12
hub/Dockerfile-test
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
# Use the official Python image as the base image
|
||||||
|
FROM python:3.9-slim
|
||||||
|
# Set the working directory inside the container
|
||||||
|
WORKDIR /app
|
||||||
|
# Copy the requirements.txt file and install dependencies
|
||||||
|
COPY hub/requirements.txt .
|
||||||
|
|
||||||
|
RUN pip install --no-cache-dir -r requirements.txt
|
||||||
|
# Copy the entire application into the container
|
||||||
|
COPY hub/. .
|
||||||
|
# Run the main.py script inside the container when it starts
|
||||||
|
CMD ["./test-entry.sh"]
|
||||||
@@ -13,12 +13,39 @@ class StoreApiAdapter(StoreGateway):
|
|||||||
def __init__(self, api_base_url):
|
def __init__(self, api_base_url):
|
||||||
self.api_base_url = api_base_url
|
self.api_base_url = api_base_url
|
||||||
|
|
||||||
|
def processed_agent_data_batch_to_payload(self, processed_agent_data_batch: List[ProcessedAgentData]):
|
||||||
|
if not processed_agent_data_batch:
|
||||||
|
return False
|
||||||
|
|
||||||
|
# Extract user_id from the first element
|
||||||
|
user_id = processed_agent_data_batch[0].agent_data.user_id
|
||||||
|
|
||||||
|
payload = {
|
||||||
|
"data": [item.model_dump(mode='json') for item in processed_agent_data_batch],
|
||||||
|
"user_id": user_id
|
||||||
|
}
|
||||||
|
|
||||||
|
return payload
|
||||||
|
|
||||||
def save_data(self, processed_agent_data_batch: List[ProcessedAgentData]):
|
def save_data(self, processed_agent_data_batch: List[ProcessedAgentData]):
|
||||||
"""
|
payload = self.processed_agent_data_batch_to_payload(processed_agent_data_batch)
|
||||||
Save the processed road data to the Store API.
|
|
||||||
Parameters:
|
if payload == False:
|
||||||
processed_agent_data_batch (dict): Processed road data to be saved.
|
return False
|
||||||
Returns:
|
|
||||||
bool: True if the data is successfully saved, False otherwise.
|
try:
|
||||||
"""
|
# Perform a POST request to the Store API with a 10-second timeout
|
||||||
# Implement it
|
response = requests.post(
|
||||||
|
f"{self.api_base_url}/processed_agent_data/",
|
||||||
|
json=payload,
|
||||||
|
timeout=10
|
||||||
|
)
|
||||||
|
if response.status_code == 200:
|
||||||
|
logging.info(f"Batch of {len(processed_agent_data_batch)} items sent to Store.")
|
||||||
|
return True
|
||||||
|
else:
|
||||||
|
logging.error(f"Store API error: {response.status_code} - {response.text}")
|
||||||
|
return False
|
||||||
|
except Exception as e:
|
||||||
|
logging.error(f"Failed to send data to Store: {e}")
|
||||||
|
return False
|
||||||
|
|||||||
41
hub/app/adapters/store_api_adapter_test.py
Normal file
41
hub/app/adapters/store_api_adapter_test.py
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
from app.adapters.store_api_adapter import StoreApiAdapter
|
||||||
|
from app.entities.agent_data import AccelerometerData, AgentData, GpsData
|
||||||
|
from app.entities.processed_agent_data import ProcessedAgentData
|
||||||
|
|
||||||
|
def _test_processed_agent_data_batch_to_payload():
|
||||||
|
processed_data_batch = [
|
||||||
|
ProcessedAgentData(road_state = "normal",
|
||||||
|
agent_data = AgentData(user_id = 1,
|
||||||
|
accelerometer = AccelerometerData(x = 0.1, y = 0.2, z = 0.3),
|
||||||
|
gps = GpsData(latitude = 10.123, longitude = 20.456),
|
||||||
|
timestamp = "2023-07-21T12:34:56Z")
|
||||||
|
),
|
||||||
|
ProcessedAgentData(road_state = "normal",
|
||||||
|
agent_data = AgentData(user_id = 2,
|
||||||
|
accelerometer = AccelerometerData(x = 0.1, y = 0.2, z = 0.3),
|
||||||
|
gps = GpsData(latitude = 10.123, longitude = 20.456),
|
||||||
|
timestamp = "2023-07-21T12:34:56Z")
|
||||||
|
),
|
||||||
|
ProcessedAgentData(road_state = "normal",
|
||||||
|
agent_data = AgentData(user_id = 3,
|
||||||
|
accelerometer = AccelerometerData(x = 0.1, y = 0.2, z = 0.3),
|
||||||
|
gps = GpsData(latitude = 10.123, longitude = 20.456),
|
||||||
|
timestamp = "2023-07-21T12:34:56Z")
|
||||||
|
),
|
||||||
|
]
|
||||||
|
|
||||||
|
res = StoreApiAdapter(None).processed_agent_data_batch_to_payload(processed_data_batch)
|
||||||
|
|
||||||
|
assert res["data"][0]["agent_data"]["user_id"] == 1
|
||||||
|
assert res["data"][1]["agent_data"]["user_id"] == 2
|
||||||
|
assert res["data"][2]["agent_data"]["user_id"] == 3
|
||||||
|
|
||||||
|
assert StoreApiAdapter(None).processed_agent_data_batch_to_payload([]) == False
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
test_functions = [i for i in dir() if i.startswith('_test_')]
|
||||||
|
|
||||||
|
for i in test_functions:
|
||||||
|
print(i)
|
||||||
|
eval(i)()
|
||||||
16
hub/main.py
16
hub/main.py
@@ -70,18 +70,20 @@ def on_message(client, userdata, msg):
|
|||||||
processed_agent_data = ProcessedAgentData.model_validate_json(
|
processed_agent_data = ProcessedAgentData.model_validate_json(
|
||||||
payload, strict=True
|
payload, strict=True
|
||||||
)
|
)
|
||||||
|
|
||||||
redis_client.lpush(
|
redis_client.lpush(
|
||||||
"processed_agent_data", processed_agent_data.model_dump_json()
|
"processed_agent_data", processed_agent_data.model_dump_json()
|
||||||
)
|
)
|
||||||
processed_agent_data_batch: List[ProcessedAgentData] = []
|
|
||||||
if redis_client.llen("processed_agent_data") >= BATCH_SIZE:
|
if redis_client.llen("processed_agent_data") >= BATCH_SIZE:
|
||||||
|
processed_agent_data_batch: List[ProcessedAgentData] = []
|
||||||
for _ in range(BATCH_SIZE):
|
for _ in range(BATCH_SIZE):
|
||||||
processed_agent_data = ProcessedAgentData.model_validate_json(
|
raw_data = redis_client.lpop("processed_agent_data")
|
||||||
redis_client.lpop("processed_agent_data")
|
if raw_data:
|
||||||
)
|
data_item = ProcessedAgentData.model_validate_json(raw_data)
|
||||||
processed_agent_data_batch.append(processed_agent_data)
|
processed_agent_data_batch.append(data_item)
|
||||||
store_adapter.save_data(processed_agent_data_batch=processed_agent_data_batch)
|
|
||||||
|
store_adapter.save_data(processed_agent_data_batch=processed_agent_data_batch)
|
||||||
|
|
||||||
return {"status": "ok"}
|
return {"status": "ok"}
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.info(f"Error processing MQTT message: {e}")
|
logging.info(f"Error processing MQTT message: {e}")
|
||||||
|
|||||||
3
hub/test-entry.sh
Executable file
3
hub/test-entry.sh
Executable file
@@ -0,0 +1,3 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
PYTHONPATH=$PWD python3 app/adapters/store_api_adapter_test.py
|
||||||
5
store/.gitignore
vendored
5
store/.gitignore
vendored
@@ -1,5 +0,0 @@
|
|||||||
venv
|
|
||||||
__pycache__
|
|
||||||
.idea
|
|
||||||
|
|
||||||
.idea/
|
|
||||||
@@ -3,9 +3,10 @@ FROM python:latest
|
|||||||
# Set the working directory inside the container
|
# Set the working directory inside the container
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
# Copy the requirements.txt file and install dependencies
|
# Copy the requirements.txt file and install dependencies
|
||||||
COPY requirements.txt .
|
COPY store/requirements.txt .
|
||||||
|
|
||||||
RUN pip install --no-cache-dir -r requirements.txt
|
RUN pip install --no-cache-dir -r requirements.txt
|
||||||
# Copy the entire application into the container
|
# Copy the entire application into the container
|
||||||
COPY . .
|
COPY store/. .
|
||||||
# Run the main.py script inside the container when it starts
|
# Run the main.py script inside the container when it starts
|
||||||
CMD ["uvicorn", "main:app", "--host", "0.0.0.0"]
|
CMD ["uvicorn", "main:app", "--host", "0.0.0.0"]
|
||||||
13
store/Dockerfile-test
Normal file
13
store/Dockerfile-test
Normal file
@@ -0,0 +1,13 @@
|
|||||||
|
# Use the official Python image as the base image
|
||||||
|
FROM python:latest
|
||||||
|
# Set the working directory inside the container
|
||||||
|
WORKDIR /app
|
||||||
|
# Copy the requirements.txt file and install dependencies
|
||||||
|
COPY store/requirements.txt .
|
||||||
|
|
||||||
|
RUN pip install --no-cache-dir -r requirements.txt
|
||||||
|
# Copy the entire application into the container
|
||||||
|
COPY store/. .
|
||||||
|
# Run the main.py script inside the container when it starts
|
||||||
|
#CMD ["uvicorn", "main:app", "--host", "0.0.0.0"]
|
||||||
|
CMD ["./test-entry.sh"]
|
||||||
@@ -1,59 +0,0 @@
|
|||||||
name: "road_vision__database"
|
|
||||||
services:
|
|
||||||
postgres_db:
|
|
||||||
image: postgres:17
|
|
||||||
container_name: postgres_db
|
|
||||||
restart: always
|
|
||||||
environment:
|
|
||||||
POSTGRES_USER: user
|
|
||||||
POSTGRES_PASSWORD: pass
|
|
||||||
POSTGRES_DB: test_db
|
|
||||||
volumes:
|
|
||||||
- postgres_data:/var/lib/postgresql/data
|
|
||||||
- ./db/structure.sql:/docker-entrypoint-initdb.d/structure.sql
|
|
||||||
ports:
|
|
||||||
- "5432:5432"
|
|
||||||
networks:
|
|
||||||
db_network:
|
|
||||||
|
|
||||||
|
|
||||||
pgadmin:
|
|
||||||
container_name: pgadmin4
|
|
||||||
image: dpage/pgadmin4
|
|
||||||
restart: always
|
|
||||||
environment:
|
|
||||||
PGADMIN_DEFAULT_EMAIL: admin@admin.com
|
|
||||||
PGADMIN_DEFAULT_PASSWORD: root
|
|
||||||
volumes:
|
|
||||||
- pgadmin-data:/var/lib/pgadmin
|
|
||||||
ports:
|
|
||||||
- "5050:80"
|
|
||||||
networks:
|
|
||||||
db_network:
|
|
||||||
|
|
||||||
|
|
||||||
store:
|
|
||||||
container_name: store
|
|
||||||
build: ..
|
|
||||||
depends_on:
|
|
||||||
- postgres_db
|
|
||||||
restart: always
|
|
||||||
environment:
|
|
||||||
POSTGRES_USER: user
|
|
||||||
POSTGRES_PASSWORD: pass
|
|
||||||
POSTGRES_DB: test_db
|
|
||||||
POSTGRES_HOST: postgres_db
|
|
||||||
POSTGRES_PORT: 5432
|
|
||||||
ports:
|
|
||||||
- "8000:8000"
|
|
||||||
networks:
|
|
||||||
db_network:
|
|
||||||
|
|
||||||
|
|
||||||
networks:
|
|
||||||
db_network:
|
|
||||||
|
|
||||||
|
|
||||||
volumes:
|
|
||||||
postgres_data:
|
|
||||||
pgadmin-data:
|
|
||||||
@@ -33,41 +33,48 @@ processed_agent_data = Table(
|
|||||||
)
|
)
|
||||||
|
|
||||||
# WebSocket subscriptions
|
# WebSocket subscriptions
|
||||||
subscriptions: Dict[int, Set[WebSocket]] = {}
|
subscriptions: Set[WebSocket] = set()
|
||||||
|
|
||||||
|
|
||||||
# FastAPI WebSocket endpoint
|
# FastAPI WebSocket endpoint
|
||||||
@app.websocket("/ws/{user_id}")
|
@app.websocket("/ws/{user_id}")
|
||||||
async def websocket_endpoint(websocket: WebSocket, user_id: int):
|
async def websocket_endpoint(websocket: WebSocket, user_id: int):
|
||||||
await websocket.accept()
|
await websocket.accept()
|
||||||
if user_id not in subscriptions:
|
|
||||||
subscriptions[user_id] = set()
|
subscriptions.add(websocket)
|
||||||
subscriptions[user_id].add(websocket)
|
|
||||||
try:
|
try:
|
||||||
|
# send already available data
|
||||||
|
r = processed_agent_data.select().order_by(processed_agent_data.c.timestamp)
|
||||||
|
stored_data = SessionLocal().execute(r).fetchall()
|
||||||
|
|
||||||
|
jsonable_data = [{c.name: getattr(i, c.name) for c in processed_agent_data.columns} for i in stored_data]
|
||||||
|
for i in jsonable_data:
|
||||||
|
i['timestamp'] = i['timestamp'].strftime("%Y-%m-%dT%H:%M:%SZ")
|
||||||
|
|
||||||
|
for i in jsonable_data:
|
||||||
|
await websocket.send_json(json.dumps([i]))
|
||||||
|
|
||||||
|
# receive forever
|
||||||
while True:
|
while True:
|
||||||
await websocket.receive_text()
|
await websocket.receive_text()
|
||||||
except WebSocketDisconnect:
|
except WebSocketDisconnect:
|
||||||
subscriptions[user_id].remove(websocket)
|
subscriptions.remove(websocket)
|
||||||
|
|
||||||
|
|
||||||
# Function to send data to subscribed users
|
# Function to send data to subscribed users
|
||||||
async def send_data_to_subscribers(user_id: int, data):
|
async def send_data_to_subscribers(data):
|
||||||
if user_id in subscriptions:
|
for websocket in subscriptions:
|
||||||
for websocket in subscriptions[user_id]:
|
await websocket.send_json(json.dumps([data]))
|
||||||
await websocket.send_json(json.dumps(data))
|
|
||||||
|
|
||||||
|
|
||||||
# FastAPI CRUDL endpoints
|
# FastAPI CRUDL endpoints
|
||||||
|
|
||||||
|
def ProcessedAgentData_to_td(data: List[ProcessedAgentData]):
|
||||||
@app.post("/processed_agent_data/")
|
return [
|
||||||
async def create_processed_agent_data(data: List[ProcessedAgentData], user_id: int = Body(..., embed=True)):
|
|
||||||
session = SessionLocal()
|
|
||||||
try:
|
|
||||||
created_data = [
|
|
||||||
{
|
{
|
||||||
"road_state": item.road_state,
|
"road_state": item.road_state,
|
||||||
"user_id": user_id,
|
"user_id": item.agent_data.user_id,
|
||||||
"x": item.agent_data.accelerometer.x,
|
"x": item.agent_data.accelerometer.x,
|
||||||
"y": item.agent_data.accelerometer.y,
|
"y": item.agent_data.accelerometer.y,
|
||||||
"z": item.agent_data.accelerometer.z,
|
"z": item.agent_data.accelerometer.z,
|
||||||
@@ -77,13 +84,20 @@ async def create_processed_agent_data(data: List[ProcessedAgentData], user_id: i
|
|||||||
}
|
}
|
||||||
for item in data
|
for item in data
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
|
@app.post("/processed_agent_data/")
|
||||||
|
async def create_processed_agent_data(data: List[ProcessedAgentData], user_id: int = Body(..., embed=True)):
|
||||||
|
session = SessionLocal()
|
||||||
|
try:
|
||||||
|
created_data = ProcessedAgentData_to_td(data)
|
||||||
stmt = processed_agent_data.insert().values(created_data).returning(processed_agent_data)
|
stmt = processed_agent_data.insert().values(created_data).returning(processed_agent_data)
|
||||||
result = session.execute(stmt)
|
result = session.execute(stmt)
|
||||||
created_records = [dict(row._mapping) for row in result.fetchall()]
|
created_records = [dict(row._mapping) for row in result.fetchall()]
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
||||||
for record in created_records:
|
for record in sorted(created_records, key = lambda x: x['timestamp']):
|
||||||
await send_data_to_subscribers(user_id, jsonable_encoder(record))
|
await send_data_to_subscribers(jsonable_encoder(record))
|
||||||
return created_records
|
return created_records
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
session.rollback()
|
session.rollback()
|
||||||
|
|||||||
3
store/test-entry.sh
Executable file
3
store/test-entry.sh
Executable file
@@ -0,0 +1,3 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
PYTHONPATH=$PWD python3 test/main_test.py
|
||||||
39
store/test/main_test.py
Normal file
39
store/test/main_test.py
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
from schemas import AccelerometerData, AgentData, GpsData, ProcessedAgentData
|
||||||
|
|
||||||
|
import main
|
||||||
|
|
||||||
|
def _test_ProcessedAgentData_to_td():
|
||||||
|
processed_data_batch = [
|
||||||
|
ProcessedAgentData(road_state = "normal",
|
||||||
|
agent_data = AgentData(user_id = 1,
|
||||||
|
accelerometer = AccelerometerData(x = 0.1, y = 0.2, z = 0.3),
|
||||||
|
gps = GpsData(latitude = 10.123, longitude = 20.456),
|
||||||
|
timestamp = "2023-07-21T12:34:56Z")
|
||||||
|
),
|
||||||
|
ProcessedAgentData(road_state = "normal",
|
||||||
|
agent_data = AgentData(user_id = 2,
|
||||||
|
accelerometer = AccelerometerData(x = 0.1, y = 0.2, z = 0.3),
|
||||||
|
gps = GpsData(latitude = 10.123, longitude = 20.456),
|
||||||
|
timestamp = "2023-07-21T12:34:56Z")
|
||||||
|
),
|
||||||
|
ProcessedAgentData(road_state = "normal",
|
||||||
|
agent_data = AgentData(user_id = 3,
|
||||||
|
accelerometer = AccelerometerData(x = 0.1, y = 0.2, z = 0.3),
|
||||||
|
gps = GpsData(latitude = 10.123, longitude = 20.456),
|
||||||
|
timestamp = "2023-07-21T12:34:56Z")
|
||||||
|
),
|
||||||
|
]
|
||||||
|
|
||||||
|
res = main.ProcessedAgentData_to_td(processed_data_batch)
|
||||||
|
|
||||||
|
assert res[0]["user_id"] == 1
|
||||||
|
assert res[1]["user_id"] == 2
|
||||||
|
assert res[2]["user_id"] == 3
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
test_functions = [i for i in dir() if i.startswith('_test_')]
|
||||||
|
|
||||||
|
for i in test_functions:
|
||||||
|
print(i)
|
||||||
|
eval(i)()
|
||||||
23
utils/check-up.py
Normal file
23
utils/check-up.py
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
import sys
|
||||||
|
import os
|
||||||
|
|
||||||
|
print("Checking for dead containers...")
|
||||||
|
|
||||||
|
l = [i for i in sys.stdin.read().split("\n") if i]
|
||||||
|
header, statuses = l[0], l[1:]
|
||||||
|
|
||||||
|
status_index = header.find('STATUS')
|
||||||
|
name_index = header.find('NAMES')
|
||||||
|
|
||||||
|
exit_code = 0
|
||||||
|
|
||||||
|
for i in statuses:
|
||||||
|
if not i[status_index:].startswith("Up "):
|
||||||
|
service_name = i[name_index:]
|
||||||
|
print(f"Crash detected in {service_name}")
|
||||||
|
print(f"docker logs for the container:\n")
|
||||||
|
os.system(f"docker logs {i.split(' ')[0]}")
|
||||||
|
print()
|
||||||
|
exit_code = 1
|
||||||
|
|
||||||
|
sys.exit(exit_code)
|
||||||
Reference in New Issue
Block a user