mirror of
https://github.com/home-assistant/supervisor.git
synced 2025-08-23 16:09:21 +00:00
Compare commits
216 Commits
Author | SHA1 | Date | |
---|---|---|---|
![]() |
8233083392 | ||
![]() |
106378d1d0 | ||
![]() |
01d18d5ff3 | ||
![]() |
6d23f3bd1c | ||
![]() |
ef96579a29 | ||
![]() |
44f0a9f21a | ||
![]() |
d854307acb | ||
![]() |
334b41de71 | ||
![]() |
1da50eab7a | ||
![]() |
b119a42f4d | ||
![]() |
99aa438817 | ||
![]() |
99fa91f480 | ||
![]() |
93969d264d | ||
![]() |
711e199977 | ||
![]() |
4e645332c3 | ||
![]() |
df8afb3337 | ||
![]() |
255a33fc08 | ||
![]() |
d15b6f0294 | ||
![]() |
1aa24e40ae | ||
![]() |
c0bde4a488 | ||
![]() |
2a09b70294 | ||
![]() |
e35b0a54c1 | ||
![]() |
8287330c67 | ||
![]() |
6b16da93cd | ||
![]() |
c1cd9bba45 | ||
![]() |
e33420f26e | ||
![]() |
abd9683e11 | ||
![]() |
8cbeabbe21 | ||
![]() |
df7d988d2f | ||
![]() |
544c009b9c | ||
![]() |
b2e0babc60 | ||
![]() |
f7c79cbd3a | ||
![]() |
587e9618da | ||
![]() |
cb2dd3b81c | ||
![]() |
8d4dd7de3f | ||
![]() |
6927c989d0 | ||
![]() |
97853d1691 | ||
![]() |
0cdef0d118 | ||
![]() |
0b17ffc243 | ||
![]() |
c516d46f16 | ||
![]() |
cb8ec22b6d | ||
![]() |
4a5fbd79c1 | ||
![]() |
b636a03567 | ||
![]() |
c96faf7c0a | ||
![]() |
2e1cd4076a | ||
![]() |
9984a638ba | ||
![]() |
a492bccc03 | ||
![]() |
e7a0e0f565 | ||
![]() |
030e081d45 | ||
![]() |
8537536368 | ||
![]() |
f03f323aac | ||
![]() |
58c0c67796 | ||
![]() |
f5e196a663 | ||
![]() |
808df68e57 | ||
![]() |
fa51c2e6e9 | ||
![]() |
ba3760e770 | ||
![]() |
ad1a8557b8 | ||
![]() |
fe91f812d9 | ||
![]() |
4cc11305c7 | ||
![]() |
898c0330c8 | ||
![]() |
33e5f94f1f | ||
![]() |
da4ee63890 | ||
![]() |
d34203b133 | ||
![]() |
23addfb9a6 | ||
![]() |
81e1227a7b | ||
![]() |
75be8666a6 | ||
![]() |
6031a60084 | ||
![]() |
39d5785118 | ||
![]() |
bddcdcadb2 | ||
![]() |
3eac6a3366 | ||
![]() |
3c7b962cf9 | ||
![]() |
bd756e2a9c | ||
![]() |
e7920bee2a | ||
![]() |
ebcc21370e | ||
![]() |
34c4acf199 | ||
![]() |
47e45dfc9f | ||
![]() |
2ecea7c1b4 | ||
![]() |
5c0eccd12f | ||
![]() |
f34ab9402b | ||
![]() |
2569a82caf | ||
![]() |
4bdd256000 | ||
![]() |
6f4f6338c5 | ||
![]() |
7cb72b55a8 | ||
![]() |
1a9a08cbfb | ||
![]() |
237ee0363d | ||
![]() |
86180ddc34 | ||
![]() |
eed41d30ec | ||
![]() |
0b0fd6b910 | ||
![]() |
1f887b47ab | ||
![]() |
affd8057ca | ||
![]() |
7a8ee2c46a | ||
![]() |
35fe1f464c | ||
![]() |
0955bafebd | ||
![]() |
2e0c540c63 | ||
![]() |
6e9ef17a28 | ||
![]() |
eb3cdbfeb9 | ||
![]() |
f4cb16ad09 | ||
![]() |
956af2bd62 | ||
![]() |
b76cd5c004 | ||
![]() |
61d9301dcc | ||
![]() |
2ded05be83 | ||
![]() |
899d6766c5 | ||
![]() |
c67d57cef4 | ||
![]() |
b5cca7d341 | ||
![]() |
8919f13911 | ||
![]() |
990ae49608 | ||
![]() |
c2ba02722c | ||
![]() |
5bd1957337 | ||
![]() |
f59f0793bc | ||
![]() |
63b96700e0 | ||
![]() |
dffbcc2c7e | ||
![]() |
0dbe1ecc2a | ||
![]() |
da8526fcec | ||
![]() |
933b6f4d1e | ||
![]() |
16f2dfeebd | ||
![]() |
bc6eb5cab4 | ||
![]() |
8833845b2e | ||
![]() |
391be6afac | ||
![]() |
a4f74676b6 | ||
![]() |
600b32f75b | ||
![]() |
f199a5cf95 | ||
![]() |
5896fde441 | ||
![]() |
9998f9720f | ||
![]() |
f37589daa6 | ||
![]() |
ce2513f175 | ||
![]() |
1a4c5d24a4 | ||
![]() |
886d202f39 | ||
![]() |
5a42019ed7 | ||
![]() |
354093c121 | ||
![]() |
aa9c300d7c | ||
![]() |
d9ad5daae3 | ||
![]() |
4680ba6d0d | ||
![]() |
1423062ac3 | ||
![]() |
a036096684 | ||
![]() |
a1c443a6f2 | ||
![]() |
e6e1367cd6 | ||
![]() |
303e741289 | ||
![]() |
79cc23e273 | ||
![]() |
046ce0230a | ||
![]() |
33a66bee01 | ||
![]() |
f76749a933 | ||
![]() |
385af5bef5 | ||
![]() |
19b72b1a79 | ||
![]() |
f6048467ad | ||
![]() |
dbe6b860c7 | ||
![]() |
0a1e6b3e2d | ||
![]() |
f178dde589 | ||
![]() |
545d45ecf0 | ||
![]() |
c333f94cfa | ||
![]() |
76a999f650 | ||
![]() |
d2f8e35622 | ||
![]() |
1e4ed9c9d1 | ||
![]() |
57c21b4eb5 | ||
![]() |
088cc3ef15 | ||
![]() |
9e326f6324 | ||
![]() |
5840290df7 | ||
![]() |
58fdabb8ff | ||
![]() |
b8fc002fbc | ||
![]() |
d5e349266b | ||
![]() |
4c135dd617 | ||
![]() |
a98a0d1a1a | ||
![]() |
768b4d2b1a | ||
![]() |
ff640c598d | ||
![]() |
c76408e4e8 | ||
![]() |
2e168d089c | ||
![]() |
a61311e928 | ||
![]() |
85ffba90b2 | ||
![]() |
6623ec9bbc | ||
![]() |
7a4ed4029c | ||
![]() |
d4c52ce298 | ||
![]() |
c0aab8497f | ||
![]() |
cabe5b143f | ||
![]() |
3a791bace6 | ||
![]() |
19dfdb094b | ||
![]() |
ec5caa483e | ||
![]() |
e6967f8db5 | ||
![]() |
759356ff2e | ||
![]() |
4d8dbdb486 | ||
![]() |
cddb40a104 | ||
![]() |
36128d119a | ||
![]() |
dadb72aca8 | ||
![]() |
390d4fa6c7 | ||
![]() |
9559c39351 | ||
![]() |
9d95b70534 | ||
![]() |
3bad896978 | ||
![]() |
9f406df129 | ||
![]() |
a9b4174590 | ||
![]() |
9109e3803b | ||
![]() |
422dd78489 | ||
![]() |
9e1d6c9d2b | ||
![]() |
c8e3f2b48a | ||
![]() |
a287f52e47 | ||
![]() |
76952db3eb | ||
![]() |
871721f04b | ||
![]() |
3c0ebdf643 | ||
![]() |
0e258a4ae0 | ||
![]() |
645a8e2372 | ||
![]() |
c6cc8adbb7 | ||
![]() |
dd38c73b85 | ||
![]() |
c916314704 | ||
![]() |
e0dcce5895 | ||
![]() |
906616e224 | ||
![]() |
db20ea95d9 | ||
![]() |
d142ea5d23 | ||
![]() |
5d52404dab | ||
![]() |
43f4b36cfe | ||
![]() |
0393db19e6 | ||
![]() |
b197578df4 | ||
![]() |
ed428c0df4 | ||
![]() |
d38707821c | ||
![]() |
cfb392054e | ||
![]() |
0ea65efeb3 | ||
![]() |
c4ce7d1a74 | ||
![]() |
7ac95b98bc | ||
![]() |
f8413d8d63 | ||
![]() |
709b80b864 |
3
.gitmodules
vendored
Normal file
3
.gitmodules
vendored
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
[submodule "home-assistant-polymer"]
|
||||||
|
path = home-assistant-polymer
|
||||||
|
url = https://github.com/home-assistant/home-assistant-polymer
|
165
API.md
165
API.md
@@ -22,29 +22,72 @@ On success
|
|||||||
|
|
||||||
### HassIO
|
### HassIO
|
||||||
|
|
||||||
- `/supervisor/ping`
|
- GET `/supervisor/ping`
|
||||||
|
|
||||||
- `/supervisor/info`
|
- GET `/supervisor/info`
|
||||||
|
|
||||||
|
The addons from `addons` are only installed one.
|
||||||
|
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"version": "INSTALL_VERSION",
|
"version": "INSTALL_VERSION",
|
||||||
"last_version": "LAST_VERSION",
|
"last_version": "LAST_VERSION",
|
||||||
|
"arch": "armhf|aarch64|i386|amd64",
|
||||||
"beta_channel": "true|false",
|
"beta_channel": "true|false",
|
||||||
|
"timezone": "TIMEZONE",
|
||||||
"addons": [
|
"addons": [
|
||||||
{
|
{
|
||||||
"name": "xy bla",
|
"name": "xy bla",
|
||||||
"slug": "xy",
|
"slug": "xy",
|
||||||
|
"description": "description",
|
||||||
|
"arch": ["armhf", "aarch64", "i386", "amd64"],
|
||||||
|
"repository": "12345678|null",
|
||||||
|
"version": "LAST_VERSION",
|
||||||
|
"installed": "INSTALL_VERSION",
|
||||||
|
"detached": "bool",
|
||||||
|
"build": "bool",
|
||||||
|
"url": "null|url"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"addons_repositories": [
|
||||||
|
"REPO_URL"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
- GET `/supervisor/addons`
|
||||||
|
|
||||||
|
Get all available addons
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"addons": [
|
||||||
|
{
|
||||||
|
"name": "xy bla",
|
||||||
|
"slug": "xy",
|
||||||
|
"description": "description",
|
||||||
|
"arch": ["armhf", "aarch64", "i386", "amd64"],
|
||||||
|
"repository": "core|local|REP_ID",
|
||||||
"version": "LAST_VERSION",
|
"version": "LAST_VERSION",
|
||||||
"installed": "none|INSTALL_VERSION",
|
"installed": "none|INSTALL_VERSION",
|
||||||
"dedicated": "bool",
|
"detached": "bool",
|
||||||
"description": "description"
|
"build": "bool",
|
||||||
|
"url": "null|url"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"repositories": [
|
||||||
|
{
|
||||||
|
"slug": "12345678",
|
||||||
|
"name": "Repitory Name",
|
||||||
|
"source": "URL_OF_REPOSITORY",
|
||||||
|
"url": "null|WEBSITE",
|
||||||
|
"maintainer": "null|BLA BLU <fla@dld.ch>"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/supervisor/update`
|
- POST `/supervisor/update`
|
||||||
Optional:
|
Optional:
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
@@ -52,28 +95,66 @@ Optional:
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/supervisor/option`
|
- POST `/supervisor/options`
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"beta_channel": "true|false"
|
"beta_channel": "true|false",
|
||||||
|
"timezone": "TIMEZONE",
|
||||||
|
"addons_repositories": [
|
||||||
|
"REPO_URL"
|
||||||
|
]
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/supervisor/reload`
|
- POST `/supervisor/reload`
|
||||||
|
|
||||||
Reload addons/version.
|
Reload addons/version.
|
||||||
|
|
||||||
- `/supervisor/logs`
|
- GET `/supervisor/logs`
|
||||||
|
|
||||||
Output the raw docker log
|
Output the raw docker log
|
||||||
|
|
||||||
|
### Security
|
||||||
|
|
||||||
|
- GET `/security/info`
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"initialize": "bool",
|
||||||
|
"totp": "bool"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
- POST `/security/options`
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"password": "xy"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
- POST `/security/totp`
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"password": "xy"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Return QR-Code
|
||||||
|
|
||||||
|
- POST `/security/session`
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"password": "xy",
|
||||||
|
"totp": "null|123456"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
### Host
|
### Host
|
||||||
|
|
||||||
- `/host/shutdown`
|
- POST `/host/shutdown`
|
||||||
|
|
||||||
- `/host/reboot`
|
- POST `/host/reboot`
|
||||||
|
|
||||||
- `/host/info`
|
- GET `/host/info`
|
||||||
See HostControl info command.
|
See HostControl info command.
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
@@ -86,7 +167,7 @@ See HostControl info command.
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/host/update`
|
- POST `/host/update`
|
||||||
Optional:
|
Optional:
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
@@ -96,9 +177,14 @@ Optional:
|
|||||||
|
|
||||||
### Network
|
### Network
|
||||||
|
|
||||||
- `/network/info`
|
- GET `/network/info`
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"hostname": ""
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
- `/network/options`
|
- POST `/network/options`
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"hostname": "",
|
"hostname": "",
|
||||||
@@ -112,16 +198,17 @@ Optional:
|
|||||||
|
|
||||||
### HomeAssistant
|
### HomeAssistant
|
||||||
|
|
||||||
- `/homeassistant/info`
|
- GET `/homeassistant/info`
|
||||||
|
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"version": "INSTALL_VERSION",
|
"version": "INSTALL_VERSION",
|
||||||
"last_version": "LAST_VERSION"
|
"last_version": "LAST_VERSION",
|
||||||
|
"devices": []
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/homeassistant/update`
|
- POST `/homeassistant/update`
|
||||||
Optional:
|
Optional:
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
@@ -129,24 +216,39 @@ Optional:
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/homeassistant/logs`
|
- GET `/homeassistant/logs`
|
||||||
|
|
||||||
Output the raw docker log
|
Output the raw docker log
|
||||||
|
|
||||||
### REST API addons
|
- POST `/homeassistant/restart`
|
||||||
|
|
||||||
- `/addons/{addon}/info`
|
- POST `/homeassistant/options`
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
|
"devices": [],
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### REST API addons
|
||||||
|
|
||||||
|
- GET `/addons/{addon}/info`
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"name": "xy bla",
|
||||||
|
"description": "description",
|
||||||
|
"url": "null|url of addon",
|
||||||
|
"detached": "bool",
|
||||||
|
"repository": "12345678|null",
|
||||||
"version": "VERSION",
|
"version": "VERSION",
|
||||||
"last_version": "LAST_VERSION",
|
"last_version": "LAST_VERSION",
|
||||||
"state": "started|stopped",
|
"state": "started|stopped",
|
||||||
"boot": "auto|manual",
|
"boot": "auto|manual",
|
||||||
|
"build": "bool",
|
||||||
"options": {},
|
"options": {},
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/addons/{addon}/options`
|
- POST `/addons/{addon}/options`
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"boot": "auto|manual",
|
"boot": "auto|manual",
|
||||||
@@ -154,11 +256,11 @@ Output the raw docker log
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/addons/{addon}/start`
|
- POST `/addons/{addon}/start`
|
||||||
|
|
||||||
- `/addons/{addon}/stop`
|
- POST `/addons/{addon}/stop`
|
||||||
|
|
||||||
- `/addons/{addon}/install`
|
- POST `/addons/{addon}/install`
|
||||||
Optional:
|
Optional:
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
@@ -166,9 +268,9 @@ Optional:
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/addons/{addon}/uninstall`
|
- POST `/addons/{addon}/uninstall`
|
||||||
|
|
||||||
- `/addons/{addon}/update`
|
- POST `/addons/{addon}/update`
|
||||||
Optional:
|
Optional:
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
@@ -176,10 +278,12 @@ Optional:
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
- `/addons/{addon}/logs`
|
- GET `/addons/{addon}/logs`
|
||||||
|
|
||||||
Output the raw docker log
|
Output the raw docker log
|
||||||
|
|
||||||
|
- POST `/addons/{addon}/restart`
|
||||||
|
|
||||||
## Host Control
|
## Host Control
|
||||||
|
|
||||||
Communicate over unix socket with a host daemon.
|
Communicate over unix socket with a host daemon.
|
||||||
@@ -192,8 +296,10 @@ Communicate over unix socket with a host daemon.
|
|||||||
# shutdown
|
# shutdown
|
||||||
# host-update [v]
|
# host-update [v]
|
||||||
|
|
||||||
|
# hostname xy
|
||||||
|
|
||||||
# network info
|
# network info
|
||||||
# network hostname xy
|
-> {}
|
||||||
# network wlan ssd xy
|
# network wlan ssd xy
|
||||||
# network wlan password xy
|
# network wlan password xy
|
||||||
# network int ip xy
|
# network int ip xy
|
||||||
@@ -205,6 +311,7 @@ features:
|
|||||||
- shutdown
|
- shutdown
|
||||||
- reboot
|
- reboot
|
||||||
- update
|
- update
|
||||||
|
- hostname
|
||||||
- network_info
|
- network_info
|
||||||
- network_control
|
- network_control
|
||||||
|
|
||||||
|
3
MANIFEST.in
Normal file
3
MANIFEST.in
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
include LICENSE.md
|
||||||
|
graft hassio
|
||||||
|
recursive-exclude * *.py[co]
|
56
README.md
56
README.md
@@ -1,56 +1,14 @@
|
|||||||
# HassIO
|
# HassIO
|
||||||
First private cloud solution for home automation.
|
### First private cloud solution for home automation.
|
||||||
|
|
||||||
It is a docker image (supervisor) they manage HomeAssistant docker and give a interface to control itself over UI. It have a own eco system with addons to extend the functionality in a easy way.
|
Hass.io is a Docker based system for managing your Home Assistant installation and related applications. The system is controlled via Home Assistant which communicates with the supervisor. The supervisor provides an API to manage the installation. This includes changing network settings or installing and updating software.
|
||||||
|
|
||||||
[HassIO-Addons](https://github.com/pvizeli/hassio-addons) | [HassIO-Build](https://github.com/pvizeli/hassio-build)
|

|
||||||
|
|
||||||
**HassIO is at the moment on development and not ready to use productive!**
|
[HassIO-Addons](https://github.com/home-assistant/hassio-addons) | [HassIO-Build](https://github.com/home-assistant/hassio-build)
|
||||||
|
|
||||||
## Feature in progress
|
**HassIO is under active development and is not ready yet for production use.**
|
||||||
- Backup/Restore
|
|
||||||
- MQTT addon
|
|
||||||
- DHCP-Server addon
|
|
||||||
|
|
||||||
# HomeAssistant
|
## Installation
|
||||||
|
|
||||||
## SSL
|
Installation instructions can be found at [https://home-assistant.io/hassio](https://home-assistant.io/hassio).
|
||||||
|
|
||||||
All addons they can create SSL certs do that in same schema. So you can put follow lines to your `configuration.yaml`.
|
|
||||||
```yaml
|
|
||||||
http:
|
|
||||||
ssl_certificate: /ssl/fullchain.pem
|
|
||||||
ssl_key: /ssl/privkey.pem
|
|
||||||
```
|
|
||||||
|
|
||||||
# Hardware Image
|
|
||||||
The image is based on ResinOS and Yocto Linux. It comes with the HassIO supervisor pre-installed. This includes support to update the supervisor over the air. After flashing your host OS will not require any more maintenance! The image does not include Home Assistant, instead it will downloaded when the image boots up for the first time.
|
|
||||||
|
|
||||||
Download can be found here: https://drive.google.com/drive/folders/0B2o1Uz6l1wVNbFJnb2gwNXJja28?usp=sharing
|
|
||||||
|
|
||||||
After extracting the archive, flash it to a drive using [Etcher](https://etcher.io/).
|
|
||||||
|
|
||||||
## History
|
|
||||||
- **0.1**: First techpreview with dumy supervisor (ResinOS 2.0.0-RC5)
|
|
||||||
- **0.2**: Fix some bugs and update it to HassIO 0.2
|
|
||||||
- **0.3**: Update HostControl and feature for HassIO 0.3 (ResinOS 2.0.0 / need reflash)
|
|
||||||
- **0.4**: Update HostControl and bring resinos OTA (resinhub) back (ResinOS 2.0.0-rev3)
|
|
||||||
|
|
||||||
## Configuring the image
|
|
||||||
You can configure the WiFi network that the image should connect to after flashing using [`resin-device-toolbox`](https://resinos.io/docs/raspberrypi3/gettingstarted/#install-resin-device-toolbox).
|
|
||||||
|
|
||||||
## Developer access to ResinOS host
|
|
||||||
Create an `authorized_keys` file in the boot partition of your SD card with your public key. After a boot it, you can acces your device as root over ssh on port 22222.
|
|
||||||
|
|
||||||
## Troubleshooting
|
|
||||||
|
|
||||||
Read logoutput from supervisor:
|
|
||||||
```bash
|
|
||||||
journalctl -f -u resin-supervisor.service
|
|
||||||
docker logs homeassistant
|
|
||||||
```
|
|
||||||
|
|
||||||
## Install on a own System
|
|
||||||
|
|
||||||
We have a installer to install HassIO on own linux device without our hardware image:
|
|
||||||
https://github.com/pvizeli/hassio-build/tree/master/install
|
|
||||||
|
@@ -1,11 +1,10 @@
|
|||||||
"""Init file for HassIO addons."""
|
"""Init file for HassIO addons."""
|
||||||
import asyncio
|
import asyncio
|
||||||
import logging
|
import logging
|
||||||
import os
|
|
||||||
import shutil
|
import shutil
|
||||||
|
|
||||||
from .data import AddonsData
|
from .data import AddonsData
|
||||||
from .git import AddonsRepo
|
from .git import AddonsRepoHassIO, AddonsRepoCustom
|
||||||
from ..const import STATE_STOPPED, STATE_STARTED
|
from ..const import STATE_STOPPED, STATE_STARTED
|
||||||
from ..dock.addon import DockerAddon
|
from ..dock.addon import DockerAddon
|
||||||
|
|
||||||
@@ -21,16 +20,29 @@ class AddonManager(AddonsData):
|
|||||||
|
|
||||||
self.loop = loop
|
self.loop = loop
|
||||||
self.dock = dock
|
self.dock = dock
|
||||||
self.repo = AddonsRepo(config, loop)
|
self.repositories = []
|
||||||
self.dockers = {}
|
self.dockers = {}
|
||||||
|
|
||||||
async def prepare(self, arch):
|
async def prepare(self, arch):
|
||||||
"""Startup addon management."""
|
"""Startup addon management."""
|
||||||
self.arch = arch
|
self.arch = arch
|
||||||
|
|
||||||
|
# init hassio repository
|
||||||
|
self.repositories.append(AddonsRepoHassIO(self.config, self.loop))
|
||||||
|
|
||||||
|
# init custom repositories
|
||||||
|
for url in self.config.addons_repositories:
|
||||||
|
self.repositories.append(
|
||||||
|
AddonsRepoCustom(self.config, self.loop, url))
|
||||||
|
|
||||||
# load addon repository
|
# load addon repository
|
||||||
if await self.repo.load():
|
tasks = [addon.load() for addon in self.repositories]
|
||||||
self.read_addons_repo()
|
if tasks:
|
||||||
|
await asyncio.wait(tasks, loop=self.loop)
|
||||||
|
|
||||||
|
# read data from repositories
|
||||||
|
self.read_data_from_repositories()
|
||||||
|
self.merge_update_config()
|
||||||
|
|
||||||
# load installed addons
|
# load installed addons
|
||||||
for addon in self.list_installed:
|
for addon in self.list_installed:
|
||||||
@@ -38,23 +50,53 @@ class AddonManager(AddonsData):
|
|||||||
self.config, self.loop, self.dock, self, addon)
|
self.config, self.loop, self.dock, self, addon)
|
||||||
await self.dockers[addon].attach()
|
await self.dockers[addon].attach()
|
||||||
|
|
||||||
|
async def add_git_repository(self, url):
|
||||||
|
"""Add a new custom repository."""
|
||||||
|
if url in self.config.addons_repositories:
|
||||||
|
_LOGGER.warning("Repository already exists %s", url)
|
||||||
|
return False
|
||||||
|
|
||||||
|
repo = AddonsRepoCustom(self.config, self.loop, url)
|
||||||
|
|
||||||
|
if not await repo.load():
|
||||||
|
_LOGGER.error("Can't load from repository %s", url)
|
||||||
|
return False
|
||||||
|
|
||||||
|
self.config.addons_repositories = url
|
||||||
|
self.repositories.append(repo)
|
||||||
|
return True
|
||||||
|
|
||||||
|
def drop_git_repository(self, url):
|
||||||
|
"""Remove a custom repository."""
|
||||||
|
for repo in self.repositories:
|
||||||
|
if repo.url == url:
|
||||||
|
self.repositories.remove(repo)
|
||||||
|
self.config.drop_addon_repository(url)
|
||||||
|
repo.remove()
|
||||||
|
return True
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
async def reload(self):
|
async def reload(self):
|
||||||
"""Update addons from repo and reload list."""
|
"""Update addons from repo and reload list."""
|
||||||
if not await self.repo.pull():
|
tasks = [addon.pull() for addon in self.repositories]
|
||||||
|
if not tasks:
|
||||||
return
|
return
|
||||||
self.read_addons_repo()
|
|
||||||
|
await asyncio.wait(tasks, loop=self.loop)
|
||||||
|
|
||||||
|
# read data from repositories
|
||||||
|
self.read_data_from_repositories()
|
||||||
|
self.merge_update_config()
|
||||||
|
|
||||||
# remove stalled addons
|
# remove stalled addons
|
||||||
for addon in self.list_removed:
|
for addon in self.list_detached:
|
||||||
_LOGGER.warning("Dedicated addon '%s' found!", addon)
|
_LOGGER.warning("Dedicated addon '%s' found!", addon)
|
||||||
|
|
||||||
async def auto_boot(self, start_type):
|
async def auto_boot(self, start_type):
|
||||||
"""Boot addons with mode auto."""
|
"""Boot addons with mode auto."""
|
||||||
boot_list = self.list_startup(start_type)
|
boot_list = self.list_startup(start_type)
|
||||||
tasks = []
|
tasks = [self.start(addon) for addon in boot_list]
|
||||||
|
|
||||||
for addon in boot_list:
|
|
||||||
tasks.append(self.loop.create_task(self.start(addon)))
|
|
||||||
|
|
||||||
_LOGGER.info("Startup %s run %d addons", start_type, len(tasks))
|
_LOGGER.info("Startup %s run %d addons", start_type, len(tasks))
|
||||||
if tasks:
|
if tasks:
|
||||||
@@ -66,14 +108,18 @@ class AddonManager(AddonsData):
|
|||||||
_LOGGER.error("Addon %s not exists for install", addon)
|
_LOGGER.error("Addon %s not exists for install", addon)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
if self.arch not in self.get_arch(addon):
|
||||||
|
_LOGGER.error("Addon %s not supported on %s", addon, self.arch)
|
||||||
|
return False
|
||||||
|
|
||||||
if self.is_installed(addon):
|
if self.is_installed(addon):
|
||||||
_LOGGER.error("Addon %s is already installed", addon)
|
_LOGGER.error("Addon %s is already installed", addon)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
if not os.path.isdir(self.path_data(addon)):
|
if not self.path_data(addon).is_dir():
|
||||||
_LOGGER.info("Create Home-Assistant addon data folder %s",
|
_LOGGER.info("Create Home-Assistant addon data folder %s",
|
||||||
self.path_data(addon))
|
self.path_data(addon))
|
||||||
os.mkdir(self.path_data(addon))
|
self.path_data(addon).mkdir()
|
||||||
|
|
||||||
addon_docker = DockerAddon(
|
addon_docker = DockerAddon(
|
||||||
self.config, self.loop, self.dock, self, addon)
|
self.config, self.loop, self.dock, self, addon)
|
||||||
@@ -99,10 +145,10 @@ class AddonManager(AddonsData):
|
|||||||
if not await self.dockers[addon].remove():
|
if not await self.dockers[addon].remove():
|
||||||
return False
|
return False
|
||||||
|
|
||||||
if os.path.isdir(self.path_data(addon)):
|
if self.path_data(addon).is_dir():
|
||||||
_LOGGER.info("Remove Home-Assistant addon data folder %s",
|
_LOGGER.info("Remove Home-Assistant addon data folder %s",
|
||||||
self.path_data(addon))
|
self.path_data(addon))
|
||||||
shutil.rmtree(self.path_data(addon))
|
shutil.rmtree(str(self.path_data(addon)))
|
||||||
|
|
||||||
self.dockers.pop(addon)
|
self.dockers.pop(addon)
|
||||||
self.set_addon_uninstall(addon)
|
self.set_addon_uninstall(addon)
|
||||||
@@ -145,15 +191,25 @@ class AddonManager(AddonsData):
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
version = version or self.get_last_version(addon)
|
version = version or self.get_last_version(addon)
|
||||||
is_running = self.dockers[addon].is_running()
|
|
||||||
|
|
||||||
# update
|
# update
|
||||||
if await self.dockers[addon].update(version):
|
if not await self.dockers[addon].update(version):
|
||||||
self.set_addon_update(addon, version)
|
return False
|
||||||
if is_running:
|
|
||||||
await self.start(addon)
|
self.set_addon_update(addon, version)
|
||||||
return True
|
return True
|
||||||
return False
|
|
||||||
|
async def restart(self, addon):
|
||||||
|
"""Restart addon."""
|
||||||
|
if addon not in self.dockers:
|
||||||
|
_LOGGER.error("No docker found for addon %s", addon)
|
||||||
|
return False
|
||||||
|
|
||||||
|
if not self.write_addon_options(addon):
|
||||||
|
_LOGGER.error("Can't write options for addon %s", addon)
|
||||||
|
return False
|
||||||
|
|
||||||
|
return await self.dockers[addon].restart()
|
||||||
|
|
||||||
async def logs(self, addon):
|
async def logs(self, addon):
|
||||||
"""Return addons log output."""
|
"""Return addons log output."""
|
||||||
|
14
hassio/addons/built-in.json
Normal file
14
hassio/addons/built-in.json
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
{
|
||||||
|
"local": {
|
||||||
|
"slug": "local",
|
||||||
|
"name": "Local Add-Ons",
|
||||||
|
"url": "https://home-assistant.io/hassio",
|
||||||
|
"maintainer": "By our self"
|
||||||
|
},
|
||||||
|
"core": {
|
||||||
|
"slug": "core",
|
||||||
|
"name": "Built-in Add-Ons",
|
||||||
|
"url": "https://home-assistant.io/addons",
|
||||||
|
"maintainer": "Home Assistant authors"
|
||||||
|
}
|
||||||
|
}
|
@@ -1,24 +1,35 @@
|
|||||||
"""Init file for HassIO addons."""
|
"""Init file for HassIO addons."""
|
||||||
|
import copy
|
||||||
import logging
|
import logging
|
||||||
import glob
|
import json
|
||||||
|
from pathlib import Path, PurePath
|
||||||
|
import re
|
||||||
|
|
||||||
import voluptuous as vol
|
import voluptuous as vol
|
||||||
from voluptuous.humanize import humanize_error
|
from voluptuous.humanize import humanize_error
|
||||||
|
|
||||||
from .validate import validate_options, SCHEMA_ADDON_CONFIG
|
from .util import extract_hash_from_path
|
||||||
|
from .validate import (
|
||||||
|
validate_options, SCHEMA_ADDON_CONFIG, SCHEMA_REPOSITORY_CONFIG,
|
||||||
|
MAP_VOLUME)
|
||||||
from ..const import (
|
from ..const import (
|
||||||
FILE_HASSIO_ADDONS, ATTR_NAME, ATTR_VERSION, ATTR_SLUG, ATTR_DESCRIPTON,
|
FILE_HASSIO_ADDONS, ATTR_NAME, ATTR_VERSION, ATTR_SLUG, ATTR_DESCRIPTON,
|
||||||
ATTR_STARTUP, ATTR_BOOT, ATTR_MAP_SSL, ATTR_MAP_CONFIG, ATTR_OPTIONS,
|
ATTR_STARTUP, ATTR_BOOT, ATTR_MAP, ATTR_OPTIONS, ATTR_PORTS, BOOT_AUTO,
|
||||||
ATTR_PORTS, BOOT_AUTO, DOCKER_REPO, ATTR_INSTALLED, ATTR_SCHEMA,
|
ATTR_SCHEMA, ATTR_IMAGE, ATTR_REPOSITORY, ATTR_URL, ATTR_ARCH,
|
||||||
ATTR_IMAGE, ATTR_DEDICATED)
|
ATTR_LOCATON, ATTR_DEVICES, ATTR_ENVIRONMENT, ATTR_HOST_NETWORK,
|
||||||
|
ATTR_TMPFS, ATTR_PRIVILEGED)
|
||||||
from ..config import Config
|
from ..config import Config
|
||||||
from ..tools import read_json_file, write_json_file
|
from ..tools import read_json_file, write_json_file
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
ADDONS_REPO_PATTERN = "{}/*/config.json"
|
SYSTEM = 'system'
|
||||||
SYSTEM = "system"
|
USER = 'user'
|
||||||
USER = "user"
|
|
||||||
|
REPOSITORY_CORE = 'core'
|
||||||
|
REPOSITORY_LOCAL = 'local'
|
||||||
|
|
||||||
|
RE_VOLUME = re.compile(MAP_VOLUME)
|
||||||
|
|
||||||
|
|
||||||
class AddonsData(Config):
|
class AddonsData(Config):
|
||||||
@@ -30,7 +41,8 @@ class AddonsData(Config):
|
|||||||
self.config = config
|
self.config = config
|
||||||
self._system_data = self._data.get(SYSTEM, {})
|
self._system_data = self._data.get(SYSTEM, {})
|
||||||
self._user_data = self._data.get(USER, {})
|
self._user_data = self._data.get(USER, {})
|
||||||
self._current_data = {}
|
self._addons_cache = {}
|
||||||
|
self._repositories_data = {}
|
||||||
self.arch = None
|
self.arch = None
|
||||||
|
|
||||||
def save(self):
|
def save(self):
|
||||||
@@ -41,56 +53,131 @@ class AddonsData(Config):
|
|||||||
}
|
}
|
||||||
super().save()
|
super().save()
|
||||||
|
|
||||||
def read_addons_repo(self):
|
def read_data_from_repositories(self):
|
||||||
"""Read data from addons repository."""
|
"""Read data from addons repository."""
|
||||||
self._current_data = {}
|
self._addons_cache = {}
|
||||||
|
self._repositories_data = {}
|
||||||
|
|
||||||
self._read_addons_folder(self.config.path_addons_repo)
|
# read core repository
|
||||||
self._read_addons_folder(self.config.path_addons_custom)
|
self._read_addons_folder(
|
||||||
|
self.config.path_addons_core, REPOSITORY_CORE)
|
||||||
|
|
||||||
def _read_addons_folder(self, folder):
|
# read local repository
|
||||||
|
self._read_addons_folder(
|
||||||
|
self.config.path_addons_local, REPOSITORY_LOCAL)
|
||||||
|
|
||||||
|
# add built-in repositories information
|
||||||
|
self._set_builtin_repositories()
|
||||||
|
|
||||||
|
# read custom git repositories
|
||||||
|
for repository_element in self.config.path_addons_git.iterdir():
|
||||||
|
if repository_element.is_dir():
|
||||||
|
self._read_git_repository(repository_element)
|
||||||
|
|
||||||
|
def _read_git_repository(self, path):
|
||||||
|
"""Process a custom repository folder."""
|
||||||
|
slug = extract_hash_from_path(path)
|
||||||
|
repository_info = {ATTR_SLUG: slug}
|
||||||
|
|
||||||
|
# exists repository json
|
||||||
|
repository_file = Path(path, "repository.json")
|
||||||
|
try:
|
||||||
|
repository_info.update(SCHEMA_REPOSITORY_CONFIG(
|
||||||
|
read_json_file(repository_file)
|
||||||
|
))
|
||||||
|
|
||||||
|
except OSError:
|
||||||
|
_LOGGER.warning("Can't read repository information from %s",
|
||||||
|
repository_file)
|
||||||
|
return
|
||||||
|
|
||||||
|
except vol.Invalid:
|
||||||
|
_LOGGER.warning("Repository parse error %s", repository_file)
|
||||||
|
return
|
||||||
|
|
||||||
|
# process data
|
||||||
|
self._repositories_data[slug] = repository_info
|
||||||
|
self._read_addons_folder(path, slug)
|
||||||
|
|
||||||
|
def _read_addons_folder(self, path, repository):
|
||||||
"""Read data from addons folder."""
|
"""Read data from addons folder."""
|
||||||
pattern = ADDONS_REPO_PATTERN.format(folder)
|
for addon in path.glob("**/config.json"):
|
||||||
|
|
||||||
for addon in glob.iglob(pattern):
|
|
||||||
try:
|
try:
|
||||||
addon_config = read_json_file(addon)
|
addon_config = read_json_file(addon)
|
||||||
|
|
||||||
|
# validate
|
||||||
addon_config = SCHEMA_ADDON_CONFIG(addon_config)
|
addon_config = SCHEMA_ADDON_CONFIG(addon_config)
|
||||||
self._current_data[addon_config[ATTR_SLUG]] = addon_config
|
|
||||||
|
|
||||||
except (OSError, KeyError):
|
# Generate slug
|
||||||
|
addon_slug = "{}_{}".format(
|
||||||
|
repository, addon_config[ATTR_SLUG])
|
||||||
|
|
||||||
|
# store
|
||||||
|
addon_config[ATTR_REPOSITORY] = repository
|
||||||
|
addon_config[ATTR_LOCATON] = str(addon.parent)
|
||||||
|
self._addons_cache[addon_slug] = addon_config
|
||||||
|
|
||||||
|
except OSError:
|
||||||
_LOGGER.warning("Can't read %s", addon)
|
_LOGGER.warning("Can't read %s", addon)
|
||||||
|
|
||||||
except vol.Invalid as ex:
|
except vol.Invalid as ex:
|
||||||
_LOGGER.warning("Can't read %s -> %s", addon,
|
_LOGGER.warning("Can't read %s -> %s", addon,
|
||||||
humanize_error(addon_config, ex))
|
humanize_error(addon_config, ex))
|
||||||
|
|
||||||
|
def _set_builtin_repositories(self):
|
||||||
|
"""Add local built-in repository into dataset."""
|
||||||
|
try:
|
||||||
|
builtin_file = Path(__file__).parent.joinpath('built-in.json')
|
||||||
|
builtin_data = read_json_file(builtin_file)
|
||||||
|
except (OSError, json.JSONDecodeError) as err:
|
||||||
|
_LOGGER.warning("Can't read built-in.json -> %s", err)
|
||||||
|
return
|
||||||
|
|
||||||
|
# if core addons are available
|
||||||
|
for data in self._addons_cache.values():
|
||||||
|
if data[ATTR_REPOSITORY] == REPOSITORY_CORE:
|
||||||
|
self._repositories_data[REPOSITORY_CORE] = \
|
||||||
|
builtin_data[REPOSITORY_CORE]
|
||||||
|
break
|
||||||
|
|
||||||
|
# if local addons are available
|
||||||
|
for data in self._addons_cache.values():
|
||||||
|
if data[ATTR_REPOSITORY] == REPOSITORY_LOCAL:
|
||||||
|
self._repositories_data[REPOSITORY_LOCAL] = \
|
||||||
|
builtin_data[REPOSITORY_LOCAL]
|
||||||
|
break
|
||||||
|
|
||||||
|
def merge_update_config(self):
|
||||||
|
"""Update local config if they have update.
|
||||||
|
|
||||||
|
It need to be the same version as the local version is.
|
||||||
|
"""
|
||||||
|
have_change = False
|
||||||
|
|
||||||
|
for addon in self.list_installed:
|
||||||
|
# detached
|
||||||
|
if addon not in self._addons_cache:
|
||||||
|
continue
|
||||||
|
|
||||||
|
cache = self._addons_cache[addon]
|
||||||
|
data = self._system_data[addon]
|
||||||
|
if data[ATTR_VERSION] == cache[ATTR_VERSION]:
|
||||||
|
if data != cache:
|
||||||
|
self._system_data[addon] = copy.deepcopy(cache)
|
||||||
|
have_change = True
|
||||||
|
|
||||||
|
if have_change:
|
||||||
|
self.save()
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def list_installed(self):
|
def list_installed(self):
|
||||||
"""Return a list of installed addons."""
|
"""Return a list of installed addons."""
|
||||||
return set(self._system_data.keys())
|
return set(self._system_data)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def list_api(self):
|
def list_all(self):
|
||||||
"""Return a list of available addons for api."""
|
"""Return a dict of all addons."""
|
||||||
data = []
|
return set(self._system_data) | set(self._addons_cache)
|
||||||
all_addons = {**self._system_data, **self._current_data}
|
|
||||||
dedicated = self.list_removed
|
|
||||||
|
|
||||||
for addon, values in all_addons.items():
|
|
||||||
i_version = self._user_data.get(addon, {}).get(ATTR_VERSION)
|
|
||||||
|
|
||||||
data.append({
|
|
||||||
ATTR_NAME: values[ATTR_NAME],
|
|
||||||
ATTR_SLUG: values[ATTR_SLUG],
|
|
||||||
ATTR_DESCRIPTON: values[ATTR_DESCRIPTON],
|
|
||||||
ATTR_VERSION: values[ATTR_VERSION],
|
|
||||||
ATTR_INSTALLED: i_version,
|
|
||||||
ATTR_DEDICATED: addon in dedicated,
|
|
||||||
})
|
|
||||||
|
|
||||||
return data
|
|
||||||
|
|
||||||
def list_startup(self, start_type):
|
def list_startup(self, start_type):
|
||||||
"""Get list of installed addon with need start by type."""
|
"""Get list of installed addon with need start by type."""
|
||||||
@@ -109,18 +196,23 @@ class AddonsData(Config):
|
|||||||
return addon_list
|
return addon_list
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def list_removed(self):
|
def list_detached(self):
|
||||||
"""Return local addons they not support from repo."""
|
"""Return local addons they not support from repo."""
|
||||||
addon_list = set()
|
addon_list = set()
|
||||||
for addon in self._system_data.keys():
|
for addon in self._system_data.keys():
|
||||||
if addon not in self._current_data:
|
if addon not in self._addons_cache:
|
||||||
addon_list.add(addon)
|
addon_list.add(addon)
|
||||||
|
|
||||||
return addon_list
|
return addon_list
|
||||||
|
|
||||||
|
@property
|
||||||
|
def list_repositories(self):
|
||||||
|
"""Return list of addon repositories."""
|
||||||
|
return list(self._repositories_data.values())
|
||||||
|
|
||||||
def exists_addon(self, addon):
|
def exists_addon(self, addon):
|
||||||
"""Return True if a addon exists."""
|
"""Return True if a addon exists."""
|
||||||
return addon in self._current_data or addon in self._system_data
|
return addon in self._addons_cache or addon in self._system_data
|
||||||
|
|
||||||
def is_installed(self, addon):
|
def is_installed(self, addon):
|
||||||
"""Return True if a addon is installed."""
|
"""Return True if a addon is installed."""
|
||||||
@@ -128,11 +220,11 @@ class AddonsData(Config):
|
|||||||
|
|
||||||
def version_installed(self, addon):
|
def version_installed(self, addon):
|
||||||
"""Return installed version."""
|
"""Return installed version."""
|
||||||
return self._user_data[addon][ATTR_VERSION]
|
return self._user_data.get(addon, {}).get(ATTR_VERSION)
|
||||||
|
|
||||||
def set_addon_install(self, addon, version):
|
def set_addon_install(self, addon, version):
|
||||||
"""Set addon as installed."""
|
"""Set addon as installed."""
|
||||||
self._system_data[addon] = self._current_data[addon]
|
self._system_data[addon] = copy.deepcopy(self._addons_cache[addon])
|
||||||
self._user_data[addon] = {
|
self._user_data[addon] = {
|
||||||
ATTR_OPTIONS: {},
|
ATTR_OPTIONS: {},
|
||||||
ATTR_VERSION: version,
|
ATTR_VERSION: version,
|
||||||
@@ -147,13 +239,13 @@ class AddonsData(Config):
|
|||||||
|
|
||||||
def set_addon_update(self, addon, version):
|
def set_addon_update(self, addon, version):
|
||||||
"""Update version of addon."""
|
"""Update version of addon."""
|
||||||
self._system_data[addon] = self._current_data[addon]
|
self._system_data[addon] = copy.deepcopy(self._addons_cache[addon])
|
||||||
self._user_data[addon][ATTR_VERSION] = version
|
self._user_data[addon][ATTR_VERSION] = version
|
||||||
self.save()
|
self.save()
|
||||||
|
|
||||||
def set_options(self, addon, options):
|
def set_options(self, addon, options):
|
||||||
"""Store user addon options."""
|
"""Store user addon options."""
|
||||||
self._user_data[addon][ATTR_OPTIONS] = options
|
self._user_data[addon][ATTR_OPTIONS] = copy.deepcopy(options)
|
||||||
self.save()
|
self.save()
|
||||||
|
|
||||||
def set_boot(self, addon, boot):
|
def set_boot(self, addon, boot):
|
||||||
@@ -177,50 +269,111 @@ class AddonsData(Config):
|
|||||||
|
|
||||||
def get_name(self, addon):
|
def get_name(self, addon):
|
||||||
"""Return name of addon."""
|
"""Return name of addon."""
|
||||||
|
if addon in self._addons_cache:
|
||||||
|
return self._addons_cache[addon][ATTR_NAME]
|
||||||
return self._system_data[addon][ATTR_NAME]
|
return self._system_data[addon][ATTR_NAME]
|
||||||
|
|
||||||
def get_description(self, addon):
|
def get_description(self, addon):
|
||||||
"""Return description of addon."""
|
"""Return description of addon."""
|
||||||
|
if addon in self._addons_cache:
|
||||||
|
return self._addons_cache[addon][ATTR_DESCRIPTON]
|
||||||
return self._system_data[addon][ATTR_DESCRIPTON]
|
return self._system_data[addon][ATTR_DESCRIPTON]
|
||||||
|
|
||||||
|
def get_repository(self, addon):
|
||||||
|
"""Return repository of addon."""
|
||||||
|
if addon in self._addons_cache:
|
||||||
|
return self._addons_cache[addon][ATTR_REPOSITORY]
|
||||||
|
return self._system_data[addon][ATTR_REPOSITORY]
|
||||||
|
|
||||||
def get_last_version(self, addon):
|
def get_last_version(self, addon):
|
||||||
"""Return version of addon."""
|
"""Return version of addon."""
|
||||||
if addon not in self._current_data:
|
if addon in self._addons_cache:
|
||||||
return self.version_installed(addon)
|
return self._addons_cache[addon][ATTR_VERSION]
|
||||||
return self._current_data[addon][ATTR_VERSION]
|
return self.version_installed(addon)
|
||||||
|
|
||||||
def get_ports(self, addon):
|
def get_ports(self, addon):
|
||||||
"""Return ports of addon."""
|
"""Return ports of addon."""
|
||||||
return self._system_data[addon].get(ATTR_PORTS)
|
return self._system_data[addon].get(ATTR_PORTS)
|
||||||
|
|
||||||
|
def get_network_mode(self, addon):
|
||||||
|
"""Return network mode of addon."""
|
||||||
|
if self._system_data[addon][ATTR_HOST_NETWORK]:
|
||||||
|
return 'host'
|
||||||
|
return 'bridge'
|
||||||
|
|
||||||
|
def get_devices(self, addon):
|
||||||
|
"""Return devices of addon."""
|
||||||
|
return self._system_data[addon].get(ATTR_DEVICES)
|
||||||
|
|
||||||
|
def get_tmpfs(self, addon):
|
||||||
|
"""Return tmpfs of addon."""
|
||||||
|
return self._system_data[addon].get(ATTR_TMPFS)
|
||||||
|
|
||||||
|
def get_environment(self, addon):
|
||||||
|
"""Return environment of addon."""
|
||||||
|
return self._system_data[addon].get(ATTR_ENVIRONMENT)
|
||||||
|
|
||||||
|
def get_privileged(self, addon):
|
||||||
|
"""Return list of privilege."""
|
||||||
|
return self._system_data[addon].get(ATTR_PRIVILEGED)
|
||||||
|
|
||||||
|
def get_url(self, addon):
|
||||||
|
"""Return url of addon."""
|
||||||
|
if addon in self._addons_cache:
|
||||||
|
return self._addons_cache[addon].get(ATTR_URL)
|
||||||
|
return self._system_data[addon].get(ATTR_URL)
|
||||||
|
|
||||||
|
def get_arch(self, addon):
|
||||||
|
"""Return list of supported arch."""
|
||||||
|
if addon in self._addons_cache:
|
||||||
|
return self._addons_cache[addon][ATTR_ARCH]
|
||||||
|
return self._system_data[addon][ATTR_ARCH]
|
||||||
|
|
||||||
def get_image(self, addon):
|
def get_image(self, addon):
|
||||||
"""Return image name of addon."""
|
"""Return image name of addon."""
|
||||||
addon_data = self._system_data.get(addon, self._current_data[addon])
|
addon_data = self._system_data.get(
|
||||||
|
addon, self._addons_cache.get(addon)
|
||||||
|
)
|
||||||
|
|
||||||
if ATTR_IMAGE not in addon_data:
|
# Repository with dockerhub images
|
||||||
return "{}/{}-addon-{}".format(DOCKER_REPO, self.arch, addon)
|
if ATTR_IMAGE in addon_data:
|
||||||
|
return addon_data[ATTR_IMAGE].format(arch=self.arch)
|
||||||
|
|
||||||
return addon_data[ATTR_IMAGE]
|
# local build
|
||||||
|
return "{}/{}-addon-{}".format(
|
||||||
|
addon_data[ATTR_REPOSITORY], self.arch, addon_data[ATTR_SLUG])
|
||||||
|
|
||||||
def need_config(self, addon):
|
def need_build(self, addon):
|
||||||
"""Return True if config map is needed."""
|
"""Return True if this addon need a local build."""
|
||||||
return self._system_data[addon][ATTR_MAP_CONFIG]
|
addon_data = self._system_data.get(
|
||||||
|
addon, self._addons_cache.get(addon)
|
||||||
|
)
|
||||||
|
return ATTR_IMAGE not in addon_data
|
||||||
|
|
||||||
def need_ssl(self, addon):
|
def map_volumes(self, addon):
|
||||||
"""Return True if ssl map is needed."""
|
"""Return a dict of {volume: policy} from addon."""
|
||||||
return self._system_data[addon][ATTR_MAP_SSL]
|
volumes = {}
|
||||||
|
for volume in self._system_data[addon][ATTR_MAP]:
|
||||||
|
result = RE_VOLUME.match(volume)
|
||||||
|
volumes[result.group(1)] = result.group(2) or 'ro'
|
||||||
|
|
||||||
|
return volumes
|
||||||
|
|
||||||
def path_data(self, addon):
|
def path_data(self, addon):
|
||||||
"""Return addon data path inside supervisor."""
|
"""Return addon data path inside supervisor."""
|
||||||
return "{}/{}".format(self.config.path_addons_data, addon)
|
return Path(self.config.path_addons_data, addon)
|
||||||
|
|
||||||
def path_data_docker(self, addon):
|
def path_extern_data(self, addon):
|
||||||
"""Return addon data path external for docker."""
|
"""Return addon data path external for docker."""
|
||||||
return "{}/{}".format(self.config.path_addons_data_docker, addon)
|
return PurePath(self.config.path_extern_addons_data, addon)
|
||||||
|
|
||||||
def path_addon_options(self, addon):
|
def path_addon_options(self, addon):
|
||||||
"""Return path to addons options."""
|
"""Return path to addons options."""
|
||||||
return "{}/options.json".format(self.path_data(addon))
|
return Path(self.path_data(addon), "options.json")
|
||||||
|
|
||||||
|
def path_addon_location(self, addon):
|
||||||
|
"""Return path to this addon."""
|
||||||
|
return Path(self._addons_cache[addon][ATTR_LOCATON])
|
||||||
|
|
||||||
def write_addon_options(self, addon):
|
def write_addon_options(self, addon):
|
||||||
"""Return True if addon options is written to data."""
|
"""Return True if addon options is written to data."""
|
||||||
@@ -240,5 +393,7 @@ class AddonsData(Config):
|
|||||||
"""Create a schema for addon options."""
|
"""Create a schema for addon options."""
|
||||||
raw_schema = self._system_data[addon][ATTR_SCHEMA]
|
raw_schema = self._system_data[addon][ATTR_SCHEMA]
|
||||||
|
|
||||||
schema = vol.Schema(vol.All(dict, validate_options(raw_schema)))
|
if isinstance(raw_schema, bool):
|
||||||
return schema
|
return vol.Schema(dict)
|
||||||
|
|
||||||
|
return vol.Schema(vol.All(dict, validate_options(raw_schema)))
|
||||||
|
@@ -1,10 +1,12 @@
|
|||||||
"""Init file for HassIO addons git."""
|
"""Init file for HassIO addons git."""
|
||||||
import asyncio
|
import asyncio
|
||||||
import logging
|
import logging
|
||||||
import os
|
from pathlib import Path
|
||||||
|
import shutil
|
||||||
|
|
||||||
import git
|
import git
|
||||||
|
|
||||||
|
from .util import get_hash_from_repository
|
||||||
from ..const import URL_HASSIO_ADDONS
|
from ..const import URL_HASSIO_ADDONS
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
@@ -13,26 +15,29 @@ _LOGGER = logging.getLogger(__name__)
|
|||||||
class AddonsRepo(object):
|
class AddonsRepo(object):
|
||||||
"""Manage addons git repo."""
|
"""Manage addons git repo."""
|
||||||
|
|
||||||
def __init__(self, config, loop):
|
def __init__(self, config, loop, path, url):
|
||||||
"""Initialize docker base wrapper."""
|
"""Initialize git base wrapper."""
|
||||||
self.config = config
|
self.config = config
|
||||||
self.loop = loop
|
self.loop = loop
|
||||||
self.repo = None
|
self.repo = None
|
||||||
|
self.path = path
|
||||||
|
self.url = url
|
||||||
self._lock = asyncio.Lock(loop=loop)
|
self._lock = asyncio.Lock(loop=loop)
|
||||||
|
|
||||||
async def load(self):
|
async def load(self):
|
||||||
"""Init git addon repo."""
|
"""Init git addon repo."""
|
||||||
if not os.path.isdir(self.config.path_addons_repo):
|
if not self.path.is_dir():
|
||||||
return await self.clone()
|
return await self.clone()
|
||||||
|
|
||||||
async with self._lock:
|
async with self._lock:
|
||||||
try:
|
try:
|
||||||
_LOGGER.info("Load addons repository")
|
_LOGGER.info("Load addon %s repository", self.path)
|
||||||
self.repo = await self.loop.run_in_executor(
|
self.repo = await self.loop.run_in_executor(
|
||||||
None, git.Repo, self.config.path_addons_repo)
|
None, git.Repo, str(self.path))
|
||||||
|
|
||||||
except (git.InvalidGitRepositoryError, git.NoSuchPathError) as err:
|
except (git.InvalidGitRepositoryError, git.NoSuchPathError,
|
||||||
_LOGGER.error("Can't load addons repo: %s.", err)
|
git.GitCommandError) as err:
|
||||||
|
_LOGGER.error("Can't load %s repo: %s.", self.path, err)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
return True
|
return True
|
||||||
@@ -41,13 +46,13 @@ class AddonsRepo(object):
|
|||||||
"""Clone git addon repo."""
|
"""Clone git addon repo."""
|
||||||
async with self._lock:
|
async with self._lock:
|
||||||
try:
|
try:
|
||||||
_LOGGER.info("Clone addons repository")
|
_LOGGER.info("Clone addon %s repository", self.url)
|
||||||
self.repo = await self.loop.run_in_executor(
|
self.repo = await self.loop.run_in_executor(
|
||||||
None, git.Repo.clone_from, URL_HASSIO_ADDONS,
|
None, git.Repo.clone_from, self.url, str(self.path))
|
||||||
self.config.path_addons_repo)
|
|
||||||
|
|
||||||
except (git.InvalidGitRepositoryError, git.NoSuchPathError) as err:
|
except (git.InvalidGitRepositoryError, git.NoSuchPathError,
|
||||||
_LOGGER.error("Can't clone addons repo: %s.", err)
|
git.GitCommandError) as err:
|
||||||
|
_LOGGER.error("Can't clone %s repo: %s.", self.url, err)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
return True
|
return True
|
||||||
@@ -60,12 +65,43 @@ class AddonsRepo(object):
|
|||||||
|
|
||||||
async with self._lock:
|
async with self._lock:
|
||||||
try:
|
try:
|
||||||
_LOGGER.info("Pull addons repository")
|
_LOGGER.info("Pull addon %s repository", self.url)
|
||||||
await self.loop.run_in_executor(
|
await self.loop.run_in_executor(
|
||||||
None, self.repo.remotes.origin.pull)
|
None, self.repo.remotes.origin.pull)
|
||||||
|
|
||||||
except (git.InvalidGitRepositoryError, git.NoSuchPathError) as err:
|
except (git.InvalidGitRepositoryError, git.NoSuchPathError,
|
||||||
_LOGGER.error("Can't pull addons repo: %s.", err)
|
git.exc.GitCommandError) as err:
|
||||||
|
_LOGGER.error("Can't pull %s repo: %s.", self.url, err)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
class AddonsRepoHassIO(AddonsRepo):
|
||||||
|
"""HassIO addons repository."""
|
||||||
|
|
||||||
|
def __init__(self, config, loop):
|
||||||
|
"""Initialize git hassio addon repository."""
|
||||||
|
super().__init__(
|
||||||
|
config, loop, config.path_addons_core, URL_HASSIO_ADDONS)
|
||||||
|
|
||||||
|
|
||||||
|
class AddonsRepoCustom(AddonsRepo):
|
||||||
|
"""Custom addons repository."""
|
||||||
|
|
||||||
|
def __init__(self, config, loop, url):
|
||||||
|
"""Initialize git hassio addon repository."""
|
||||||
|
path = Path(config.path_addons_git, get_hash_from_repository(url))
|
||||||
|
|
||||||
|
super().__init__(config, loop, path, url)
|
||||||
|
|
||||||
|
def remove(self):
|
||||||
|
"""Remove a custom addon."""
|
||||||
|
if self.path.is_dir():
|
||||||
|
_LOGGER.info("Remove custom addon repository %s", self.url)
|
||||||
|
|
||||||
|
def log_err(funct, path, _):
|
||||||
|
"""Log error."""
|
||||||
|
_LOGGER.warning("Can't remove %s", path)
|
||||||
|
|
||||||
|
shutil.rmtree(str(self.path), onerror=log_err)
|
||||||
|
26
hassio/addons/util.py
Normal file
26
hassio/addons/util.py
Normal file
@@ -0,0 +1,26 @@
|
|||||||
|
"""Util addons functions."""
|
||||||
|
import hashlib
|
||||||
|
import re
|
||||||
|
|
||||||
|
RE_SLUGIFY = re.compile(r'[^a-z0-9_]+')
|
||||||
|
RE_SHA1 = re.compile(r"[a-f0-9]{8}")
|
||||||
|
|
||||||
|
|
||||||
|
def get_hash_from_repository(name):
|
||||||
|
"""Generate a hash from repository."""
|
||||||
|
key = name.lower().encode()
|
||||||
|
return hashlib.sha1(key).hexdigest()[:8]
|
||||||
|
|
||||||
|
|
||||||
|
def extract_hash_from_path(path):
|
||||||
|
"""Extract repo id from path."""
|
||||||
|
repo_dir = path.parts[-1]
|
||||||
|
|
||||||
|
if not RE_SHA1.match(repo_dir):
|
||||||
|
return get_hash_from_repository(repo_dir)
|
||||||
|
return repo_dir
|
||||||
|
|
||||||
|
|
||||||
|
def create_hash_index_list(name_list):
|
||||||
|
"""Create a dict with hash from repositories list."""
|
||||||
|
return {get_hash_from_repository(repo): repo for repo in name_list}
|
@@ -3,9 +3,14 @@ import voluptuous as vol
|
|||||||
|
|
||||||
from ..const import (
|
from ..const import (
|
||||||
ATTR_NAME, ATTR_VERSION, ATTR_SLUG, ATTR_DESCRIPTON, ATTR_STARTUP,
|
ATTR_NAME, ATTR_VERSION, ATTR_SLUG, ATTR_DESCRIPTON, ATTR_STARTUP,
|
||||||
ATTR_BOOT, ATTR_MAP_SSL, ATTR_MAP_CONFIG, ATTR_OPTIONS,
|
ATTR_BOOT, ATTR_MAP, ATTR_OPTIONS, ATTR_PORTS, STARTUP_ONCE, STARTUP_AFTER,
|
||||||
ATTR_PORTS, STARTUP_ONCE, STARTUP_AFTER, STARTUP_BEFORE, BOOT_AUTO,
|
STARTUP_BEFORE, STARTUP_INITIALIZE, BOOT_AUTO, BOOT_MANUAL, ATTR_SCHEMA,
|
||||||
BOOT_MANUAL, ATTR_SCHEMA, ATTR_IMAGE)
|
ATTR_IMAGE, ATTR_URL, ATTR_MAINTAINER, ATTR_ARCH, ATTR_DEVICES,
|
||||||
|
ATTR_ENVIRONMENT, ATTR_HOST_NETWORK, ARCH_ARMHF, ARCH_AARCH64, ARCH_AMD64,
|
||||||
|
ARCH_I386, ATTR_TMPFS, ATTR_PRIVILEGED)
|
||||||
|
|
||||||
|
|
||||||
|
MAP_VOLUME = r"^(config|ssl|addons|backup|share)(?::(rw|:ro))?$"
|
||||||
|
|
||||||
V_STR = 'str'
|
V_STR = 'str'
|
||||||
V_INT = 'int'
|
V_INT = 'int'
|
||||||
@@ -16,27 +21,62 @@ V_URL = 'url'
|
|||||||
|
|
||||||
ADDON_ELEMENT = vol.In([V_STR, V_INT, V_FLOAT, V_BOOL, V_EMAIL, V_URL])
|
ADDON_ELEMENT = vol.In([V_STR, V_INT, V_FLOAT, V_BOOL, V_EMAIL, V_URL])
|
||||||
|
|
||||||
|
ARCH_ALL = [
|
||||||
|
ARCH_ARMHF, ARCH_AARCH64, ARCH_AMD64, ARCH_I386
|
||||||
|
]
|
||||||
|
|
||||||
|
PRIVILEGE_ALL = [
|
||||||
|
"NET_ADMIN"
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def check_network(data):
|
||||||
|
"""Validate network settings."""
|
||||||
|
host_network = data[ATTR_HOST_NETWORK]
|
||||||
|
|
||||||
|
if ATTR_PORTS in data and host_network:
|
||||||
|
raise vol.Invalid("Hostnetwork & ports are not allow!")
|
||||||
|
|
||||||
|
return data
|
||||||
|
|
||||||
|
|
||||||
# pylint: disable=no-value-for-parameter
|
# pylint: disable=no-value-for-parameter
|
||||||
SCHEMA_ADDON_CONFIG = vol.Schema({
|
SCHEMA_ADDON_CONFIG = vol.Schema(vol.All({
|
||||||
vol.Required(ATTR_NAME): vol.Coerce(str),
|
vol.Required(ATTR_NAME): vol.Coerce(str),
|
||||||
vol.Required(ATTR_VERSION): vol.Coerce(str),
|
vol.Required(ATTR_VERSION): vol.Coerce(str),
|
||||||
vol.Required(ATTR_SLUG): vol.Coerce(str),
|
vol.Required(ATTR_SLUG): vol.Coerce(str),
|
||||||
vol.Required(ATTR_DESCRIPTON): vol.Coerce(str),
|
vol.Required(ATTR_DESCRIPTON): vol.Coerce(str),
|
||||||
|
vol.Optional(ATTR_URL): vol.Url(),
|
||||||
|
vol.Optional(ATTR_ARCH, default=ARCH_ALL): [vol.In(ARCH_ALL)],
|
||||||
vol.Required(ATTR_STARTUP):
|
vol.Required(ATTR_STARTUP):
|
||||||
vol.In([STARTUP_BEFORE, STARTUP_AFTER, STARTUP_ONCE]),
|
vol.In([STARTUP_BEFORE, STARTUP_AFTER, STARTUP_ONCE,
|
||||||
|
STARTUP_INITIALIZE]),
|
||||||
vol.Required(ATTR_BOOT):
|
vol.Required(ATTR_BOOT):
|
||||||
vol.In([BOOT_AUTO, BOOT_MANUAL]),
|
vol.In([BOOT_AUTO, BOOT_MANUAL]),
|
||||||
vol.Optional(ATTR_PORTS): dict,
|
vol.Optional(ATTR_PORTS): dict,
|
||||||
vol.Optional(ATTR_MAP_CONFIG, default=False): vol.Boolean(),
|
vol.Optional(ATTR_HOST_NETWORK, default=False): vol.Boolean(),
|
||||||
vol.Optional(ATTR_MAP_SSL, default=False): vol.Boolean(),
|
vol.Optional(ATTR_DEVICES): [vol.Match(r"^(.*):(.*):([rwm]{1,3})$")],
|
||||||
|
vol.Optional(ATTR_TMPFS):
|
||||||
|
vol.Match(r"^size=(\d)*[kmg](,uid=\d{1,4})?(,rw)?$"),
|
||||||
|
vol.Optional(ATTR_MAP, default=[]): [vol.Match(MAP_VOLUME)],
|
||||||
|
vol.Optional(ATTR_ENVIRONMENT): {vol.Match(r"\w*"): vol.Coerce(str)},
|
||||||
|
vol.Optional(ATTR_PRIVILEGED): [vol.In(PRIVILEGE_ALL)],
|
||||||
vol.Required(ATTR_OPTIONS): dict,
|
vol.Required(ATTR_OPTIONS): dict,
|
||||||
vol.Required(ATTR_SCHEMA): {
|
vol.Required(ATTR_SCHEMA): vol.Any({
|
||||||
vol.Coerce(str): vol.Any(ADDON_ELEMENT, [
|
vol.Coerce(str): vol.Any(ADDON_ELEMENT, [
|
||||||
vol.Any(ADDON_ELEMENT, {vol.Coerce(str): ADDON_ELEMENT})
|
vol.Any(ADDON_ELEMENT, {vol.Coerce(str): ADDON_ELEMENT})
|
||||||
])
|
])
|
||||||
},
|
}, False),
|
||||||
vol.Optional(ATTR_IMAGE): vol.Match(r"\w*/\w*"),
|
vol.Optional(ATTR_IMAGE): vol.Match(r"\w*/\w*"),
|
||||||
})
|
}, check_network), extra=vol.ALLOW_EXTRA)
|
||||||
|
|
||||||
|
|
||||||
|
# pylint: disable=no-value-for-parameter
|
||||||
|
SCHEMA_REPOSITORY_CONFIG = vol.Schema({
|
||||||
|
vol.Required(ATTR_NAME): vol.Coerce(str),
|
||||||
|
vol.Optional(ATTR_URL): vol.Url(),
|
||||||
|
vol.Optional(ATTR_MAINTAINER): vol.Coerce(str),
|
||||||
|
}, extra=vol.ALLOW_EXTRA)
|
||||||
|
|
||||||
|
|
||||||
def validate_options(raw_schema):
|
def validate_options(raw_schema):
|
||||||
@@ -54,10 +94,10 @@ def validate_options(raw_schema):
|
|||||||
try:
|
try:
|
||||||
if isinstance(typ, list):
|
if isinstance(typ, list):
|
||||||
# nested value
|
# nested value
|
||||||
options[key] = _nested_validate(typ[0], value)
|
options[key] = _nested_validate(typ[0], value, key)
|
||||||
else:
|
else:
|
||||||
# normal value
|
# normal value
|
||||||
options[key] = _single_validate(typ, value)
|
options[key] = _single_validate(typ, value, key)
|
||||||
except (IndexError, KeyError):
|
except (IndexError, KeyError):
|
||||||
raise vol.Invalid(
|
raise vol.Invalid(
|
||||||
"Type error for {}.".format(key)) from None
|
"Type error for {}.".format(key)) from None
|
||||||
@@ -68,9 +108,13 @@ def validate_options(raw_schema):
|
|||||||
|
|
||||||
|
|
||||||
# pylint: disable=no-value-for-parameter
|
# pylint: disable=no-value-for-parameter
|
||||||
def _single_validate(typ, value):
|
def _single_validate(typ, value, key):
|
||||||
"""Validate a single element."""
|
"""Validate a single element."""
|
||||||
try:
|
try:
|
||||||
|
# if required argument
|
||||||
|
if value is None:
|
||||||
|
raise vol.Invalid("Missing required option '{}'.".format(key))
|
||||||
|
|
||||||
if typ == V_STR:
|
if typ == V_STR:
|
||||||
return str(value)
|
return str(value)
|
||||||
elif typ == V_INT:
|
elif typ == V_INT:
|
||||||
@@ -84,13 +128,13 @@ def _single_validate(typ, value):
|
|||||||
elif typ == V_URL:
|
elif typ == V_URL:
|
||||||
return vol.Url()(value)
|
return vol.Url()(value)
|
||||||
|
|
||||||
raise vol.Invalid("Fatal error for {}.".format(value))
|
raise vol.Invalid("Fatal error for {} type {}.".format(key, typ))
|
||||||
except TypeError:
|
except ValueError:
|
||||||
raise vol.Invalid(
|
raise vol.Invalid(
|
||||||
"Type {} error for {}.".format(typ, value)) from None
|
"Type {} error for '{}' on {}.".format(typ, value, key)) from None
|
||||||
|
|
||||||
|
|
||||||
def _nested_validate(typ, data_list):
|
def _nested_validate(typ, data_list, key):
|
||||||
"""Validate nested items."""
|
"""Validate nested items."""
|
||||||
options = []
|
options = []
|
||||||
|
|
||||||
@@ -103,10 +147,10 @@ def _nested_validate(typ, data_list):
|
|||||||
raise vol.Invalid(
|
raise vol.Invalid(
|
||||||
"Unknown nested options {}.".format(c_key))
|
"Unknown nested options {}.".format(c_key))
|
||||||
|
|
||||||
c_options[c_key] = _single_validate(typ[c_key], c_value)
|
c_options[c_key] = _single_validate(typ[c_key], c_value, c_key)
|
||||||
options.append(c_options)
|
options.append(c_options)
|
||||||
# normal list
|
# normal list
|
||||||
else:
|
else:
|
||||||
options.append(_single_validate(typ, element))
|
options.append(_single_validate(typ, element, key))
|
||||||
|
|
||||||
return options
|
return options
|
||||||
|
@@ -1,5 +1,6 @@
|
|||||||
"""Init file for HassIO rest api."""
|
"""Init file for HassIO rest api."""
|
||||||
import logging
|
import logging
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
from aiohttp import web
|
from aiohttp import web
|
||||||
|
|
||||||
@@ -8,6 +9,7 @@ from .homeassistant import APIHomeAssistant
|
|||||||
from .host import APIHost
|
from .host import APIHost
|
||||||
from .network import APINetwork
|
from .network import APINetwork
|
||||||
from .supervisor import APISupervisor
|
from .supervisor import APISupervisor
|
||||||
|
from .security import APISecurity
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
@@ -30,16 +32,16 @@ class RestAPI(object):
|
|||||||
api_host = APIHost(self.config, self.loop, host_control)
|
api_host = APIHost(self.config, self.loop, host_control)
|
||||||
|
|
||||||
self.webapp.router.add_get('/host/info', api_host.info)
|
self.webapp.router.add_get('/host/info', api_host.info)
|
||||||
self.webapp.router.add_get('/host/reboot', api_host.reboot)
|
self.webapp.router.add_post('/host/reboot', api_host.reboot)
|
||||||
self.webapp.router.add_get('/host/shutdown', api_host.shutdown)
|
self.webapp.router.add_post('/host/shutdown', api_host.shutdown)
|
||||||
self.webapp.router.add_get('/host/update', api_host.update)
|
self.webapp.router.add_post('/host/update', api_host.update)
|
||||||
|
|
||||||
def register_network(self, host_control):
|
def register_network(self, host_control):
|
||||||
"""Register network function."""
|
"""Register network function."""
|
||||||
api_net = APINetwork(self.config, self.loop, host_control)
|
api_net = APINetwork(self.config, self.loop, host_control)
|
||||||
|
|
||||||
self.webapp.router.add_get('/network/info', api_net.info)
|
self.webapp.router.add_get('/network/info', api_net.info)
|
||||||
self.webapp.router.add_get('/network/options', api_net.options)
|
self.webapp.router.add_post('/network/options', api_net.options)
|
||||||
|
|
||||||
def register_supervisor(self, supervisor, addons, host_control):
|
def register_supervisor(self, supervisor, addons, host_control):
|
||||||
"""Register supervisor function."""
|
"""Register supervisor function."""
|
||||||
@@ -48,9 +50,13 @@ class RestAPI(object):
|
|||||||
|
|
||||||
self.webapp.router.add_get('/supervisor/ping', api_supervisor.ping)
|
self.webapp.router.add_get('/supervisor/ping', api_supervisor.ping)
|
||||||
self.webapp.router.add_get('/supervisor/info', api_supervisor.info)
|
self.webapp.router.add_get('/supervisor/info', api_supervisor.info)
|
||||||
self.webapp.router.add_get('/supervisor/update', api_supervisor.update)
|
|
||||||
self.webapp.router.add_get('/supervisor/reload', api_supervisor.reload)
|
|
||||||
self.webapp.router.add_get(
|
self.webapp.router.add_get(
|
||||||
|
'/supervisor/addons', api_supervisor.available_addons)
|
||||||
|
self.webapp.router.add_post(
|
||||||
|
'/supervisor/update', api_supervisor.update)
|
||||||
|
self.webapp.router.add_post(
|
||||||
|
'/supervisor/reload', api_supervisor.reload)
|
||||||
|
self.webapp.router.add_post(
|
||||||
'/supervisor/options', api_supervisor.options)
|
'/supervisor/options', api_supervisor.options)
|
||||||
self.webapp.router.add_get('/supervisor/logs', api_supervisor.logs)
|
self.webapp.router.add_get('/supervisor/logs', api_supervisor.logs)
|
||||||
|
|
||||||
@@ -59,7 +65,9 @@ class RestAPI(object):
|
|||||||
api_hass = APIHomeAssistant(self.config, self.loop, dock_homeassistant)
|
api_hass = APIHomeAssistant(self.config, self.loop, dock_homeassistant)
|
||||||
|
|
||||||
self.webapp.router.add_get('/homeassistant/info', api_hass.info)
|
self.webapp.router.add_get('/homeassistant/info', api_hass.info)
|
||||||
self.webapp.router.add_get('/homeassistant/update', api_hass.update)
|
self.webapp.router.add_post('/homeassistant/options', api_hass.options)
|
||||||
|
self.webapp.router.add_post('/homeassistant/update', api_hass.update)
|
||||||
|
self.webapp.router.add_post('/homeassistant/restart', api_hass.restart)
|
||||||
self.webapp.router.add_get('/homeassistant/logs', api_hass.logs)
|
self.webapp.router.add_get('/homeassistant/logs', api_hass.logs)
|
||||||
|
|
||||||
def register_addons(self, addons):
|
def register_addons(self, addons):
|
||||||
@@ -67,17 +75,39 @@ class RestAPI(object):
|
|||||||
api_addons = APIAddons(self.config, self.loop, addons)
|
api_addons = APIAddons(self.config, self.loop, addons)
|
||||||
|
|
||||||
self.webapp.router.add_get('/addons/{addon}/info', api_addons.info)
|
self.webapp.router.add_get('/addons/{addon}/info', api_addons.info)
|
||||||
self.webapp.router.add_get(
|
self.webapp.router.add_post(
|
||||||
'/addons/{addon}/install', api_addons.install)
|
'/addons/{addon}/install', api_addons.install)
|
||||||
self.webapp.router.add_get(
|
self.webapp.router.add_post(
|
||||||
'/addons/{addon}/uninstall', api_addons.uninstall)
|
'/addons/{addon}/uninstall', api_addons.uninstall)
|
||||||
self.webapp.router.add_get('/addons/{addon}/start', api_addons.start)
|
self.webapp.router.add_post('/addons/{addon}/start', api_addons.start)
|
||||||
self.webapp.router.add_get('/addons/{addon}/stop', api_addons.stop)
|
self.webapp.router.add_post('/addons/{addon}/stop', api_addons.stop)
|
||||||
self.webapp.router.add_get('/addons/{addon}/update', api_addons.update)
|
self.webapp.router.add_post(
|
||||||
self.webapp.router.add_get(
|
'/addons/{addon}/restart', api_addons.restart)
|
||||||
|
self.webapp.router.add_post(
|
||||||
|
'/addons/{addon}/update', api_addons.update)
|
||||||
|
self.webapp.router.add_post(
|
||||||
'/addons/{addon}/options', api_addons.options)
|
'/addons/{addon}/options', api_addons.options)
|
||||||
self.webapp.router.add_get('/addons/{addon}/logs', api_addons.logs)
|
self.webapp.router.add_get('/addons/{addon}/logs', api_addons.logs)
|
||||||
|
|
||||||
|
def register_security(self):
|
||||||
|
"""Register security function."""
|
||||||
|
api_security = APISecurity(self.config, self.loop)
|
||||||
|
|
||||||
|
self.webapp.router.add_get('/security/info', api_security.info)
|
||||||
|
self.webapp.router.add_post('/security/options', api_security.options)
|
||||||
|
self.webapp.router.add_post('/security/totp', api_security.totp)
|
||||||
|
self.webapp.router.add_post('/security/session', api_security.session)
|
||||||
|
|
||||||
|
def register_panel(self):
|
||||||
|
"""Register panel for homeassistant."""
|
||||||
|
panel = Path(__file__).parents[1].joinpath('panel/hassio-main.html')
|
||||||
|
|
||||||
|
def get_panel(request):
|
||||||
|
"""Return file response with panel."""
|
||||||
|
return web.FileResponse(panel)
|
||||||
|
|
||||||
|
self.webapp.router.add_get('/panel', get_panel)
|
||||||
|
|
||||||
async def start(self):
|
async def start(self):
|
||||||
"""Run rest api webserver."""
|
"""Run rest api webserver."""
|
||||||
self._handler = self.webapp.make_handler(loop=self.loop)
|
self._handler = self.webapp.make_handler(loop=self.loop)
|
||||||
|
@@ -8,7 +8,8 @@ from voluptuous.humanize import humanize_error
|
|||||||
from .util import api_process, api_process_raw, api_validate
|
from .util import api_process, api_process_raw, api_validate
|
||||||
from ..const import (
|
from ..const import (
|
||||||
ATTR_VERSION, ATTR_LAST_VERSION, ATTR_STATE, ATTR_BOOT, ATTR_OPTIONS,
|
ATTR_VERSION, ATTR_LAST_VERSION, ATTR_STATE, ATTR_BOOT, ATTR_OPTIONS,
|
||||||
STATE_STOPPED, STATE_STARTED, BOOT_AUTO, BOOT_MANUAL)
|
ATTR_URL, ATTR_DESCRIPTON, ATTR_DETACHED, ATTR_NAME, ATTR_REPOSITORY,
|
||||||
|
ATTR_BUILD, STATE_STOPPED, STATE_STARTED, BOOT_AUTO, BOOT_MANUAL)
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
@@ -48,11 +49,17 @@ class APIAddons(object):
|
|||||||
addon = self._extract_addon(request)
|
addon = self._extract_addon(request)
|
||||||
|
|
||||||
return {
|
return {
|
||||||
|
ATTR_NAME: self.addons.get_name(addon),
|
||||||
|
ATTR_DESCRIPTON: self.addons.get_description(addon),
|
||||||
ATTR_VERSION: self.addons.version_installed(addon),
|
ATTR_VERSION: self.addons.version_installed(addon),
|
||||||
|
ATTR_REPOSITORY: self.addons.get_repository(addon),
|
||||||
ATTR_LAST_VERSION: self.addons.get_last_version(addon),
|
ATTR_LAST_VERSION: self.addons.get_last_version(addon),
|
||||||
ATTR_STATE: await self.addons.state(addon),
|
ATTR_STATE: await self.addons.state(addon),
|
||||||
ATTR_BOOT: self.addons.get_boot(addon),
|
ATTR_BOOT: self.addons.get_boot(addon),
|
||||||
ATTR_OPTIONS: self.addons.get_options(addon),
|
ATTR_OPTIONS: self.addons.get_options(addon),
|
||||||
|
ATTR_URL: self.addons.get_url(addon),
|
||||||
|
ATTR_DETACHED: addon in self.addons.list_detached,
|
||||||
|
ATTR_BUILD: self.addons.need_build(addon),
|
||||||
}
|
}
|
||||||
|
|
||||||
@api_process
|
@api_process
|
||||||
@@ -82,6 +89,11 @@ class APIAddons(object):
|
|||||||
version = body.get(
|
version = body.get(
|
||||||
ATTR_VERSION, self.addons.get_last_version(addon))
|
ATTR_VERSION, self.addons.get_last_version(addon))
|
||||||
|
|
||||||
|
# check if arch supported
|
||||||
|
if self.addons.arch not in self.addons.get_arch(addon):
|
||||||
|
raise RuntimeError(
|
||||||
|
"Addon is not supported on {}".format(self.addons.arch))
|
||||||
|
|
||||||
return await asyncio.shield(
|
return await asyncio.shield(
|
||||||
self.addons.install(addon, version), loop=self.loop)
|
self.addons.install(addon, version), loop=self.loop)
|
||||||
|
|
||||||
@@ -137,6 +149,12 @@ class APIAddons(object):
|
|||||||
return await asyncio.shield(
|
return await asyncio.shield(
|
||||||
self.addons.update(addon, version), loop=self.loop)
|
self.addons.update(addon, version), loop=self.loop)
|
||||||
|
|
||||||
|
@api_process
|
||||||
|
async def restart(self, request):
|
||||||
|
"""Restart addon."""
|
||||||
|
addon = self._extract_addon(request)
|
||||||
|
return await asyncio.shield(self.addons.restart(addon), loop=self.loop)
|
||||||
|
|
||||||
@api_process_raw
|
@api_process_raw
|
||||||
def logs(self, request):
|
def logs(self, request):
|
||||||
"""Return logs from addon."""
|
"""Return logs from addon."""
|
||||||
|
@@ -5,10 +5,15 @@ import logging
|
|||||||
import voluptuous as vol
|
import voluptuous as vol
|
||||||
|
|
||||||
from .util import api_process, api_process_raw, api_validate
|
from .util import api_process, api_process_raw, api_validate
|
||||||
from ..const import ATTR_VERSION, ATTR_LAST_VERSION
|
from ..const import ATTR_VERSION, ATTR_LAST_VERSION, ATTR_DEVICES
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
SCHEMA_OPTIONS = vol.Schema({
|
||||||
|
vol.Optional(ATTR_DEVICES): [vol.Match(r"^[^/]*$")],
|
||||||
|
})
|
||||||
|
|
||||||
SCHEMA_VERSION = vol.Schema({
|
SCHEMA_VERSION = vol.Schema({
|
||||||
vol.Optional(ATTR_VERSION): vol.Coerce(str),
|
vol.Optional(ATTR_VERSION): vol.Coerce(str),
|
||||||
})
|
})
|
||||||
@@ -26,16 +31,25 @@ class APIHomeAssistant(object):
|
|||||||
@api_process
|
@api_process
|
||||||
async def info(self, request):
|
async def info(self, request):
|
||||||
"""Return host information."""
|
"""Return host information."""
|
||||||
info = {
|
return {
|
||||||
ATTR_VERSION: self.homeassistant.version,
|
ATTR_VERSION: self.homeassistant.version,
|
||||||
ATTR_LAST_VERSION: self.config.last_homeassistant,
|
ATTR_LAST_VERSION: self.config.last_homeassistant,
|
||||||
|
ATTR_DEVICES: self.config.homeassistant_devices,
|
||||||
}
|
}
|
||||||
|
|
||||||
return info
|
@api_process
|
||||||
|
async def options(self, request):
|
||||||
|
"""Set homeassistant options."""
|
||||||
|
body = await api_validate(SCHEMA_OPTIONS, request)
|
||||||
|
|
||||||
|
if ATTR_DEVICES in body:
|
||||||
|
self.config.homeassistant_devices = body[ATTR_DEVICES]
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
@api_process
|
@api_process
|
||||||
async def update(self, request):
|
async def update(self, request):
|
||||||
"""Update host OS."""
|
"""Update homeassistant."""
|
||||||
body = await api_validate(SCHEMA_VERSION, request)
|
body = await api_validate(SCHEMA_VERSION, request)
|
||||||
version = body.get(ATTR_VERSION, self.config.last_homeassistant)
|
version = body.get(ATTR_VERSION, self.config.last_homeassistant)
|
||||||
|
|
||||||
@@ -48,6 +62,15 @@ class APIHomeAssistant(object):
|
|||||||
return await asyncio.shield(
|
return await asyncio.shield(
|
||||||
self.homeassistant.update(version), loop=self.loop)
|
self.homeassistant.update(version), loop=self.loop)
|
||||||
|
|
||||||
|
@api_process
|
||||||
|
async def restart(self, request):
|
||||||
|
"""Restart homeassistant."""
|
||||||
|
if self.homeassistant.in_progress:
|
||||||
|
raise RuntimeError("Other task is in progress")
|
||||||
|
|
||||||
|
return await asyncio.shield(
|
||||||
|
self.homeassistant.restart(), loop=self.loop)
|
||||||
|
|
||||||
@api_process_raw
|
@api_process_raw
|
||||||
def logs(self, request):
|
def logs(self, request):
|
||||||
"""Return homeassistant docker logs.
|
"""Return homeassistant docker logs.
|
||||||
|
@@ -1,4 +1,5 @@
|
|||||||
"""Init file for HassIO host rest api."""
|
"""Init file for HassIO host rest api."""
|
||||||
|
import asyncio
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
import voluptuous as vol
|
import voluptuous as vol
|
||||||
@@ -30,7 +31,7 @@ class APIHost(object):
|
|||||||
return {
|
return {
|
||||||
ATTR_TYPE: self.host_control.type,
|
ATTR_TYPE: self.host_control.type,
|
||||||
ATTR_VERSION: self.host_control.version,
|
ATTR_VERSION: self.host_control.version,
|
||||||
ATTR_LAST_VERSION: self.host_control.last,
|
ATTR_LAST_VERSION: self.host_control.last_version,
|
||||||
ATTR_FEATURES: self.host_control.features,
|
ATTR_FEATURES: self.host_control.features,
|
||||||
ATTR_HOSTNAME: self.host_control.hostname,
|
ATTR_HOSTNAME: self.host_control.hostname,
|
||||||
ATTR_OS: self.host_control.os_info,
|
ATTR_OS: self.host_control.os_info,
|
||||||
@@ -50,9 +51,10 @@ class APIHost(object):
|
|||||||
async def update(self, request):
|
async def update(self, request):
|
||||||
"""Update host OS."""
|
"""Update host OS."""
|
||||||
body = await api_validate(SCHEMA_VERSION, request)
|
body = await api_validate(SCHEMA_VERSION, request)
|
||||||
version = body.get(ATTR_VERSION)
|
version = body.get(ATTR_VERSION, self.host_control.last_version)
|
||||||
|
|
||||||
if version == self.host_control.version:
|
if version == self.host_control.version:
|
||||||
raise RuntimeError("Version is already in use")
|
raise RuntimeError("Version is already in use")
|
||||||
|
|
||||||
return await self.host_control.update(version=version)
|
return await asyncio.shield(
|
||||||
|
self.host_control.update(version=version), loop=self.loop)
|
||||||
|
@@ -1,11 +1,19 @@
|
|||||||
"""Init file for HassIO network rest api."""
|
"""Init file for HassIO network rest api."""
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from .util import api_process_hostcontrol
|
import voluptuous as vol
|
||||||
|
|
||||||
|
from .util import api_process, api_process_hostcontrol, api_validate
|
||||||
|
from ..const import ATTR_HOSTNAME
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
SCHEMA_OPTIONS = vol.Schema({
|
||||||
|
vol.Optional(ATTR_HOSTNAME): vol.Coerce(str),
|
||||||
|
})
|
||||||
|
|
||||||
|
|
||||||
class APINetwork(object):
|
class APINetwork(object):
|
||||||
"""Handle rest api for network functions."""
|
"""Handle rest api for network functions."""
|
||||||
|
|
||||||
@@ -15,12 +23,21 @@ class APINetwork(object):
|
|||||||
self.loop = loop
|
self.loop = loop
|
||||||
self.host_control = host_control
|
self.host_control = host_control
|
||||||
|
|
||||||
@api_process_hostcontrol
|
@api_process
|
||||||
def info(self, request):
|
async def info(self, request):
|
||||||
"""Show network settings."""
|
"""Show network settings."""
|
||||||
pass
|
return {
|
||||||
|
ATTR_HOSTNAME: self.host_control.hostname,
|
||||||
|
}
|
||||||
|
|
||||||
@api_process_hostcontrol
|
@api_process_hostcontrol
|
||||||
def options(self, request):
|
async def options(self, request):
|
||||||
"""Edit network settings."""
|
"""Edit network settings."""
|
||||||
pass
|
body = await api_validate(SCHEMA_OPTIONS, request)
|
||||||
|
|
||||||
|
# hostname
|
||||||
|
if ATTR_HOSTNAME in body:
|
||||||
|
if self.host_control.hostname != body[ATTR_HOSTNAME]:
|
||||||
|
await self.host_control.set_hostname(body[ATTR_HOSTNAME])
|
||||||
|
|
||||||
|
return True
|
||||||
|
102
hassio/api/security.py
Normal file
102
hassio/api/security.py
Normal file
@@ -0,0 +1,102 @@
|
|||||||
|
"""Init file for HassIO security rest api."""
|
||||||
|
from datetime import datetime, timedelta
|
||||||
|
import io
|
||||||
|
import logging
|
||||||
|
import hashlib
|
||||||
|
import os
|
||||||
|
|
||||||
|
from aiohttp import web
|
||||||
|
import voluptuous as vol
|
||||||
|
import pyotp
|
||||||
|
import pyqrcode
|
||||||
|
|
||||||
|
from .util import api_process, api_validate, hash_password
|
||||||
|
from ..const import ATTR_INITIALIZE, ATTR_PASSWORD, ATTR_TOTP, ATTR_SESSION
|
||||||
|
|
||||||
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
SCHEMA_PASSWORD = vol.Schema({
|
||||||
|
vol.Required(ATTR_PASSWORD): vol.Coerce(str),
|
||||||
|
})
|
||||||
|
|
||||||
|
SCHEMA_SESSION = SCHEMA_PASSWORD.extend({
|
||||||
|
vol.Optional(ATTR_TOTP, default=None): vol.Coerce(str),
|
||||||
|
})
|
||||||
|
|
||||||
|
|
||||||
|
class APISecurity(object):
|
||||||
|
"""Handle rest api for security functions."""
|
||||||
|
|
||||||
|
def __init__(self, config, loop):
|
||||||
|
"""Initialize security rest api part."""
|
||||||
|
self.config = config
|
||||||
|
self.loop = loop
|
||||||
|
|
||||||
|
def _check_password(self, body):
|
||||||
|
"""Check if password is valid and security is initialize."""
|
||||||
|
if not self.config.security_initialize:
|
||||||
|
raise RuntimeError("First set a password")
|
||||||
|
|
||||||
|
password = hash_password(body[ATTR_PASSWORD])
|
||||||
|
if password != self.config.security_password:
|
||||||
|
raise RuntimeError("Wrong password")
|
||||||
|
|
||||||
|
@api_process
|
||||||
|
async def info(self, request):
|
||||||
|
"""Return host information."""
|
||||||
|
return {
|
||||||
|
ATTR_INITIALIZE: self.config.security_initialize,
|
||||||
|
ATTR_TOTP: self.config.security_totp is not None,
|
||||||
|
}
|
||||||
|
|
||||||
|
@api_process
|
||||||
|
async def options(self, request):
|
||||||
|
"""Set options / password."""
|
||||||
|
body = await api_validate(SCHEMA_PASSWORD, request)
|
||||||
|
|
||||||
|
if self.config.security_initialize:
|
||||||
|
raise RuntimeError("Password is already set!")
|
||||||
|
|
||||||
|
self.config.security_password = hash_password(body[ATTR_PASSWORD])
|
||||||
|
self.config.security_initialize = True
|
||||||
|
return True
|
||||||
|
|
||||||
|
@api_process
|
||||||
|
async def totp(self, request):
|
||||||
|
"""Set and initialze TOTP."""
|
||||||
|
body = await api_validate(SCHEMA_PASSWORD, request)
|
||||||
|
self._check_password(body)
|
||||||
|
|
||||||
|
# generate TOTP
|
||||||
|
totp_init_key = pyotp.random_base32()
|
||||||
|
totp = pyotp.TOTP(totp_init_key)
|
||||||
|
|
||||||
|
# init qrcode
|
||||||
|
buff = io.BytesIO()
|
||||||
|
|
||||||
|
qrcode = pyqrcode.create(totp.provisioning_uri("Hass.IO"))
|
||||||
|
qrcode.svg(buff)
|
||||||
|
|
||||||
|
# finish
|
||||||
|
self.config.security_totp = totp_init_key
|
||||||
|
return web.Response(body=buff.getvalue(), content_type='image/svg+xml')
|
||||||
|
|
||||||
|
@api_process
|
||||||
|
async def session(self, request):
|
||||||
|
"""Set and initialze session."""
|
||||||
|
body = await api_validate(SCHEMA_SESSION, request)
|
||||||
|
self._check_password(body)
|
||||||
|
|
||||||
|
# check TOTP
|
||||||
|
if self.config.security_totp:
|
||||||
|
totp = pyotp.TOTP(self.config.security_totp)
|
||||||
|
if body[ATTR_TOTP] != totp.now():
|
||||||
|
raise RuntimeError("Invalid TOTP token!")
|
||||||
|
|
||||||
|
# create session
|
||||||
|
valid_until = datetime.now() + timedelta(days=1)
|
||||||
|
session = hashlib.sha256(os.urandom(54)).hexdigest()
|
||||||
|
|
||||||
|
# store session
|
||||||
|
self.config.security_sessions = (session, valid_until)
|
||||||
|
return {ATTR_SESSION: session}
|
@@ -5,15 +5,22 @@ import logging
|
|||||||
import voluptuous as vol
|
import voluptuous as vol
|
||||||
|
|
||||||
from .util import api_process, api_process_raw, api_validate
|
from .util import api_process, api_process_raw, api_validate
|
||||||
|
from ..addons.util import create_hash_index_list
|
||||||
from ..const import (
|
from ..const import (
|
||||||
ATTR_ADDONS, ATTR_VERSION, ATTR_LAST_VERSION, ATTR_BETA_CHANNEL,
|
ATTR_ADDONS, ATTR_VERSION, ATTR_LAST_VERSION, ATTR_BETA_CHANNEL,
|
||||||
HASSIO_VERSION)
|
HASSIO_VERSION, ATTR_ADDONS_REPOSITORIES, ATTR_REPOSITORIES,
|
||||||
|
ATTR_REPOSITORY, ATTR_DESCRIPTON, ATTR_NAME, ATTR_SLUG, ATTR_INSTALLED,
|
||||||
|
ATTR_DETACHED, ATTR_SOURCE, ATTR_MAINTAINER, ATTR_URL, ATTR_ARCH,
|
||||||
|
ATTR_BUILD, ATTR_TIMEZONE)
|
||||||
|
from ..tools import validate_timezone
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
SCHEMA_OPTIONS = vol.Schema({
|
SCHEMA_OPTIONS = vol.Schema({
|
||||||
# pylint: disable=no-value-for-parameter
|
# pylint: disable=no-value-for-parameter
|
||||||
vol.Optional(ATTR_BETA_CHANNEL): vol.Boolean(),
|
vol.Optional(ATTR_BETA_CHANNEL): vol.Boolean(),
|
||||||
|
vol.Optional(ATTR_ADDONS_REPOSITORIES): [vol.Url()],
|
||||||
|
vol.Optional(ATTR_TIMEZONE): validate_timezone,
|
||||||
})
|
})
|
||||||
|
|
||||||
SCHEMA_VERSION = vol.Schema({
|
SCHEMA_VERSION = vol.Schema({
|
||||||
@@ -32,6 +39,48 @@ class APISupervisor(object):
|
|||||||
self.addons = addons
|
self.addons = addons
|
||||||
self.host_control = host_control
|
self.host_control = host_control
|
||||||
|
|
||||||
|
def _addons_list(self, only_installed=False):
|
||||||
|
"""Return a list of addons."""
|
||||||
|
detached = self.addons.list_detached
|
||||||
|
|
||||||
|
if only_installed:
|
||||||
|
addons = self.addons.list_installed
|
||||||
|
else:
|
||||||
|
addons = self.addons.list_all
|
||||||
|
|
||||||
|
data = []
|
||||||
|
for addon in addons:
|
||||||
|
data.append({
|
||||||
|
ATTR_NAME: self.addons.get_name(addon),
|
||||||
|
ATTR_SLUG: addon,
|
||||||
|
ATTR_DESCRIPTON: self.addons.get_description(addon),
|
||||||
|
ATTR_VERSION: self.addons.get_last_version(addon),
|
||||||
|
ATTR_INSTALLED: self.addons.version_installed(addon),
|
||||||
|
ATTR_ARCH: self.addons.get_arch(addon),
|
||||||
|
ATTR_DETACHED: addon in detached,
|
||||||
|
ATTR_REPOSITORY: self.addons.get_repository(addon),
|
||||||
|
ATTR_BUILD: self.addons.need_build(addon),
|
||||||
|
ATTR_URL: self.addons.get_url(addon),
|
||||||
|
})
|
||||||
|
|
||||||
|
return data
|
||||||
|
|
||||||
|
def _repositories_list(self):
|
||||||
|
"""Return a list of addons repositories."""
|
||||||
|
data = []
|
||||||
|
list_id = create_hash_index_list(self.config.addons_repositories)
|
||||||
|
|
||||||
|
for repository in self.addons.list_repositories:
|
||||||
|
data.append({
|
||||||
|
ATTR_SLUG: repository[ATTR_SLUG],
|
||||||
|
ATTR_NAME: repository[ATTR_NAME],
|
||||||
|
ATTR_SOURCE: list_id.get(repository[ATTR_SLUG]),
|
||||||
|
ATTR_URL: repository.get(ATTR_URL),
|
||||||
|
ATTR_MAINTAINER: repository.get(ATTR_MAINTAINER),
|
||||||
|
})
|
||||||
|
|
||||||
|
return data
|
||||||
|
|
||||||
@api_process
|
@api_process
|
||||||
async def ping(self, request):
|
async def ping(self, request):
|
||||||
"""Return ok for signal that the api is ready."""
|
"""Return ok for signal that the api is ready."""
|
||||||
@@ -44,7 +93,18 @@ class APISupervisor(object):
|
|||||||
ATTR_VERSION: HASSIO_VERSION,
|
ATTR_VERSION: HASSIO_VERSION,
|
||||||
ATTR_LAST_VERSION: self.config.last_hassio,
|
ATTR_LAST_VERSION: self.config.last_hassio,
|
||||||
ATTR_BETA_CHANNEL: self.config.upstream_beta,
|
ATTR_BETA_CHANNEL: self.config.upstream_beta,
|
||||||
ATTR_ADDONS: self.addons.list_api,
|
ATTR_ARCH: self.addons.arch,
|
||||||
|
ATTR_TIMEZONE: self.config.timezone,
|
||||||
|
ATTR_ADDONS: self._addons_list(only_installed=True),
|
||||||
|
ATTR_ADDONS_REPOSITORIES: self.config.addons_repositories,
|
||||||
|
}
|
||||||
|
|
||||||
|
@api_process
|
||||||
|
async def available_addons(self, request):
|
||||||
|
"""Return information for all available addons."""
|
||||||
|
return {
|
||||||
|
ATTR_ADDONS: self._addons_list(),
|
||||||
|
ATTR_REPOSITORIES: self._repositories_list(),
|
||||||
}
|
}
|
||||||
|
|
||||||
@api_process
|
@api_process
|
||||||
@@ -55,7 +115,28 @@ class APISupervisor(object):
|
|||||||
if ATTR_BETA_CHANNEL in body:
|
if ATTR_BETA_CHANNEL in body:
|
||||||
self.config.upstream_beta = body[ATTR_BETA_CHANNEL]
|
self.config.upstream_beta = body[ATTR_BETA_CHANNEL]
|
||||||
|
|
||||||
return self.config.save()
|
if ATTR_TIMEZONE in body:
|
||||||
|
self.config.timezone = body[ATTR_TIMEZONE]
|
||||||
|
|
||||||
|
if ATTR_ADDONS_REPOSITORIES in body:
|
||||||
|
new = set(body[ATTR_ADDONS_REPOSITORIES])
|
||||||
|
old = set(self.config.addons_repositories)
|
||||||
|
|
||||||
|
# add new repositories
|
||||||
|
tasks = [self.addons.add_git_repository(url) for url in
|
||||||
|
set(new - old)]
|
||||||
|
if tasks:
|
||||||
|
await asyncio.shield(
|
||||||
|
asyncio.wait(tasks, loop=self.loop), loop=self.loop)
|
||||||
|
|
||||||
|
# remove old repositories
|
||||||
|
for url in set(old - new):
|
||||||
|
self.addons.drop_git_repository(url)
|
||||||
|
|
||||||
|
# read repository
|
||||||
|
self.addons.read_data_from_repositories()
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
@api_process
|
@api_process
|
||||||
async def update(self, request):
|
async def update(self, request):
|
||||||
|
@@ -1,5 +1,6 @@
|
|||||||
"""Init file for HassIO util for rest api."""
|
"""Init file for HassIO util for rest api."""
|
||||||
import json
|
import json
|
||||||
|
import hashlib
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from aiohttp import web
|
from aiohttp import web
|
||||||
@@ -32,6 +33,8 @@ def api_process(method):
|
|||||||
|
|
||||||
if isinstance(answer, dict):
|
if isinstance(answer, dict):
|
||||||
return api_return_ok(data=answer)
|
return api_return_ok(data=answer)
|
||||||
|
if isinstance(answer, web.Response):
|
||||||
|
return answer
|
||||||
elif answer:
|
elif answer:
|
||||||
return api_return_ok()
|
return api_return_ok()
|
||||||
return api_return_error()
|
return api_return_error()
|
||||||
@@ -81,7 +84,7 @@ def api_return_error(message=None):
|
|||||||
return web.json_response({
|
return web.json_response({
|
||||||
JSON_RESULT: RESULT_ERROR,
|
JSON_RESULT: RESULT_ERROR,
|
||||||
JSON_MESSAGE: message,
|
JSON_MESSAGE: message,
|
||||||
})
|
}, status=400)
|
||||||
|
|
||||||
|
|
||||||
def api_return_ok(data=None):
|
def api_return_ok(data=None):
|
||||||
@@ -101,3 +104,9 @@ async def api_validate(schema, request):
|
|||||||
raise RuntimeError(humanize_error(data, ex)) from None
|
raise RuntimeError(humanize_error(data, ex)) from None
|
||||||
|
|
||||||
return data
|
return data
|
||||||
|
|
||||||
|
|
||||||
|
def hash_password(password):
|
||||||
|
"""Hash and salt our passwords."""
|
||||||
|
key = ")*()*SALT_HASSIO2123{}6554547485HSKA!!*JSLAfdasda$".format(password)
|
||||||
|
return hashlib.sha256(key.encode()).hexdigest()
|
||||||
|
@@ -1,7 +1,6 @@
|
|||||||
"""Bootstrap HassIO."""
|
"""Bootstrap HassIO."""
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
import stat
|
|
||||||
import signal
|
import signal
|
||||||
|
|
||||||
from colorlog import ColoredFormatter
|
from colorlog import ColoredFormatter
|
||||||
@@ -17,26 +16,46 @@ def initialize_system_data(websession):
|
|||||||
config = CoreConfig(websession)
|
config = CoreConfig(websession)
|
||||||
|
|
||||||
# homeassistant config folder
|
# homeassistant config folder
|
||||||
if not os.path.isdir(config.path_config):
|
if not config.path_config.is_dir():
|
||||||
_LOGGER.info(
|
_LOGGER.info(
|
||||||
"Create Home-Assistant config folder %s", config.path_config)
|
"Create Home-Assistant config folder %s", config.path_config)
|
||||||
os.mkdir(config.path_config)
|
config.path_config.mkdir()
|
||||||
|
|
||||||
# homeassistant ssl folder
|
# hassio ssl folder
|
||||||
if not os.path.isdir(config.path_ssl):
|
if not config.path_ssl.is_dir():
|
||||||
_LOGGER.info("Create Home-Assistant ssl folder %s", config.path_ssl)
|
_LOGGER.info("Create hassio ssl folder %s", config.path_ssl)
|
||||||
os.mkdir(config.path_ssl)
|
config.path_ssl.mkdir()
|
||||||
|
|
||||||
# homeassistant addon data folder
|
# hassio addon data folder
|
||||||
if not os.path.isdir(config.path_addons_data):
|
if not config.path_addons_data.is_dir():
|
||||||
_LOGGER.info("Create Home-Assistant addon data folder %s",
|
_LOGGER.info(
|
||||||
config.path_addons_data)
|
"Create hassio addon data folder %s", config.path_addons_data)
|
||||||
os.mkdir(config.path_addons_data)
|
config.path_addons_data.mkdir(parents=True)
|
||||||
|
|
||||||
if not os.path.isdir(config.path_addons_custom):
|
if not config.path_addons_local.is_dir():
|
||||||
_LOGGER.info("Create Home-Assistant addon custom folder %s",
|
_LOGGER.info("Create hassio addon local repository folder %s",
|
||||||
config.path_addons_custom)
|
config.path_addons_local)
|
||||||
os.mkdir(config.path_addons_custom)
|
config.path_addons_local.mkdir(parents=True)
|
||||||
|
|
||||||
|
if not config.path_addons_git.is_dir():
|
||||||
|
_LOGGER.info("Create hassio addon git repositories folder %s",
|
||||||
|
config.path_addons_git)
|
||||||
|
config.path_addons_git.mkdir(parents=True)
|
||||||
|
|
||||||
|
if not config.path_addons_build.is_dir():
|
||||||
|
_LOGGER.info("Create Home-Assistant addon build folder %s",
|
||||||
|
config.path_addons_build)
|
||||||
|
config.path_addons_build.mkdir(parents=True)
|
||||||
|
|
||||||
|
# hassio backup folder
|
||||||
|
if not config.path_backup.is_dir():
|
||||||
|
_LOGGER.info("Create hassio backup folder %s", config.path_backup)
|
||||||
|
config.path_backup.mkdir()
|
||||||
|
|
||||||
|
# share folder
|
||||||
|
if not config.path_share.is_dir():
|
||||||
|
_LOGGER.info("Create hassio share folder %s", config.path_share)
|
||||||
|
config.path_share.mkdir()
|
||||||
|
|
||||||
return config
|
return config
|
||||||
|
|
||||||
@@ -76,8 +95,7 @@ def check_environment():
|
|||||||
_LOGGER.fatal("Can't find %s in env!", key)
|
_LOGGER.fatal("Can't find %s in env!", key)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
mode = os.stat(SOCKET_DOCKER)[stat.ST_MODE]
|
if not SOCKET_DOCKER.is_socket():
|
||||||
if not stat.S_ISSOCK(mode):
|
|
||||||
_LOGGER.fatal("Can't find docker socket!")
|
_LOGGER.fatal("Can't find docker socket!")
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
257
hassio/config.py
257
hassio/config.py
@@ -1,28 +1,64 @@
|
|||||||
"""Bootstrap HassIO."""
|
"""Bootstrap HassIO."""
|
||||||
|
from datetime import datetime
|
||||||
import logging
|
import logging
|
||||||
|
import json
|
||||||
import os
|
import os
|
||||||
|
from pathlib import Path, PurePath
|
||||||
|
|
||||||
|
import voluptuous as vol
|
||||||
|
from voluptuous.humanize import humanize_error
|
||||||
|
|
||||||
from .const import FILE_HASSIO_CONFIG, HASSIO_SHARE
|
from .const import FILE_HASSIO_CONFIG, HASSIO_SHARE
|
||||||
from .tools import (
|
from .tools import (
|
||||||
fetch_current_versions, write_json_file, read_json_file)
|
fetch_last_versions, write_json_file, read_json_file, validate_timezone)
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
HOMEASSISTANT_CONFIG = "{}/homeassistant"
|
DATETIME_FORMAT = "%Y%m%d %H:%M:%S"
|
||||||
HOMEASSISTANT_IMAGE = 'homeassistant_image'
|
|
||||||
|
HOMEASSISTANT_CONFIG = PurePath("homeassistant")
|
||||||
HOMEASSISTANT_LAST = 'homeassistant_last'
|
HOMEASSISTANT_LAST = 'homeassistant_last'
|
||||||
|
HOMEASSISTANT_DEVICES = 'homeassistant_devices'
|
||||||
|
|
||||||
HASSIO_SSL = "{}/ssl"
|
HASSIO_SSL = PurePath("ssl")
|
||||||
HASSIO_LAST = 'hassio_last'
|
HASSIO_LAST = 'hassio_last'
|
||||||
HASSIO_CLEANUP = 'hassio_cleanup'
|
|
||||||
|
|
||||||
ADDONS_REPO = "{}/addons"
|
ADDONS_CORE = PurePath("addons/core")
|
||||||
ADDONS_DATA = "{}/addons_data"
|
ADDONS_LOCAL = PurePath("addons/local")
|
||||||
ADDONS_CUSTOM = "{}/addons_custom"
|
ADDONS_GIT = PurePath("addons/git")
|
||||||
|
ADDONS_DATA = PurePath("addons/data")
|
||||||
|
ADDONS_BUILD = PurePath("addons/build")
|
||||||
|
ADDONS_CUSTOM_LIST = 'addons_custom_list'
|
||||||
|
|
||||||
|
BACKUP_DATA = PurePath("backup")
|
||||||
|
|
||||||
|
SHARE_DATA = PurePath("share")
|
||||||
|
|
||||||
UPSTREAM_BETA = 'upstream_beta'
|
UPSTREAM_BETA = 'upstream_beta'
|
||||||
|
|
||||||
API_ENDPOINT = 'api_endpoint'
|
API_ENDPOINT = 'api_endpoint'
|
||||||
|
TIMEZONE = 'timezone'
|
||||||
|
|
||||||
|
SECURITY_INITIALIZE = 'security_initialize'
|
||||||
|
SECURITY_TOTP = 'security_totp'
|
||||||
|
SECURITY_PASSWORD = 'security_password'
|
||||||
|
SECURITY_SESSIONS = 'security_sessions'
|
||||||
|
|
||||||
|
|
||||||
|
# pylint: disable=no-value-for-parameter
|
||||||
|
SCHEMA_CONFIG = vol.Schema({
|
||||||
|
vol.Optional(UPSTREAM_BETA, default=False): vol.Boolean(),
|
||||||
|
vol.Optional(API_ENDPOINT): vol.Coerce(str),
|
||||||
|
vol.Optional(TIMEZONE, default='UTC'): validate_timezone,
|
||||||
|
vol.Optional(HOMEASSISTANT_LAST): vol.Coerce(str),
|
||||||
|
vol.Optional(HOMEASSISTANT_DEVICES, default=[]): [vol.Coerce(str)],
|
||||||
|
vol.Optional(HASSIO_LAST): vol.Coerce(str),
|
||||||
|
vol.Optional(ADDONS_CUSTOM_LIST, default=[]): [vol.Url()],
|
||||||
|
vol.Optional(SECURITY_INITIALIZE, default=False): vol.Boolean(),
|
||||||
|
vol.Optional(SECURITY_TOTP): vol.Coerce(str),
|
||||||
|
vol.Optional(SECURITY_PASSWORD): vol.Coerce(str),
|
||||||
|
vol.Optional(SECURITY_SESSIONS, default={}):
|
||||||
|
{vol.Coerce(str): vol.Coerce(str)},
|
||||||
|
}, extra=vol.REMOVE_EXTRA)
|
||||||
|
|
||||||
|
|
||||||
class Config(object):
|
class Config(object):
|
||||||
@@ -30,20 +66,21 @@ class Config(object):
|
|||||||
|
|
||||||
def __init__(self, config_file):
|
def __init__(self, config_file):
|
||||||
"""Initialize config object."""
|
"""Initialize config object."""
|
||||||
self._filename = config_file
|
self._file = config_file
|
||||||
self._data = {}
|
self._data = {}
|
||||||
|
|
||||||
# init or load data
|
# init or load data
|
||||||
if os.path.isfile(self._filename):
|
if self._file.is_file():
|
||||||
try:
|
try:
|
||||||
self._data = read_json_file(self._filename)
|
self._data = read_json_file(self._file)
|
||||||
except OSError:
|
except (OSError, json.JSONDecodeError):
|
||||||
_LOGGER.warning("Can't read %s", self._filename)
|
_LOGGER.warning("Can't read %s", self._file)
|
||||||
|
self._data = {}
|
||||||
|
|
||||||
def save(self):
|
def save(self):
|
||||||
"""Store data to config file."""
|
"""Store data to config file."""
|
||||||
if not write_json_file(self._filename, self._data):
|
if not write_json_file(self._file, self._data):
|
||||||
_LOGGER.exception("Can't store config in %s", self._filename)
|
_LOGGER.error("Can't store config in %s", self._file)
|
||||||
return False
|
return False
|
||||||
return True
|
return True
|
||||||
|
|
||||||
@@ -57,17 +94,17 @@ class CoreConfig(Config):
|
|||||||
|
|
||||||
super().__init__(FILE_HASSIO_CONFIG)
|
super().__init__(FILE_HASSIO_CONFIG)
|
||||||
|
|
||||||
# init data
|
# validate data
|
||||||
if not self._data:
|
try:
|
||||||
self._data.update({
|
self._data = SCHEMA_CONFIG(self._data)
|
||||||
HOMEASSISTANT_IMAGE: os.environ['HOMEASSISTANT_REPOSITORY'],
|
|
||||||
UPSTREAM_BETA: False,
|
|
||||||
})
|
|
||||||
self.save()
|
self.save()
|
||||||
|
except vol.Invalid as ex:
|
||||||
|
_LOGGER.warning(
|
||||||
|
"Invalid config %s", humanize_error(self._data, ex))
|
||||||
|
|
||||||
async def fetch_update_infos(self):
|
async def fetch_update_infos(self):
|
||||||
"""Read current versions from web."""
|
"""Read current versions from web."""
|
||||||
last = await fetch_current_versions(
|
last = await fetch_last_versions(
|
||||||
self.websession, beta=self.upstream_beta)
|
self.websession, beta=self.upstream_beta)
|
||||||
|
|
||||||
if last:
|
if last:
|
||||||
@@ -93,31 +130,40 @@ class CoreConfig(Config):
|
|||||||
@property
|
@property
|
||||||
def upstream_beta(self):
|
def upstream_beta(self):
|
||||||
"""Return True if we run in beta upstream."""
|
"""Return True if we run in beta upstream."""
|
||||||
return self._data.get(UPSTREAM_BETA, False)
|
return self._data[UPSTREAM_BETA]
|
||||||
|
|
||||||
@upstream_beta.setter
|
@upstream_beta.setter
|
||||||
def upstream_beta(self, value):
|
def upstream_beta(self, value):
|
||||||
"""Set beta upstream mode."""
|
"""Set beta upstream mode."""
|
||||||
self._data[UPSTREAM_BETA] = bool(value)
|
self._data[UPSTREAM_BETA] = bool(value)
|
||||||
|
self.save()
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def hassio_cleanup(self):
|
def timezone(self):
|
||||||
"""Return Version they need to cleanup."""
|
"""Return system timezone."""
|
||||||
return self._data.get(HASSIO_CLEANUP)
|
return self._data[TIMEZONE]
|
||||||
|
|
||||||
@hassio_cleanup.setter
|
@timezone.setter
|
||||||
def hassio_cleanup(self, version):
|
def timezone(self, value):
|
||||||
"""Set or remove cleanup flag."""
|
"""Set system timezone."""
|
||||||
if version is None:
|
self._data[TIMEZONE] = value
|
||||||
self._data.pop(HASSIO_CLEANUP, None)
|
self.save()
|
||||||
else:
|
|
||||||
self._data[HASSIO_CLEANUP] = version
|
@property
|
||||||
|
def homeassistant_devices(self):
|
||||||
|
"""Return list of special device to map into homeassistant."""
|
||||||
|
return self._data[HOMEASSISTANT_DEVICES]
|
||||||
|
|
||||||
|
@homeassistant_devices.setter
|
||||||
|
def homeassistant_devices(self, value):
|
||||||
|
"""Set list of special device."""
|
||||||
|
self._data[HOMEASSISTANT_DEVICES] = value
|
||||||
self.save()
|
self.save()
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def homeassistant_image(self):
|
def homeassistant_image(self):
|
||||||
"""Return docker homeassistant repository."""
|
"""Return docker homeassistant repository."""
|
||||||
return self._data.get(HOMEASSISTANT_IMAGE)
|
return os.environ['HOMEASSISTANT_REPOSITORY']
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def last_homeassistant(self):
|
def last_homeassistant(self):
|
||||||
@@ -130,46 +176,155 @@ class CoreConfig(Config):
|
|||||||
return self._data.get(HASSIO_LAST)
|
return self._data.get(HASSIO_LAST)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_hassio_docker(self):
|
def path_extern_hassio(self):
|
||||||
"""Return hassio data path extern for docker."""
|
"""Return hassio data path extern for docker."""
|
||||||
return os.environ['SUPERVISOR_SHARE']
|
return PurePath(os.environ['SUPERVISOR_SHARE'])
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_config_docker(self):
|
def path_extern_config(self):
|
||||||
"""Return config path extern for docker."""
|
"""Return config path extern for docker."""
|
||||||
return HOMEASSISTANT_CONFIG.format(self.path_hassio_docker)
|
return str(PurePath(self.path_extern_hassio, HOMEASSISTANT_CONFIG))
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_config(self):
|
def path_config(self):
|
||||||
"""Return config path inside supervisor."""
|
"""Return config path inside supervisor."""
|
||||||
return HOMEASSISTANT_CONFIG.format(HASSIO_SHARE)
|
return Path(HASSIO_SHARE, HOMEASSISTANT_CONFIG)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_ssl_docker(self):
|
def path_extern_ssl(self):
|
||||||
"""Return SSL path extern for docker."""
|
"""Return SSL path extern for docker."""
|
||||||
return HASSIO_SSL.format(self.path_hassio_docker)
|
return str(PurePath(self.path_extern_hassio, HASSIO_SSL))
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_ssl(self):
|
def path_ssl(self):
|
||||||
"""Return SSL path inside supervisor."""
|
"""Return SSL path inside supervisor."""
|
||||||
return HASSIO_SSL.format(HASSIO_SHARE)
|
return Path(HASSIO_SHARE, HASSIO_SSL)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_addons_repo(self):
|
def path_addons_core(self):
|
||||||
"""Return git repo path for addons."""
|
"""Return git path for core addons."""
|
||||||
return ADDONS_REPO.format(HASSIO_SHARE)
|
return Path(HASSIO_SHARE, ADDONS_CORE)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_addons_custom(self):
|
def path_addons_git(self):
|
||||||
|
"""Return path for git addons."""
|
||||||
|
return Path(HASSIO_SHARE, ADDONS_GIT)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def path_addons_local(self):
|
||||||
"""Return path for customs addons."""
|
"""Return path for customs addons."""
|
||||||
return ADDONS_CUSTOM.format(HASSIO_SHARE)
|
return Path(HASSIO_SHARE, ADDONS_LOCAL)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def path_extern_addons_local(self):
|
||||||
|
"""Return path for customs addons."""
|
||||||
|
return PurePath(self.path_extern_hassio, ADDONS_LOCAL)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_addons_data(self):
|
def path_addons_data(self):
|
||||||
"""Return root addon data folder."""
|
"""Return root addon data folder."""
|
||||||
return ADDONS_DATA.format(HASSIO_SHARE)
|
return Path(HASSIO_SHARE, ADDONS_DATA)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def path_addons_data_docker(self):
|
def path_extern_addons_data(self):
|
||||||
"""Return root addon data folder extern for docker."""
|
"""Return root addon data folder extern for docker."""
|
||||||
return ADDONS_DATA.format(self.path_hassio_docker)
|
return PurePath(self.path_extern_hassio, ADDONS_DATA)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def path_addons_build(self):
|
||||||
|
"""Return root addon build folder."""
|
||||||
|
return Path(HASSIO_SHARE, ADDONS_BUILD)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def path_backup(self):
|
||||||
|
"""Return root backup data folder."""
|
||||||
|
return Path(HASSIO_SHARE, BACKUP_DATA)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def path_extern_backup(self):
|
||||||
|
"""Return root backup data folder extern for docker."""
|
||||||
|
return PurePath(self.path_extern_hassio, BACKUP_DATA)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def path_share(self):
|
||||||
|
"""Return root share data folder."""
|
||||||
|
return Path(HASSIO_SHARE, SHARE_DATA)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def path_extern_share(self):
|
||||||
|
"""Return root share data folder extern for docker."""
|
||||||
|
return PurePath(self.path_extern_hassio, SHARE_DATA)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def addons_repositories(self):
|
||||||
|
"""Return list of addons custom repositories."""
|
||||||
|
return self._data[ADDONS_CUSTOM_LIST]
|
||||||
|
|
||||||
|
@addons_repositories.setter
|
||||||
|
def addons_repositories(self, repo):
|
||||||
|
"""Add a custom repository to list."""
|
||||||
|
if repo in self._data[ADDONS_CUSTOM_LIST]:
|
||||||
|
return
|
||||||
|
|
||||||
|
self._data[ADDONS_CUSTOM_LIST].append(repo)
|
||||||
|
self.save()
|
||||||
|
|
||||||
|
def drop_addon_repository(self, repo):
|
||||||
|
"""Remove a custom repository from list."""
|
||||||
|
if repo not in self._data[ADDONS_CUSTOM_LIST]:
|
||||||
|
return
|
||||||
|
|
||||||
|
self._data[ADDONS_CUSTOM_LIST].remove(repo)
|
||||||
|
self.save()
|
||||||
|
|
||||||
|
@property
|
||||||
|
def security_initialize(self):
|
||||||
|
"""Return is security was initialize."""
|
||||||
|
return self._data[SECURITY_INITIALIZE]
|
||||||
|
|
||||||
|
@security_initialize.setter
|
||||||
|
def security_initialize(self, value):
|
||||||
|
"""Set is security initialize."""
|
||||||
|
self._data[SECURITY_INITIALIZE] = value
|
||||||
|
self.save()
|
||||||
|
|
||||||
|
@property
|
||||||
|
def security_totp(self):
|
||||||
|
"""Return the TOTP key."""
|
||||||
|
return self._data.get(SECURITY_TOTP)
|
||||||
|
|
||||||
|
@security_totp.setter
|
||||||
|
def security_totp(self, value):
|
||||||
|
"""Set the TOTP key."""
|
||||||
|
self._data[SECURITY_TOTP] = value
|
||||||
|
self.save()
|
||||||
|
|
||||||
|
@property
|
||||||
|
def security_password(self):
|
||||||
|
"""Return the password key."""
|
||||||
|
return self._data.get(SECURITY_PASSWORD)
|
||||||
|
|
||||||
|
@security_password.setter
|
||||||
|
def security_password(self, value):
|
||||||
|
"""Set the password key."""
|
||||||
|
self._data[SECURITY_PASSWORD] = value
|
||||||
|
self.save()
|
||||||
|
|
||||||
|
@property
|
||||||
|
def security_sessions(self):
|
||||||
|
"""Return api sessions."""
|
||||||
|
return {session: datetime.strptime(until, DATETIME_FORMAT) for
|
||||||
|
session, until in self._data[SECURITY_SESSIONS].items()}
|
||||||
|
|
||||||
|
@security_sessions.setter
|
||||||
|
def security_sessions(self, value):
|
||||||
|
"""Set the a new session."""
|
||||||
|
session, valid = value
|
||||||
|
if valid is None:
|
||||||
|
self._data[SECURITY_SESSIONS].pop(session, None)
|
||||||
|
else:
|
||||||
|
self._data[SECURITY_SESSIONS].update(
|
||||||
|
{session: valid.strftime(DATETIME_FORMAT)}
|
||||||
|
)
|
||||||
|
|
||||||
|
self.save()
|
||||||
|
@@ -1,28 +1,38 @@
|
|||||||
"""Const file for HassIO."""
|
"""Const file for HassIO."""
|
||||||
HASSIO_VERSION = '0.16'
|
from pathlib import Path
|
||||||
|
|
||||||
|
HASSIO_VERSION = '0.37'
|
||||||
|
|
||||||
URL_HASSIO_VERSION = ('https://raw.githubusercontent.com/home-assistant/'
|
URL_HASSIO_VERSION = ('https://raw.githubusercontent.com/home-assistant/'
|
||||||
'hassio/master/version.json')
|
'hassio/master/version.json')
|
||||||
URL_HASSIO_VERSION_BETA = ('https://raw.githubusercontent.com/home-assistant/'
|
URL_HASSIO_VERSION_BETA = ('https://raw.githubusercontent.com/home-assistant/'
|
||||||
'hassio/master/version_beta.json')
|
'hassio/dev/version.json')
|
||||||
|
|
||||||
URL_HASSIO_ADDONS = 'https://github.com/home-assistant/hassio-addons'
|
URL_HASSIO_ADDONS = 'https://github.com/home-assistant/hassio-addons'
|
||||||
|
|
||||||
DOCKER_REPO = "pvizeli"
|
HASSIO_SHARE = Path("/data")
|
||||||
|
|
||||||
HASSIO_SHARE = "/data"
|
|
||||||
|
|
||||||
RUN_UPDATE_INFO_TASKS = 28800
|
RUN_UPDATE_INFO_TASKS = 28800
|
||||||
RUN_UPDATE_SUPERVISOR_TASKS = 29100
|
RUN_UPDATE_SUPERVISOR_TASKS = 29100
|
||||||
RUN_RELOAD_ADDONS_TASKS = 28800
|
RUN_RELOAD_ADDONS_TASKS = 28800
|
||||||
|
RUN_WATCHDOG_HOMEASSISTANT = 15
|
||||||
|
RUN_CLEANUP_API_SESSIONS = 900
|
||||||
|
|
||||||
RESTART_EXIT_CODE = 100
|
RESTART_EXIT_CODE = 100
|
||||||
|
|
||||||
FILE_HASSIO_ADDONS = "{}/addons.json".format(HASSIO_SHARE)
|
FILE_HASSIO_ADDONS = Path(HASSIO_SHARE, "addons.json")
|
||||||
FILE_HASSIO_CONFIG = "{}/config.json".format(HASSIO_SHARE)
|
FILE_HASSIO_CONFIG = Path(HASSIO_SHARE, "config.json")
|
||||||
|
|
||||||
SOCKET_DOCKER = "/var/run/docker.sock"
|
SOCKET_DOCKER = Path("/var/run/docker.sock")
|
||||||
SOCKET_HC = "/var/run/hassio-hc.sock"
|
SOCKET_HC = Path("/var/run/hassio-hc.sock")
|
||||||
|
|
||||||
|
LABEL_VERSION = 'io.hass.version'
|
||||||
|
LABEL_ARCH = 'io.hass.arch'
|
||||||
|
LABEL_TYPE = 'io.hass.type'
|
||||||
|
|
||||||
|
META_ADDON = 'addon'
|
||||||
|
META_SUPERVISOR = 'supervisor'
|
||||||
|
META_HOMEASSISTANT = 'homeassistant'
|
||||||
|
|
||||||
JSON_RESULT = 'result'
|
JSON_RESULT = 'result'
|
||||||
JSON_DATA = 'data'
|
JSON_DATA = 'data'
|
||||||
@@ -31,9 +41,12 @@ JSON_MESSAGE = 'message'
|
|||||||
RESULT_ERROR = 'error'
|
RESULT_ERROR = 'error'
|
||||||
RESULT_OK = 'ok'
|
RESULT_OK = 'ok'
|
||||||
|
|
||||||
|
ATTR_ARCH = 'arch'
|
||||||
ATTR_HOSTNAME = 'hostname'
|
ATTR_HOSTNAME = 'hostname'
|
||||||
|
ATTR_TIMEZONE = 'timezone'
|
||||||
ATTR_OS = 'os'
|
ATTR_OS = 'os'
|
||||||
ATTR_TYPE = 'type'
|
ATTR_TYPE = 'type'
|
||||||
|
ATTR_SOURCE = 'source'
|
||||||
ATTR_FEATURES = 'features'
|
ATTR_FEATURES = 'features'
|
||||||
ATTR_ADDONS = 'addons'
|
ATTR_ADDONS = 'addons'
|
||||||
ATTR_VERSION = 'version'
|
ATTR_VERSION = 'version'
|
||||||
@@ -45,15 +58,31 @@ ATTR_DESCRIPTON = 'description'
|
|||||||
ATTR_STARTUP = 'startup'
|
ATTR_STARTUP = 'startup'
|
||||||
ATTR_BOOT = 'boot'
|
ATTR_BOOT = 'boot'
|
||||||
ATTR_PORTS = 'ports'
|
ATTR_PORTS = 'ports'
|
||||||
ATTR_MAP_CONFIG = 'map_config'
|
ATTR_MAP = 'map'
|
||||||
ATTR_MAP_SSL = 'map_ssl'
|
|
||||||
ATTR_OPTIONS = 'options'
|
ATTR_OPTIONS = 'options'
|
||||||
ATTR_INSTALLED = 'installed'
|
ATTR_INSTALLED = 'installed'
|
||||||
ATTR_DEDICATED = 'dedicated'
|
ATTR_DETACHED = 'detached'
|
||||||
ATTR_STATE = 'state'
|
ATTR_STATE = 'state'
|
||||||
ATTR_SCHEMA = 'schema'
|
ATTR_SCHEMA = 'schema'
|
||||||
ATTR_IMAGE = 'image'
|
ATTR_IMAGE = 'image'
|
||||||
|
ATTR_ADDONS_REPOSITORIES = 'addons_repositories'
|
||||||
|
ATTR_REPOSITORY = 'repository'
|
||||||
|
ATTR_REPOSITORIES = 'repositories'
|
||||||
|
ATTR_URL = 'url'
|
||||||
|
ATTR_MAINTAINER = 'maintainer'
|
||||||
|
ATTR_PASSWORD = 'password'
|
||||||
|
ATTR_TOTP = 'totp'
|
||||||
|
ATTR_INITIALIZE = 'initialize'
|
||||||
|
ATTR_SESSION = 'session'
|
||||||
|
ATTR_LOCATON = 'location'
|
||||||
|
ATTR_BUILD = 'build'
|
||||||
|
ATTR_DEVICES = 'devices'
|
||||||
|
ATTR_ENVIRONMENT = 'environment'
|
||||||
|
ATTR_HOST_NETWORK = 'host_network'
|
||||||
|
ATTR_TMPFS = 'tmpfs'
|
||||||
|
ATTR_PRIVILEGED = 'privileged'
|
||||||
|
|
||||||
|
STARTUP_INITIALIZE = 'initialize'
|
||||||
STARTUP_BEFORE = 'before'
|
STARTUP_BEFORE = 'before'
|
||||||
STARTUP_AFTER = 'after'
|
STARTUP_AFTER = 'after'
|
||||||
STARTUP_ONCE = 'once'
|
STARTUP_ONCE = 'once'
|
||||||
@@ -63,3 +92,14 @@ BOOT_MANUAL = 'manual'
|
|||||||
|
|
||||||
STATE_STARTED = 'started'
|
STATE_STARTED = 'started'
|
||||||
STATE_STOPPED = 'stopped'
|
STATE_STOPPED = 'stopped'
|
||||||
|
|
||||||
|
MAP_CONFIG = 'config'
|
||||||
|
MAP_SSL = 'ssl'
|
||||||
|
MAP_ADDONS = 'addons'
|
||||||
|
MAP_BACKUP = 'backup'
|
||||||
|
MAP_SHARE = 'share'
|
||||||
|
|
||||||
|
ARCH_ARMHF = 'armhf'
|
||||||
|
ARCH_AARCH64 = 'aarch64'
|
||||||
|
ARCH_AMD64 = 'amd64'
|
||||||
|
ARCH_I386 = 'i386'
|
||||||
|
@@ -11,11 +11,16 @@ from .api import RestAPI
|
|||||||
from .host_control import HostControl
|
from .host_control import HostControl
|
||||||
from .const import (
|
from .const import (
|
||||||
SOCKET_DOCKER, RUN_UPDATE_INFO_TASKS, RUN_RELOAD_ADDONS_TASKS,
|
SOCKET_DOCKER, RUN_UPDATE_INFO_TASKS, RUN_RELOAD_ADDONS_TASKS,
|
||||||
RUN_UPDATE_SUPERVISOR_TASKS, STARTUP_AFTER, STARTUP_BEFORE)
|
RUN_UPDATE_SUPERVISOR_TASKS, RUN_WATCHDOG_HOMEASSISTANT,
|
||||||
|
RUN_CLEANUP_API_SESSIONS, STARTUP_AFTER, STARTUP_BEFORE,
|
||||||
|
STARTUP_INITIALIZE)
|
||||||
from .scheduler import Scheduler
|
from .scheduler import Scheduler
|
||||||
from .dock.homeassistant import DockerHomeAssistant
|
from .dock.homeassistant import DockerHomeAssistant
|
||||||
from .dock.supervisor import DockerSupervisor
|
from .dock.supervisor import DockerSupervisor
|
||||||
from .tools import get_arch_from_image, get_local_ip
|
from .tasks import (
|
||||||
|
hassio_update, homeassistant_watchdog, homeassistant_setup,
|
||||||
|
api_sessions_cleanup)
|
||||||
|
from .tools import get_arch_from_image, get_local_ip, fetch_timezone
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
@@ -32,11 +37,11 @@ class HassIO(object):
|
|||||||
self.scheduler = Scheduler(self.loop)
|
self.scheduler = Scheduler(self.loop)
|
||||||
self.api = RestAPI(self.config, self.loop)
|
self.api = RestAPI(self.config, self.loop)
|
||||||
self.dock = docker.DockerClient(
|
self.dock = docker.DockerClient(
|
||||||
base_url="unix:/{}".format(SOCKET_DOCKER), version='auto')
|
base_url="unix:/{}".format(str(SOCKET_DOCKER)), version='auto')
|
||||||
|
|
||||||
# init basic docker container
|
# init basic docker container
|
||||||
self.supervisor = DockerSupervisor(
|
self.supervisor = DockerSupervisor(
|
||||||
self.config, self.loop, self.dock, self)
|
self.config, self.loop, self.dock, self.stop)
|
||||||
self.homeassistant = DockerHomeAssistant(
|
self.homeassistant = DockerHomeAssistant(
|
||||||
self.config, self.loop, self.dock)
|
self.config, self.loop, self.dock)
|
||||||
|
|
||||||
@@ -49,20 +54,24 @@ class HassIO(object):
|
|||||||
async def setup(self):
|
async def setup(self):
|
||||||
"""Setup HassIO orchestration."""
|
"""Setup HassIO orchestration."""
|
||||||
# supervisor
|
# supervisor
|
||||||
await self.supervisor.attach()
|
if not await self.supervisor.attach():
|
||||||
|
_LOGGER.fatal("Can't attach to supervisor docker container!")
|
||||||
await self.supervisor.cleanup()
|
await self.supervisor.cleanup()
|
||||||
|
|
||||||
# set api endpoint
|
# set api endpoint
|
||||||
self.config.api_endpoint = await get_local_ip(self.loop)
|
self.config.api_endpoint = await get_local_ip(self.loop)
|
||||||
|
|
||||||
|
# update timezone
|
||||||
|
if self.config.timezone == 'UTC':
|
||||||
|
self.config.timezone = await fetch_timezone(self.websession)
|
||||||
|
|
||||||
# hostcontrol
|
# hostcontrol
|
||||||
await self.host_control.load()
|
await self.host_control.load()
|
||||||
_LOGGER.info(
|
|
||||||
"Connected to HostControl. Type: %s Version: %s Hostname: %s "
|
|
||||||
"Features: %s", self.host_control.type,
|
|
||||||
self.host_control.version, self.host_control.hostname,
|
|
||||||
self.host_control.features)
|
|
||||||
|
|
||||||
|
# schedule update info tasks
|
||||||
|
self.scheduler.register_task(
|
||||||
|
|
||||||
|
self.host_control.load, RUN_UPDATE_INFO_TASKS)
|
||||||
# rest api views
|
# rest api views
|
||||||
self.api.register_host(self.host_control)
|
self.api.register_host(self.host_control)
|
||||||
self.api.register_network(self.host_control)
|
self.api.register_network(self.host_control)
|
||||||
@@ -70,6 +79,13 @@ class HassIO(object):
|
|||||||
self.supervisor, self.addons, self.host_control)
|
self.supervisor, self.addons, self.host_control)
|
||||||
self.api.register_homeassistant(self.homeassistant)
|
self.api.register_homeassistant(self.homeassistant)
|
||||||
self.api.register_addons(self.addons)
|
self.api.register_addons(self.addons)
|
||||||
|
self.api.register_security()
|
||||||
|
self.api.register_panel()
|
||||||
|
|
||||||
|
# schedule api session cleanup
|
||||||
|
self.scheduler.register_task(
|
||||||
|
api_sessions_cleanup(self.config), RUN_CLEANUP_API_SESSIONS,
|
||||||
|
now=True)
|
||||||
|
|
||||||
# schedule update info tasks
|
# schedule update info tasks
|
||||||
self.scheduler.register_task(
|
self.scheduler.register_task(
|
||||||
@@ -79,7 +95,10 @@ class HassIO(object):
|
|||||||
# first start of supervisor?
|
# first start of supervisor?
|
||||||
if not await self.homeassistant.exists():
|
if not await self.homeassistant.exists():
|
||||||
_LOGGER.info("No HomeAssistant docker found.")
|
_LOGGER.info("No HomeAssistant docker found.")
|
||||||
await self._setup_homeassistant()
|
await homeassistant_setup(
|
||||||
|
self.config, self.loop, self.homeassistant)
|
||||||
|
else:
|
||||||
|
await self.homeassistant.attach()
|
||||||
|
|
||||||
# Load addons
|
# Load addons
|
||||||
arch = get_arch_from_image(self.supervisor.image)
|
arch = get_arch_from_image(self.supervisor.image)
|
||||||
@@ -91,7 +110,11 @@ class HassIO(object):
|
|||||||
|
|
||||||
# schedule self update task
|
# schedule self update task
|
||||||
self.scheduler.register_task(
|
self.scheduler.register_task(
|
||||||
self._hassio_update, RUN_UPDATE_SUPERVISOR_TASKS)
|
hassio_update(self.config, self.supervisor),
|
||||||
|
RUN_UPDATE_SUPERVISOR_TASKS)
|
||||||
|
|
||||||
|
# start addon mark as initialize
|
||||||
|
await self.addons.auto_boot(STARTUP_INITIALIZE)
|
||||||
|
|
||||||
async def start(self):
|
async def start(self):
|
||||||
"""Start HassIO orchestration."""
|
"""Start HassIO orchestration."""
|
||||||
@@ -99,19 +122,26 @@ class HassIO(object):
|
|||||||
await self.api.start()
|
await self.api.start()
|
||||||
_LOGGER.info("Start hassio api on %s", self.config.api_endpoint)
|
_LOGGER.info("Start hassio api on %s", self.config.api_endpoint)
|
||||||
|
|
||||||
# HomeAssistant is already running / supervisor have only reboot
|
try:
|
||||||
if await self.homeassistant.is_running():
|
# HomeAssistant is already running / supervisor have only reboot
|
||||||
_LOGGER.info("HassIO reboot detected")
|
if await self.homeassistant.is_running():
|
||||||
return
|
_LOGGER.info("HassIO reboot detected")
|
||||||
|
return
|
||||||
|
|
||||||
# start addon mark as before
|
# start addon mark as before
|
||||||
await self.addons.auto_boot(STARTUP_BEFORE)
|
await self.addons.auto_boot(STARTUP_BEFORE)
|
||||||
|
|
||||||
# run HomeAssistant
|
# run HomeAssistant
|
||||||
await self.homeassistant.run()
|
await self.homeassistant.run()
|
||||||
|
|
||||||
# start addon mark as after
|
# start addon mark as after
|
||||||
await self.addons.auto_boot(STARTUP_AFTER)
|
await self.addons.auto_boot(STARTUP_AFTER)
|
||||||
|
|
||||||
|
finally:
|
||||||
|
# schedule homeassistant watchdog
|
||||||
|
self.scheduler.register_task(
|
||||||
|
homeassistant_watchdog(self.loop, self.homeassistant),
|
||||||
|
RUN_WATCHDOG_HOMEASSISTANT)
|
||||||
|
|
||||||
async def stop(self, exit_code=0):
|
async def stop(self, exit_code=0):
|
||||||
"""Stop a running orchestration."""
|
"""Stop a running orchestration."""
|
||||||
@@ -124,28 +154,3 @@ class HassIO(object):
|
|||||||
|
|
||||||
self.exit_code = exit_code
|
self.exit_code = exit_code
|
||||||
self.loop.stop()
|
self.loop.stop()
|
||||||
|
|
||||||
async def _setup_homeassistant(self):
|
|
||||||
"""Install a homeassistant docker container."""
|
|
||||||
while True:
|
|
||||||
# read homeassistant tag and install it
|
|
||||||
if not self.config.last_homeassistant:
|
|
||||||
await self.config.fetch_update_infos()
|
|
||||||
|
|
||||||
tag = self.config.last_homeassistant
|
|
||||||
if tag and await self.homeassistant.install(tag):
|
|
||||||
break
|
|
||||||
_LOGGER.warning("Error on setup HomeAssistant. Retry in 60.")
|
|
||||||
await asyncio.sleep(60, loop=self.loop)
|
|
||||||
|
|
||||||
# store version
|
|
||||||
_LOGGER.info("HomeAssistant docker now installed.")
|
|
||||||
|
|
||||||
async def _hassio_update(self):
|
|
||||||
"""Check and run update of supervisor hassio."""
|
|
||||||
if self.config.last_hassio == self.supervisor.version:
|
|
||||||
return
|
|
||||||
|
|
||||||
_LOGGER.info(
|
|
||||||
"Found new HassIO version %s.", self.config.last_hassio)
|
|
||||||
await self.supervisor.update(self.config.last_hassio)
|
|
||||||
|
@@ -5,7 +5,7 @@ import logging
|
|||||||
|
|
||||||
import docker
|
import docker
|
||||||
|
|
||||||
from ..tools import get_version_from_env
|
from ..const import LABEL_VERSION
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
@@ -19,12 +19,11 @@ class DockerBase(object):
|
|||||||
self.loop = loop
|
self.loop = loop
|
||||||
self.dock = dock
|
self.dock = dock
|
||||||
self.image = image
|
self.image = image
|
||||||
self.container = None
|
|
||||||
self.version = None
|
self.version = None
|
||||||
self._lock = asyncio.Lock(loop=loop)
|
self._lock = asyncio.Lock(loop=loop)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def docker_name(self):
|
def name(self):
|
||||||
"""Return name of docker container."""
|
"""Return name of docker container."""
|
||||||
return None
|
return None
|
||||||
|
|
||||||
@@ -33,6 +32,19 @@ class DockerBase(object):
|
|||||||
"""Return True if a task is in progress."""
|
"""Return True if a task is in progress."""
|
||||||
return self._lock.locked()
|
return self._lock.locked()
|
||||||
|
|
||||||
|
def process_metadata(self, metadata, force=False):
|
||||||
|
"""Read metadata and set it to object."""
|
||||||
|
# read image
|
||||||
|
if not self.image:
|
||||||
|
self.image = metadata['Config']['Image']
|
||||||
|
|
||||||
|
# read metadata
|
||||||
|
need_version = force or not self.version
|
||||||
|
if need_version and LABEL_VERSION in metadata['Config']['Labels']:
|
||||||
|
self.version = metadata['Config']['Labels'][LABEL_VERSION]
|
||||||
|
elif need_version:
|
||||||
|
_LOGGER.warning("Can't read version from %s", self.name)
|
||||||
|
|
||||||
async def install(self, tag):
|
async def install(self, tag):
|
||||||
"""Pull docker image."""
|
"""Pull docker image."""
|
||||||
if self._lock.locked():
|
if self._lock.locked():
|
||||||
@@ -52,12 +64,12 @@ class DockerBase(object):
|
|||||||
image = self.dock.images.pull("{}:{}".format(self.image, tag))
|
image = self.dock.images.pull("{}:{}".format(self.image, tag))
|
||||||
|
|
||||||
image.tag(self.image, tag='latest')
|
image.tag(self.image, tag='latest')
|
||||||
self.version = get_version_from_env(image.attrs['Config']['Env'])
|
self.process_metadata(image.attrs, force=True)
|
||||||
_LOGGER.info("Tag image %s with version %s as latest",
|
|
||||||
self.image, self.version)
|
|
||||||
except docker.errors.APIError as err:
|
except docker.errors.APIError as err:
|
||||||
_LOGGER.error("Can't install %s:%s -> %s.", self.image, tag, err)
|
_LOGGER.error("Can't install %s:%s -> %s.", self.image, tag, err)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
_LOGGER.info("Tag image %s with version %s as latest", self.image, tag)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def exists(self):
|
def exists(self):
|
||||||
@@ -73,8 +85,7 @@ class DockerBase(object):
|
|||||||
Need run inside executor.
|
Need run inside executor.
|
||||||
"""
|
"""
|
||||||
try:
|
try:
|
||||||
image = self.dock.images.get(self.image)
|
self.dock.images.get(self.image)
|
||||||
self.version = get_version_from_env(image.attrs['Config']['Env'])
|
|
||||||
except docker.errors.DockerException:
|
except docker.errors.DockerException:
|
||||||
return False
|
return False
|
||||||
|
|
||||||
@@ -92,17 +103,21 @@ class DockerBase(object):
|
|||||||
|
|
||||||
Need run inside executor.
|
Need run inside executor.
|
||||||
"""
|
"""
|
||||||
if not self.container:
|
try:
|
||||||
try:
|
container = self.dock.containers.get(self.name)
|
||||||
self.container = self.dock.containers.get(self.docker_name)
|
image = self.dock.images.get(self.image)
|
||||||
self.version = get_version_from_env(
|
except docker.errors.DockerException:
|
||||||
self.container.attrs['Config']['Env'])
|
return False
|
||||||
except docker.errors.DockerException:
|
|
||||||
return False
|
|
||||||
else:
|
|
||||||
self.container.reload()
|
|
||||||
|
|
||||||
return self.container.status == 'running'
|
# container is not running
|
||||||
|
if container.status != 'running':
|
||||||
|
return False
|
||||||
|
|
||||||
|
# we run on a old image, stop and start it
|
||||||
|
if container.image.id != image.id:
|
||||||
|
return False
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
async def attach(self):
|
async def attach(self):
|
||||||
"""Attach to running docker container."""
|
"""Attach to running docker container."""
|
||||||
@@ -119,17 +134,17 @@ class DockerBase(object):
|
|||||||
Need run inside executor.
|
Need run inside executor.
|
||||||
"""
|
"""
|
||||||
try:
|
try:
|
||||||
self.container = self.dock.containers.get(self.docker_name)
|
if self.image:
|
||||||
self.image = self.container.attrs['Config']['Image']
|
obj_data = self.dock.images.get(self.image).attrs
|
||||||
self.version = get_version_from_env(
|
else:
|
||||||
self.container.attrs['Config']['Env'])
|
obj_data = self.dock.containers.get(self.name).attrs
|
||||||
_LOGGER.info("Attach to image %s with version %s",
|
except docker.errors.DockerException:
|
||||||
self.image, self.version)
|
|
||||||
except (docker.errors.DockerException, KeyError):
|
|
||||||
_LOGGER.fatal(
|
|
||||||
"Can't attach to %s docker container!", self.docker_name)
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
self.process_metadata(obj_data)
|
||||||
|
_LOGGER.info(
|
||||||
|
"Attach to image %s with version %s", self.image, self.version)
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
async def run(self):
|
async def run(self):
|
||||||
@@ -163,20 +178,19 @@ class DockerBase(object):
|
|||||||
|
|
||||||
Need run inside executor.
|
Need run inside executor.
|
||||||
"""
|
"""
|
||||||
if not self.container:
|
try:
|
||||||
|
container = self.dock.containers.get(self.name)
|
||||||
|
except docker.errors.DockerException:
|
||||||
return
|
return
|
||||||
|
|
||||||
_LOGGER.info("Stop %s docker application", self.image)
|
_LOGGER.info("Stop %s docker application", self.image)
|
||||||
|
|
||||||
self.container.reload()
|
if container.status == 'running':
|
||||||
if self.container.status == 'running':
|
|
||||||
with suppress(docker.errors.DockerException):
|
with suppress(docker.errors.DockerException):
|
||||||
self.container.stop()
|
container.stop()
|
||||||
|
|
||||||
with suppress(docker.errors.DockerException):
|
with suppress(docker.errors.DockerException):
|
||||||
self.container.remove(force=True)
|
container.remove(force=True)
|
||||||
|
|
||||||
self.container = None
|
|
||||||
|
|
||||||
async def remove(self):
|
async def remove(self):
|
||||||
"""Remove docker container."""
|
"""Remove docker container."""
|
||||||
@@ -192,17 +206,21 @@ class DockerBase(object):
|
|||||||
|
|
||||||
Need run inside executor.
|
Need run inside executor.
|
||||||
"""
|
"""
|
||||||
if self._is_running():
|
# cleanup container
|
||||||
self._stop()
|
self._stop()
|
||||||
|
|
||||||
_LOGGER.info("Remove docker %s with latest and %s",
|
_LOGGER.info(
|
||||||
self.image, self.version)
|
"Remove docker %s with latest and %s", self.image, self.version)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
self.dock.images.remove(
|
with suppress(docker.errors.ImageNotFound):
|
||||||
image="{}:latest".format(self.image), force=True)
|
self.dock.images.remove(
|
||||||
self.dock.images.remove(
|
image="{}:latest".format(self.image), force=True)
|
||||||
image="{}:{}".format(self.image, self.version), force=True)
|
|
||||||
|
with suppress(docker.errors.ImageNotFound):
|
||||||
|
self.dock.images.remove(
|
||||||
|
image="{}:{}".format(self.image, self.version), force=True)
|
||||||
|
|
||||||
except docker.errors.DockerException as err:
|
except docker.errors.DockerException as err:
|
||||||
_LOGGER.warning("Can't remove image %s -> %s", self.image, err)
|
_LOGGER.warning("Can't remove image %s -> %s", self.image, err)
|
||||||
return False
|
return False
|
||||||
@@ -223,23 +241,21 @@ class DockerBase(object):
|
|||||||
|
|
||||||
Need run inside executor.
|
Need run inside executor.
|
||||||
"""
|
"""
|
||||||
old_image = "{}:{}".format(self.image, self.version)
|
was_running = self._is_running()
|
||||||
|
|
||||||
_LOGGER.info("Update docker %s with %s:%s",
|
_LOGGER.info(
|
||||||
old_image, self.image, tag)
|
"Update docker %s with %s:%s", self.version, self.image, tag)
|
||||||
|
|
||||||
# update docker image
|
# update docker image
|
||||||
if self._install(tag):
|
if not self._install(tag):
|
||||||
_LOGGER.info("Cleanup old %s docker", old_image)
|
return False
|
||||||
self._stop()
|
|
||||||
try:
|
|
||||||
self.dock.images.remove(image=old_image, force=True)
|
|
||||||
except docker.errors.DockerException as err:
|
|
||||||
_LOGGER.warning(
|
|
||||||
"Can't remove old image %s -> %s", old_image, err)
|
|
||||||
return True
|
|
||||||
|
|
||||||
return False
|
# cleanup old stuff
|
||||||
|
if was_running:
|
||||||
|
self._run()
|
||||||
|
self._cleanup()
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
async def logs(self):
|
async def logs(self):
|
||||||
"""Return docker logs of container."""
|
"""Return docker logs of container."""
|
||||||
@@ -255,10 +271,69 @@ class DockerBase(object):
|
|||||||
|
|
||||||
Need run inside executor.
|
Need run inside executor.
|
||||||
"""
|
"""
|
||||||
if not self.container:
|
try:
|
||||||
return
|
container = self.dock.containers.get(self.name)
|
||||||
|
except docker.errors.DockerException:
|
||||||
|
return b""
|
||||||
|
|
||||||
try:
|
try:
|
||||||
return self.container.logs(tail=100, stdout=True, stderr=True)
|
return container.logs(tail=100, stdout=True, stderr=True)
|
||||||
except docker.errors.DockerException as err:
|
except docker.errors.DockerException as err:
|
||||||
_LOGGER.warning("Can't grap logs from %s -> %s", self.image, err)
|
_LOGGER.warning("Can't grap logs from %s -> %s", self.image, err)
|
||||||
|
|
||||||
|
async def restart(self):
|
||||||
|
"""Restart docker container."""
|
||||||
|
if self._lock.locked():
|
||||||
|
_LOGGER.error("Can't excute restart while a task is in progress")
|
||||||
|
return False
|
||||||
|
|
||||||
|
async with self._lock:
|
||||||
|
return await self.loop.run_in_executor(None, self._restart)
|
||||||
|
|
||||||
|
def _restart(self):
|
||||||
|
"""Restart docker container.
|
||||||
|
|
||||||
|
Need run inside executor.
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
container = self.dock.containers.get(self.name)
|
||||||
|
except docker.errors.DockerException:
|
||||||
|
return False
|
||||||
|
|
||||||
|
_LOGGER.info("Restart %s", self.image)
|
||||||
|
|
||||||
|
try:
|
||||||
|
container.restart(timeout=30)
|
||||||
|
except docker.errors.DockerException as err:
|
||||||
|
_LOGGER.warning("Can't restart %s -> %s", self.image, err)
|
||||||
|
return False
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
|
async def cleanup(self):
|
||||||
|
"""Check if old version exists and cleanup."""
|
||||||
|
if self._lock.locked():
|
||||||
|
_LOGGER.error("Can't excute cleanup while a task is in progress")
|
||||||
|
return False
|
||||||
|
|
||||||
|
async with self._lock:
|
||||||
|
await self.loop.run_in_executor(None, self._cleanup)
|
||||||
|
|
||||||
|
def _cleanup(self):
|
||||||
|
"""Check if old version exists and cleanup.
|
||||||
|
|
||||||
|
Need run inside executor.
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
latest = self.dock.images.get(self.image)
|
||||||
|
except docker.errors.DockerException:
|
||||||
|
_LOGGER.warning("Can't find %s for cleanup", self.image)
|
||||||
|
return
|
||||||
|
|
||||||
|
for image in self.dock.images.list(name=self.image):
|
||||||
|
if latest.id == image.id:
|
||||||
|
continue
|
||||||
|
|
||||||
|
with suppress(docker.errors.DockerException):
|
||||||
|
_LOGGER.info("Cleanup docker images: %s", image.tags)
|
||||||
|
self.dock.images.remove(image.id, force=True)
|
||||||
|
@@ -1,15 +1,17 @@
|
|||||||
"""Init file for HassIO addon docker object."""
|
"""Init file for HassIO addon docker object."""
|
||||||
import logging
|
import logging
|
||||||
|
from pathlib import Path
|
||||||
|
import shutil
|
||||||
|
|
||||||
import docker
|
import docker
|
||||||
|
|
||||||
from . import DockerBase
|
from . import DockerBase
|
||||||
from ..tools import get_version_from_env
|
from .util import dockerfile_template
|
||||||
|
from ..const import (
|
||||||
|
META_ADDON, MAP_CONFIG, MAP_SSL, MAP_ADDONS, MAP_BACKUP, MAP_SHARE)
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
HASS_DOCKER_NAME = 'homeassistant'
|
|
||||||
|
|
||||||
|
|
||||||
class DockerAddon(DockerBase):
|
class DockerAddon(DockerBase):
|
||||||
"""Docker hassio wrapper for HomeAssistant."""
|
"""Docker hassio wrapper for HomeAssistant."""
|
||||||
@@ -22,10 +24,70 @@ class DockerAddon(DockerBase):
|
|||||||
self.addons_data = addons_data
|
self.addons_data = addons_data
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def docker_name(self):
|
def name(self):
|
||||||
"""Return name of docker container."""
|
"""Return name of docker container."""
|
||||||
return "addon_{}".format(self.addon)
|
return "addon_{}".format(self.addon)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def environment(self):
|
||||||
|
"""Return environment for docker add-on."""
|
||||||
|
addon_env = self.addons_data.get_environment(self.addon) or {}
|
||||||
|
|
||||||
|
return {
|
||||||
|
**addon_env,
|
||||||
|
'TZ': self.config.timezone,
|
||||||
|
}
|
||||||
|
|
||||||
|
@property
|
||||||
|
def tmpfs(self):
|
||||||
|
"""Return tmpfs for docker add-on."""
|
||||||
|
options = self.addons_data.get_tmpfs(self.addon)
|
||||||
|
if options:
|
||||||
|
return {"/tmpfs": "{}".format(options)}
|
||||||
|
return None
|
||||||
|
|
||||||
|
@property
|
||||||
|
def volumes(self):
|
||||||
|
"""Generate volumes for mappings."""
|
||||||
|
volumes = {
|
||||||
|
str(self.addons_data.path_extern_data(self.addon)): {
|
||||||
|
'bind': '/data', 'mode': 'rw'
|
||||||
|
}}
|
||||||
|
|
||||||
|
addon_mapping = self.addons_data.map_volumes(self.addon)
|
||||||
|
|
||||||
|
if MAP_CONFIG in addon_mapping:
|
||||||
|
volumes.update({
|
||||||
|
str(self.config.path_extern_config): {
|
||||||
|
'bind': '/config', 'mode': addon_mapping[MAP_CONFIG]
|
||||||
|
}})
|
||||||
|
|
||||||
|
if MAP_SSL in addon_mapping:
|
||||||
|
volumes.update({
|
||||||
|
str(self.config.path_extern_ssl): {
|
||||||
|
'bind': '/ssl', 'mode': addon_mapping[MAP_SSL]
|
||||||
|
}})
|
||||||
|
|
||||||
|
if MAP_ADDONS in addon_mapping:
|
||||||
|
volumes.update({
|
||||||
|
str(self.config.path_extern_addons_local): {
|
||||||
|
'bind': '/addons', 'mode': addon_mapping[MAP_ADDONS]
|
||||||
|
}})
|
||||||
|
|
||||||
|
if MAP_BACKUP in addon_mapping:
|
||||||
|
volumes.update({
|
||||||
|
str(self.config.path_extern_backup): {
|
||||||
|
'bind': '/backup', 'mode': addon_mapping[MAP_BACKUP]
|
||||||
|
}})
|
||||||
|
|
||||||
|
if MAP_SHARE in addon_mapping:
|
||||||
|
volumes.update({
|
||||||
|
str(self.config.path_extern_share): {
|
||||||
|
'bind': '/share', 'mode': addon_mapping[MAP_SHARE]
|
||||||
|
}})
|
||||||
|
|
||||||
|
return volumes
|
||||||
|
|
||||||
def _run(self):
|
def _run(self):
|
||||||
"""Run docker image.
|
"""Run docker image.
|
||||||
|
|
||||||
@@ -34,57 +96,91 @@ class DockerAddon(DockerBase):
|
|||||||
if self._is_running():
|
if self._is_running():
|
||||||
return
|
return
|
||||||
|
|
||||||
# cleanup old container
|
# cleanup
|
||||||
self._stop()
|
self._stop()
|
||||||
|
|
||||||
# volumes
|
|
||||||
volumes = {
|
|
||||||
self.addons_data.path_data_docker(self.addon): {
|
|
||||||
'bind': '/data', 'mode': 'rw'
|
|
||||||
}}
|
|
||||||
if self.addons_data.need_config(self.addon):
|
|
||||||
volumes.update({
|
|
||||||
self.config.path_config_docker: {
|
|
||||||
'bind': '/config', 'mode': 'rw'
|
|
||||||
}})
|
|
||||||
if self.addons_data.need_ssl(self.addon):
|
|
||||||
volumes.update({
|
|
||||||
self.config.path_ssl_docker: {
|
|
||||||
'bind': '/ssl', 'mode': 'rw'
|
|
||||||
}})
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
self.container = self.dock.containers.run(
|
self.dock.containers.run(
|
||||||
self.image,
|
self.image,
|
||||||
name=self.docker_name,
|
name=self.name,
|
||||||
detach=True,
|
detach=True,
|
||||||
network_mode='bridge',
|
network_mode=self.addons_data.get_network_mode(self.addon),
|
||||||
ports=self.addons_data.get_ports(self.addon),
|
ports=self.addons_data.get_ports(self.addon),
|
||||||
volumes=volumes,
|
devices=self.addons_data.get_devices(self.addon),
|
||||||
|
cap_add=self.addons_data.get_privileged(self.addon),
|
||||||
|
environment=self.environment,
|
||||||
|
volumes=self.volumes,
|
||||||
|
tmpfs=self.tmpfs
|
||||||
)
|
)
|
||||||
|
|
||||||
self.version = get_version_from_env(
|
|
||||||
self.container.attrs['Config']['Env'])
|
|
||||||
|
|
||||||
_LOGGER.info("Start docker addon %s with version %s",
|
|
||||||
self.image, self.version)
|
|
||||||
|
|
||||||
except docker.errors.DockerException as err:
|
except docker.errors.DockerException as err:
|
||||||
_LOGGER.error("Can't run %s -> %s", self.image, err)
|
_LOGGER.error("Can't run %s -> %s", self.image, err)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
_LOGGER.info(
|
||||||
|
"Start docker addon %s with version %s", self.image, self.version)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def _attach(self):
|
def _install(self, tag):
|
||||||
"""Attach to running docker container.
|
"""Pull docker image or build it.
|
||||||
|
|
||||||
Need run inside executor.
|
Need run inside executor.
|
||||||
"""
|
"""
|
||||||
|
if self.addons_data.need_build(self.addon):
|
||||||
|
return self._build(tag)
|
||||||
|
|
||||||
|
return super()._install(tag)
|
||||||
|
|
||||||
|
async def build(self, tag):
|
||||||
|
"""Build a docker container."""
|
||||||
|
if self._lock.locked():
|
||||||
|
_LOGGER.error("Can't excute build while a task is in progress")
|
||||||
|
return False
|
||||||
|
|
||||||
|
async with self._lock:
|
||||||
|
return await self.loop.run_in_executor(None, self._build, tag)
|
||||||
|
|
||||||
|
def _build(self, tag):
|
||||||
|
"""Build a docker container.
|
||||||
|
|
||||||
|
Need run inside executor.
|
||||||
|
"""
|
||||||
|
build_dir = Path(self.config.path_addons_build, self.addon)
|
||||||
try:
|
try:
|
||||||
self.container = self.dock.containers.get(self.docker_name)
|
# prepare temporary addon build folder
|
||||||
self.version = get_version_from_env(
|
try:
|
||||||
self.container.attrs['Config']['Env'])
|
source = self.addons_data.path_addon_location(self.addon)
|
||||||
_LOGGER.info("Attach to image %s with version %s",
|
shutil.copytree(str(source), str(build_dir))
|
||||||
self.image, self.version)
|
except shutil.Error as err:
|
||||||
except (docker.errors.DockerException, KeyError):
|
_LOGGER.error("Can't copy %s to temporary build folder -> %s",
|
||||||
pass
|
source, build_dir)
|
||||||
|
return False
|
||||||
|
|
||||||
|
# prepare Dockerfile
|
||||||
|
try:
|
||||||
|
dockerfile_template(
|
||||||
|
Path(build_dir, 'Dockerfile'), self.addons_data.arch,
|
||||||
|
tag, META_ADDON)
|
||||||
|
except OSError as err:
|
||||||
|
_LOGGER.error("Can't prepare dockerfile -> %s", err)
|
||||||
|
|
||||||
|
# run docker build
|
||||||
|
try:
|
||||||
|
build_tag = "{}:{}".format(self.image, tag)
|
||||||
|
|
||||||
|
_LOGGER.info("Start build %s on %s", build_tag, build_dir)
|
||||||
|
image = self.dock.images.build(
|
||||||
|
path=str(build_dir), tag=build_tag, pull=True)
|
||||||
|
|
||||||
|
image.tag(self.image, tag='latest')
|
||||||
|
self.process_metadata(image.attrs, force=True)
|
||||||
|
|
||||||
|
except (docker.errors.DockerException, TypeError) as err:
|
||||||
|
_LOGGER.error("Can't build %s -> %s", build_tag, err)
|
||||||
|
return False
|
||||||
|
|
||||||
|
_LOGGER.info("Build %s done", build_tag)
|
||||||
|
return True
|
||||||
|
|
||||||
|
finally:
|
||||||
|
shutil.rmtree(str(build_dir), ignore_errors=True)
|
||||||
|
@@ -4,7 +4,6 @@ import logging
|
|||||||
import docker
|
import docker
|
||||||
|
|
||||||
from . import DockerBase
|
from . import DockerBase
|
||||||
from ..tools import get_version_from_env
|
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
@@ -19,10 +18,22 @@ class DockerHomeAssistant(DockerBase):
|
|||||||
super().__init__(config, loop, dock, image=config.homeassistant_image)
|
super().__init__(config, loop, dock, image=config.homeassistant_image)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def docker_name(self):
|
def name(self):
|
||||||
"""Return name of docker container."""
|
"""Return name of docker container."""
|
||||||
return HASS_DOCKER_NAME
|
return HASS_DOCKER_NAME
|
||||||
|
|
||||||
|
@property
|
||||||
|
def devices(self):
|
||||||
|
"""Create list of special device to map into docker."""
|
||||||
|
if not self.config.homeassistant_devices:
|
||||||
|
return
|
||||||
|
|
||||||
|
devices = []
|
||||||
|
for device in self.config.homeassistant_devices:
|
||||||
|
devices.append("/dev/{0}:/dev/{0}:rwm".format(device))
|
||||||
|
|
||||||
|
return devices
|
||||||
|
|
||||||
def _run(self):
|
def _run(self):
|
||||||
"""Run docker image.
|
"""Run docker image.
|
||||||
|
|
||||||
@@ -31,47 +42,34 @@ class DockerHomeAssistant(DockerBase):
|
|||||||
if self._is_running():
|
if self._is_running():
|
||||||
return
|
return
|
||||||
|
|
||||||
# cleanup old container
|
# cleanup
|
||||||
self._stop()
|
self._stop()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
self.container = self.dock.containers.run(
|
self.dock.containers.run(
|
||||||
self.image,
|
self.image,
|
||||||
name=self.docker_name,
|
name=self.name,
|
||||||
detach=True,
|
detach=True,
|
||||||
privileged=True,
|
privileged=True,
|
||||||
|
devices=self.devices,
|
||||||
network_mode='host',
|
network_mode='host',
|
||||||
environment={
|
environment={
|
||||||
'HASSIO': self.config.api_endpoint,
|
'HASSIO': self.config.api_endpoint,
|
||||||
|
'TZ': self.config.timezone,
|
||||||
},
|
},
|
||||||
volumes={
|
volumes={
|
||||||
self.config.path_config_docker:
|
str(self.config.path_extern_config):
|
||||||
{'bind': '/config', 'mode': 'rw'},
|
{'bind': '/config', 'mode': 'rw'},
|
||||||
self.config.path_ssl_docker:
|
str(self.config.path_extern_ssl):
|
||||||
{'bind': '/ssl', 'mode': 'rw'},
|
{'bind': '/ssl', 'mode': 'ro'},
|
||||||
|
str(self.config.path_extern_share):
|
||||||
|
{'bind': '/share', 'mode': 'rw'},
|
||||||
})
|
})
|
||||||
|
|
||||||
self.version = get_version_from_env(
|
|
||||||
self.container.attrs['Config']['Env'])
|
|
||||||
|
|
||||||
_LOGGER.info("Start docker addon %s with version %s",
|
|
||||||
self.image, self.version)
|
|
||||||
|
|
||||||
except docker.errors.DockerException as err:
|
except docker.errors.DockerException as err:
|
||||||
_LOGGER.error("Can't run %s -> %s", self.image, err)
|
_LOGGER.error("Can't run %s -> %s", self.image, err)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
_LOGGER.info(
|
||||||
|
"Start homeassistant %s with version %s", self.image, self.version)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
async def update(self, tag):
|
|
||||||
"""Update homeassistant docker image."""
|
|
||||||
if self._lock.locked():
|
|
||||||
_LOGGER.error("Can't excute update while a task is in progress")
|
|
||||||
return False
|
|
||||||
|
|
||||||
async with self._lock:
|
|
||||||
if await self.loop.run_in_executor(None, self._update, tag):
|
|
||||||
await self.loop.run_in_executor(None, self._run)
|
|
||||||
return True
|
|
||||||
|
|
||||||
return False
|
|
||||||
|
@@ -2,8 +2,6 @@
|
|||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
|
|
||||||
import docker
|
|
||||||
|
|
||||||
from . import DockerBase
|
from . import DockerBase
|
||||||
from ..const import RESTART_EXIT_CODE
|
from ..const import RESTART_EXIT_CODE
|
||||||
|
|
||||||
@@ -13,14 +11,13 @@ _LOGGER = logging.getLogger(__name__)
|
|||||||
class DockerSupervisor(DockerBase):
|
class DockerSupervisor(DockerBase):
|
||||||
"""Docker hassio wrapper for HomeAssistant."""
|
"""Docker hassio wrapper for HomeAssistant."""
|
||||||
|
|
||||||
def __init__(self, config, loop, dock, hassio, image=None):
|
def __init__(self, config, loop, dock, stop_callback, image=None):
|
||||||
"""Initialize docker base wrapper."""
|
"""Initialize docker base wrapper."""
|
||||||
super().__init__(config, loop, dock, image=image)
|
super().__init__(config, loop, dock, image=image)
|
||||||
|
self.stop_callback = stop_callback
|
||||||
self.hassio = hassio
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def docker_name(self):
|
def name(self):
|
||||||
"""Return name of docker container."""
|
"""Return name of docker container."""
|
||||||
return os.environ['SUPERVISOR_NAME']
|
return os.environ['SUPERVISOR_NAME']
|
||||||
|
|
||||||
@@ -31,41 +28,14 @@ class DockerSupervisor(DockerBase):
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
_LOGGER.info("Update supervisor docker to %s:%s", self.image, tag)
|
_LOGGER.info("Update supervisor docker to %s:%s", self.image, tag)
|
||||||
old_version = self.version
|
|
||||||
|
|
||||||
async with self._lock:
|
async with self._lock:
|
||||||
if await self.loop.run_in_executor(None, self._install, tag):
|
if await self.loop.run_in_executor(None, self._install, tag):
|
||||||
self.config.hassio_cleanup = old_version
|
self.loop.create_task(self.stop_callback(RESTART_EXIT_CODE))
|
||||||
self.loop.create_task(self.hassio.stop(RESTART_EXIT_CODE))
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
async def cleanup(self):
|
|
||||||
"""Check if old supervisor version exists and cleanup."""
|
|
||||||
if not self.config.hassio_cleanup:
|
|
||||||
return
|
|
||||||
|
|
||||||
async with self._lock:
|
|
||||||
if await self.loop.run_in_executor(None, self._cleanup):
|
|
||||||
self.config.hassio_cleanup = None
|
|
||||||
|
|
||||||
def _cleanup(self):
|
|
||||||
"""Remove old image.
|
|
||||||
|
|
||||||
Need run inside executor.
|
|
||||||
"""
|
|
||||||
old_image = "{}:{}".format(self.image, self.config.hassio_cleanup)
|
|
||||||
|
|
||||||
_LOGGER.info("Old supervisor docker found %s", old_image)
|
|
||||||
try:
|
|
||||||
self.dock.images.remove(image=old_image, force=True)
|
|
||||||
except docker.errors.DockerException as err:
|
|
||||||
_LOGGER.warning("Can't remove old image %s -> %s", old_image, err)
|
|
||||||
return False
|
|
||||||
|
|
||||||
return True
|
|
||||||
|
|
||||||
async def run(self):
|
async def run(self):
|
||||||
"""Run docker image."""
|
"""Run docker image."""
|
||||||
raise RuntimeError("Not support on supervisor docker container!")
|
raise RuntimeError("Not support on supervisor docker container!")
|
||||||
@@ -81,3 +51,7 @@ class DockerSupervisor(DockerBase):
|
|||||||
async def remove(self):
|
async def remove(self):
|
||||||
"""Remove docker image."""
|
"""Remove docker image."""
|
||||||
raise RuntimeError("Not support on supervisor docker container!")
|
raise RuntimeError("Not support on supervisor docker container!")
|
||||||
|
|
||||||
|
async def restart(self):
|
||||||
|
"""Restart docker container."""
|
||||||
|
raise RuntimeError("Not support on supervisor docker container!")
|
||||||
|
40
hassio/dock/util.py
Normal file
40
hassio/dock/util.py
Normal file
@@ -0,0 +1,40 @@
|
|||||||
|
"""HassIO docker utilitys."""
|
||||||
|
import re
|
||||||
|
|
||||||
|
from ..const import ARCH_AARCH64, ARCH_ARMHF, ARCH_I386, ARCH_AMD64
|
||||||
|
|
||||||
|
|
||||||
|
RESIN_BASE_IMAGE = {
|
||||||
|
ARCH_ARMHF: "homeassistant/armhf-base:latest",
|
||||||
|
ARCH_AARCH64: "homeassistant/aarch64-base:latest",
|
||||||
|
ARCH_I386: "homeassistant/i386-base:latest",
|
||||||
|
ARCH_AMD64: "homeassistant/amd64-base:latest",
|
||||||
|
}
|
||||||
|
|
||||||
|
TMPL_IMAGE = re.compile(r"%%BASE_IMAGE%%")
|
||||||
|
|
||||||
|
|
||||||
|
def dockerfile_template(dockerfile, arch, version, meta_type):
|
||||||
|
"""Prepare a Hass.IO dockerfile."""
|
||||||
|
buff = []
|
||||||
|
resin_image = RESIN_BASE_IMAGE[arch]
|
||||||
|
|
||||||
|
# read docker
|
||||||
|
with dockerfile.open('r') as dock_input:
|
||||||
|
for line in dock_input:
|
||||||
|
line = TMPL_IMAGE.sub(resin_image, line)
|
||||||
|
buff.append(line)
|
||||||
|
|
||||||
|
# add metadata
|
||||||
|
buff.append(create_metadata(version, arch, meta_type))
|
||||||
|
|
||||||
|
# write docker
|
||||||
|
with dockerfile.open('w') as dock_output:
|
||||||
|
dock_output.writelines(buff)
|
||||||
|
|
||||||
|
|
||||||
|
def create_metadata(version, arch, meta_type):
|
||||||
|
"""Generate docker label layer for hassio."""
|
||||||
|
return ('LABEL io.hass.version="{}" '
|
||||||
|
'io.hass.arch="{}" '
|
||||||
|
'io.hass.type="{}"').format(version, arch, meta_type)
|
@@ -2,8 +2,6 @@
|
|||||||
import asyncio
|
import asyncio
|
||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
import os
|
|
||||||
import stat
|
|
||||||
|
|
||||||
import async_timeout
|
import async_timeout
|
||||||
|
|
||||||
@@ -19,6 +17,7 @@ UNKNOWN = 'unknown'
|
|||||||
FEATURES_SHUTDOWN = 'shutdown'
|
FEATURES_SHUTDOWN = 'shutdown'
|
||||||
FEATURES_REBOOT = 'reboot'
|
FEATURES_REBOOT = 'reboot'
|
||||||
FEATURES_UPDATE = 'update'
|
FEATURES_UPDATE = 'update'
|
||||||
|
FEATURES_HOSTNAME = 'hostname'
|
||||||
FEATURES_NETWORK_INFO = 'network_info'
|
FEATURES_NETWORK_INFO = 'network_info'
|
||||||
FEATURES_NETWORK_CONTROL = 'network_control'
|
FEATURES_NETWORK_CONTROL = 'network_control'
|
||||||
|
|
||||||
@@ -31,14 +30,13 @@ class HostControl(object):
|
|||||||
self.loop = loop
|
self.loop = loop
|
||||||
self.active = False
|
self.active = False
|
||||||
self.version = UNKNOWN
|
self.version = UNKNOWN
|
||||||
self.last = UNKNOWN
|
self.last_version = UNKNOWN
|
||||||
self.type = UNKNOWN
|
self.type = UNKNOWN
|
||||||
self.features = []
|
self.features = []
|
||||||
self.hostname = UNKNOWN
|
self.hostname = UNKNOWN
|
||||||
self.os_info = UNKNOWN
|
self.os_info = UNKNOWN
|
||||||
|
|
||||||
mode = os.stat(SOCKET_HC)[stat.ST_MODE]
|
if SOCKET_HC.is_socket():
|
||||||
if stat.S_ISSOCK(mode):
|
|
||||||
self.active = True
|
self.active = True
|
||||||
|
|
||||||
async def _send_command(self, command):
|
async def _send_command(self, command):
|
||||||
@@ -50,7 +48,7 @@ class HostControl(object):
|
|||||||
return
|
return
|
||||||
|
|
||||||
reader, writer = await asyncio.open_unix_connection(
|
reader, writer = await asyncio.open_unix_connection(
|
||||||
SOCKET_HC, loop=self.loop)
|
str(SOCKET_HC), loop=self.loop)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# send
|
# send
|
||||||
@@ -60,8 +58,8 @@ class HostControl(object):
|
|||||||
writer.write("{}\n".format(command).encode())
|
writer.write("{}\n".format(command).encode())
|
||||||
data = await reader.readline()
|
data = await reader.readline()
|
||||||
|
|
||||||
response = data.decode()
|
response = data.decode().rstrip()
|
||||||
_LOGGER.debug("Receive from HostControl: %s.", response)
|
_LOGGER.info("Receive from HostControl: %s.", response)
|
||||||
|
|
||||||
if response == "OK":
|
if response == "OK":
|
||||||
return True
|
return True
|
||||||
@@ -73,7 +71,8 @@ class HostControl(object):
|
|||||||
try:
|
try:
|
||||||
return json.loads(response)
|
return json.loads(response)
|
||||||
except json.JSONDecodeError:
|
except json.JSONDecodeError:
|
||||||
_LOGGER.warning("Json parse error from HostControl.")
|
_LOGGER.warning("Json parse error from HostControl '%s'.",
|
||||||
|
response)
|
||||||
|
|
||||||
except asyncio.TimeoutError:
|
except asyncio.TimeoutError:
|
||||||
_LOGGER.error("Timeout from HostControl!")
|
_LOGGER.error("Timeout from HostControl!")
|
||||||
@@ -91,7 +90,7 @@ class HostControl(object):
|
|||||||
return
|
return
|
||||||
|
|
||||||
self.version = info.get(ATTR_VERSION, UNKNOWN)
|
self.version = info.get(ATTR_VERSION, UNKNOWN)
|
||||||
self.last = info.get(ATTR_LAST_VERSION, UNKNOWN)
|
self.last_version = info.get(ATTR_LAST_VERSION, UNKNOWN)
|
||||||
self.type = info.get(ATTR_TYPE, UNKNOWN)
|
self.type = info.get(ATTR_TYPE, UNKNOWN)
|
||||||
self.features = info.get(ATTR_FEATURES, [])
|
self.features = info.get(ATTR_FEATURES, [])
|
||||||
self.hostname = info.get(ATTR_HOSTNAME, UNKNOWN)
|
self.hostname = info.get(ATTR_HOSTNAME, UNKNOWN)
|
||||||
@@ -119,3 +118,7 @@ class HostControl(object):
|
|||||||
if version:
|
if version:
|
||||||
return self._send_command("update {}".format(version))
|
return self._send_command("update {}".format(version))
|
||||||
return self._send_command("update")
|
return self._send_command("update")
|
||||||
|
|
||||||
|
def set_hostname(self, hostname):
|
||||||
|
"""Update hostname on host."""
|
||||||
|
return self._send_command("hostname {}".format(hostname))
|
||||||
|
17
hassio/panel/hassio-main.html
Normal file
17
hassio/panel/hassio-main.html
Normal file
File diff suppressed because one or more lines are too long
BIN
hassio/panel/hassio-main.html.gz
Normal file
BIN
hassio/panel/hassio-main.html.gz
Normal file
Binary file not shown.
60
hassio/tasks.py
Normal file
60
hassio/tasks.py
Normal file
@@ -0,0 +1,60 @@
|
|||||||
|
"""Multible tasks."""
|
||||||
|
import asyncio
|
||||||
|
from datetime import datetime
|
||||||
|
import logging
|
||||||
|
|
||||||
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def api_sessions_cleanup(config):
|
||||||
|
"""Create scheduler task for cleanup api sessions."""
|
||||||
|
async def _api_sessions_cleanup():
|
||||||
|
"""Cleanup old api sessions."""
|
||||||
|
now = datetime.now()
|
||||||
|
for session, until_valid in config.security_sessions.items():
|
||||||
|
if now >= until_valid:
|
||||||
|
config.security_sessions = (session, None)
|
||||||
|
|
||||||
|
return _api_sessions_cleanup
|
||||||
|
|
||||||
|
|
||||||
|
def hassio_update(config, supervisor):
|
||||||
|
"""Create scheduler task for update of supervisor hassio."""
|
||||||
|
async def _hassio_update():
|
||||||
|
"""Check and run update of supervisor hassio."""
|
||||||
|
if config.last_hassio == supervisor.version:
|
||||||
|
return
|
||||||
|
|
||||||
|
_LOGGER.info("Found new HassIO version %s.", config.last_hassio)
|
||||||
|
await supervisor.update(config.last_hassio)
|
||||||
|
|
||||||
|
return _hassio_update
|
||||||
|
|
||||||
|
|
||||||
|
def homeassistant_watchdog(loop, homeassistant):
|
||||||
|
"""Create scheduler task for montoring running state."""
|
||||||
|
async def _homeassistant_watchdog():
|
||||||
|
"""Check running state and start if they is close."""
|
||||||
|
if homeassistant.in_progress or await homeassistant.is_running():
|
||||||
|
return
|
||||||
|
|
||||||
|
loop.create_task(homeassistant.run())
|
||||||
|
|
||||||
|
return _homeassistant_watchdog
|
||||||
|
|
||||||
|
|
||||||
|
async def homeassistant_setup(config, loop, homeassistant):
|
||||||
|
"""Install a homeassistant docker container."""
|
||||||
|
while True:
|
||||||
|
# read homeassistant tag and install it
|
||||||
|
if not config.last_homeassistant:
|
||||||
|
await config.fetch_update_infos()
|
||||||
|
|
||||||
|
tag = config.last_homeassistant
|
||||||
|
if tag and await homeassistant.install(tag):
|
||||||
|
break
|
||||||
|
_LOGGER.warning("Error on setup HomeAssistant. Retry in 60.")
|
||||||
|
await asyncio.sleep(60, loop=loop)
|
||||||
|
|
||||||
|
# store version
|
||||||
|
_LOGGER.info("HomeAssistant docker now installed.")
|
@@ -1,5 +1,6 @@
|
|||||||
"""Tools file for HassIO."""
|
"""Tools file for HassIO."""
|
||||||
import asyncio
|
import asyncio
|
||||||
|
from contextlib import suppress
|
||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
import re
|
import re
|
||||||
@@ -7,16 +8,20 @@ import socket
|
|||||||
|
|
||||||
import aiohttp
|
import aiohttp
|
||||||
import async_timeout
|
import async_timeout
|
||||||
|
import pytz
|
||||||
|
import voluptuous as vol
|
||||||
|
|
||||||
from .const import URL_HASSIO_VERSION, URL_HASSIO_VERSION_BETA
|
from .const import URL_HASSIO_VERSION, URL_HASSIO_VERSION_BETA
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
FREEGEOIP_URL = "https://freegeoip.io/json/"
|
||||||
|
|
||||||
_RE_VERSION = re.compile(r"VERSION=(.*)")
|
_RE_VERSION = re.compile(r"VERSION=(.*)")
|
||||||
_IMAGE_ARCH = re.compile(r".*/([a-z0-9]*)-hassio-supervisor")
|
_IMAGE_ARCH = re.compile(r".*/([a-z0-9]*)-hassio-supervisor")
|
||||||
|
|
||||||
|
|
||||||
async def fetch_current_versions(websession, beta=False):
|
async def fetch_last_versions(websession, beta=False):
|
||||||
"""Fetch current versions from github.
|
"""Fetch current versions from github.
|
||||||
|
|
||||||
Is a coroutine.
|
Is a coroutine.
|
||||||
@@ -41,17 +46,6 @@ def get_arch_from_image(image):
|
|||||||
return found.group(1)
|
return found.group(1)
|
||||||
|
|
||||||
|
|
||||||
def get_version_from_env(env_list):
|
|
||||||
"""Extract Version from ENV list."""
|
|
||||||
for env in env_list:
|
|
||||||
found = _RE_VERSION.match(env)
|
|
||||||
if found:
|
|
||||||
return found.group(1)
|
|
||||||
|
|
||||||
_LOGGER.error("Can't find VERSION in env")
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
def get_local_ip(loop):
|
def get_local_ip(loop):
|
||||||
"""Retrieve local IP address.
|
"""Retrieve local IP address.
|
||||||
|
|
||||||
@@ -77,9 +71,10 @@ def get_local_ip(loop):
|
|||||||
def write_json_file(jsonfile, data):
|
def write_json_file(jsonfile, data):
|
||||||
"""Write a json file."""
|
"""Write a json file."""
|
||||||
try:
|
try:
|
||||||
with open(jsonfile, 'w') as conf_file:
|
json_str = json.dumps(data, indent=2)
|
||||||
conf_file.write(json.dumps(data))
|
with jsonfile.open('w') as conf_file:
|
||||||
except OSError:
|
conf_file.write(json_str)
|
||||||
|
except (OSError, json.JSONDecodeError):
|
||||||
return False
|
return False
|
||||||
|
|
||||||
return True
|
return True
|
||||||
@@ -87,5 +82,30 @@ def write_json_file(jsonfile, data):
|
|||||||
|
|
||||||
def read_json_file(jsonfile):
|
def read_json_file(jsonfile):
|
||||||
"""Read a json file and return a dict."""
|
"""Read a json file and return a dict."""
|
||||||
with open(jsonfile, 'r') as cfile:
|
with jsonfile.open('r') as cfile:
|
||||||
return json.loads(cfile.read())
|
return json.loads(cfile.read())
|
||||||
|
|
||||||
|
|
||||||
|
def validate_timezone(timezone):
|
||||||
|
"""Validate voluptuous timezone."""
|
||||||
|
try:
|
||||||
|
pytz.timezone(timezone)
|
||||||
|
except pytz.exceptions.UnknownTimeZoneError:
|
||||||
|
raise vol.Invalid(
|
||||||
|
"Invalid time zone passed in. Valid options can be found here: "
|
||||||
|
"http://en.wikipedia.org/wiki/List_of_tz_database_time_zones") \
|
||||||
|
from None
|
||||||
|
|
||||||
|
return timezone
|
||||||
|
|
||||||
|
|
||||||
|
async def fetch_timezone(websession):
|
||||||
|
"""Read timezone from freegeoip."""
|
||||||
|
data = {}
|
||||||
|
with suppress(aiohttp.ClientError, asyncio.TimeoutError,
|
||||||
|
json.JSONDecodeError, KeyError):
|
||||||
|
with async_timeout.timeout(10, loop=websession.loop):
|
||||||
|
async with websession.get(FREEGEOIP_URL) as request:
|
||||||
|
data = await request.json()
|
||||||
|
|
||||||
|
return data.get('time_zone', 'UTC')
|
||||||
|
1
home-assistant-polymer
Submodule
1
home-assistant-polymer
Submodule
Submodule home-assistant-polymer added at c5a5f41d3c
BIN
misc/hassio.png
Normal file
BIN
misc/hassio.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 42 KiB |
1
misc/hassio.xml
Normal file
1
misc/hassio.xml
Normal file
@@ -0,0 +1 @@
|
|||||||
|
<mxfile userAgent="Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.81 Safari/537.36" version="6.5.6" editor="www.draw.io" type="device"><diagram name="Page-1">5Vptc6M2EP41/ng3gHj9mPiSy820c5n6Q3sfsVBsNTJyhYid/voKkABZkOBY+KYtmYnR6pVn99ld1l6A5e74laX77a80Q2ThOdlxAb4sPC8OY/G/Erw2At9xG8GG4awR9QQr/DeSQkdKS5yhQhvIKSUc73UhpHmOINdkKWP0oA97okTfdZ9ukCFYwZSY0t9xxrdS6oZJ1/GA8GYrt469sOlYp/B5w2iZy/0WHniqr6Z7l6q15IMW2zSjh54I3C3AklHKm7vdcYlIBa2CrZl3P9LbnpuhnE+Z4DUTXlJSInXikIipt09UrCAOyF8lKOFfJVUdn4paZTdigNjtKD5ERw206DtIYKrenLJdSrrJ4m5TfX5fqX3E2Zqtmg4JS7urd9hijlb7FFbtg7A2MWjLd0S03Oo0mJAlJZTVowXYKIRQyAvO6DPq9Tj1Jc+/kutLvF4Q4+g4CqHbKkbYO6I7xNmrGKImJKCZIm09SKRuD53l+Arobc9oQjkulca6aZfuFCZupM6G9QcM/X3LcaW31WvB0e5CNGGG1vF6CE0QggRkrb7sAhhNBNCzAKBvAPiFwmfELkUOokCQ/trI+SZy3hBywAJyoYHcw9JArXaFqJpRUe9MLscQDXN5HQd+4NjB0A8DHcPQxDBwTAgDCxAmBl4oE3FINinjW7qheUruOumtjmgPPXTE/I9K/DkKZPOH6srFwZq+QDV/yBX+RJy/ygiclpwKUbfxL5Tu5RrNUavzvQ20eBxaMihHRTJ4p2yDeM9uTHUwRFKOX/TVLwFX5RK20fXeQDcB3im+deMRMSweALGfBbp/JdCj0Xxi3UX48xIMN6wSjNMEYlXuEXvBhXAJagOm+h7Sovj2fTTBaMXr0aSjMwP3fbdluKflMgybVEN3aFmA4sy347ZAoLstMJB1uPGA33JtRE3Xm4Nbbo9Yyou13NJ4VbuxeUnkqveOHouiK7EIzOO6NHh1dE/iQtc89VyFwIPfVK9YQgCJYBqGSnyPidpzqm5QnpmLCWFvqcFMfrm0qlgvvlZQUm8cvaxJrPLpRjy6wLByU9dxRSmKn6CtLFR3Rd5A/t56HS1/9224ovDKXHE/O3qQ/+zG8aWBfiKtPmjxwLR4d0Sn1i3enyVUSJ30srCJCPYcTk5zpHmb8xQ2Vl+AJXtp+WpPYdeKPa5ZUrjJMpoXhhqLbbqvbveMQlQU73sn3ZVN9lX34qr9fZMTCt07XhiBxANhEHtx7PhgpqRqyJN5bmB6ssSCI1O1nDmJ0rVOHdWlqYAkU59uc7zoXEAAOfWR4vq9Q5WqneE0Wq3Q0FJO6hdSz1ynobKxTm0U7dNMs5PYJCjk1KxYKX6WO9IMALcVOzAUyKdrRB5pgTmmuRiyppzTnRhAqo7btoitVVbrMna3xg3Bm2oup+fRvCvEnpZu5QYWiHxS0wEDNR0wkJBYqciaNJ5AUifSWOq/x1LX5OgUOk5Ity8PgO97LQshEng/L0SqvXsMPBwOpvcmBO+LWg2SiZDQMrs4Tl6FQInuz3xnIKeP5iovgLcLo9K4P5DEn8mRmTLEXqzt3hyaQ3qj0faDNPFNmjTmaz+S+icmc+pN7YVAMP6tjfNQrkcjIUzZ5fQL62uAfkH1Z4d+CThJJ4boN1TdsxLBopnY17f7yGaWOT9lP8i+YAb2TVZjYJDkK+bbuekxFp2QmwUomocevnppvQo94v9LcEpCnaOR5dgU/idjk/m9+G9oX71qUYbReBXl30s+Vf6dgXyi2f0WqlFG93szcPcP</diagram></mxfile>
|
BIN
misc/security.png
Normal file
BIN
misc/security.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 36 KiB |
1
misc/security.xml
Normal file
1
misc/security.xml
Normal file
@@ -0,0 +1 @@
|
|||||||
|
<mxfile userAgent="Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:53.0) Gecko/20100101 Firefox/53.0" version="6.5.8" editor="www.draw.io" type="device"><diagram name="Page-1">5Vxdd5s4EP01fmwOkgCbx9hp2j7sNrvpnnYfiVFsTjDyghwn++tXGMmAxileEB9O+9BjBhjM3GHmzjXKhCw2L58Sf7v+jQU0mmAreJmQmwnGU4eI/zPDa24gyM0NqyQMchMqDPfhv1QaLWndhQFNKwdyxiIebqvGJYtjuuQVm58kbF897JFF1atu/RUFhvulH0Hr9zDg69w6w25h/0zD1VpdGblevufBXz6tEraL5fUmmDwe/uW7N77yJW80XfsB25dM5OOELBLGeP5p87KgURZaFbb8vNs39h6/d0Jjfs4JOD/h2Y928tZvwyTlwnTP/YTLL8lfVWA4fRF+52u+iYQBiY8pT9gTXbCIJcISs1gcOX8Mo0gz+VG4isXmUnwzKuzzZ5rwUIT8Wu7YhEGQXWa+X4ec3m/9ZXbNvcivzCGL+b38Go7aztMGeWIb3rcMRXYV+lIyyTh8omxDefIqDpF7ySw/Q6asKxHaF/gjS9rWJewVkr5MudXRcRF28UFG/jQKBKDwVypipAe/FPUtC2N+uKIznzg3mYUmobhwFtoblvA1W7HYj+4KawcxQhgGyT0Vo5mBINkgSJ/9NB1hkDAiw0XJAVFaiyhdffk6wkDZ7oCBckGg2JbGh1uKs2b2drT0wvXAOGcbsYPGwXXWfDJbxJZPP4uSqK4ryiuZTYNKU4JhK4VFRSChkc/D52rbOhUW6e0uQ7pAwNOeZ1sLbMp2yZLKk8ptRPMjoNMc4aqj/HaBowNIxzs8C7cpwE2ckdLlLgm5uNPbMH5kvaLnDIYenmrPj9sQPuLUODIH3wzCNxVxFtdz/9llrGcexiEvtibkOiNwfpTS7KjpTVtsD085mQd+uqaBPE/slmRilm29hPyH+PzBurIcuf232LauCFH7S5XwxvpZpuQQVDKlyaPfMlNsy60AjK2mmYJrHJnLFA9kip8+ZfsP+WHdfe8+E856/kk/EOqsApOGECJS48gchGqcK2GYUm4Sw8vss7hpoT5GVDlyvM6wg6NhtdGyLQ9ZLAi4G2WF+kHMK+7qULK1gr4VBHTPkkAv6nrJt7b70iFGir1Kj/K4iC6vsWPPUGMHjgzmCxxiq/mS0jQVCfNGvvyvZOk1VxQdQFcWmlbowNRtRQfsMacc0XWNpikHHL2RcgIG/7V0mJxJWyYlFA306lSk5Rv5Jg94oq+mM66egDSqW31xSm16J9OmGTOrcWSwSEF5xMi43xGSA1FL0rTd6NQSODKIJNRvfmfJxodQvmPJGlfZoN2nZo2gEHMZorWDYJQ6UxkR1DsuRLXuN0xw2L8c2brXSGE4Ug+mW6vkHn6gdpqKIbpw7RDcVcc6JtpolGv11I1g3HAcQ+MGcGQQwBOKyBnaNU/E0XhROY4zvn2fGrfKqUZ1wrDK7TSWTXCNI4NJBWWTXOYejb6tiF7fU4jbVIHQpxDgyCB6UF/IZ4Xete3x9GK3aSnXxW3X7kzcPvHrfzdi5SAypVuVKV3itqros1EzhykyxByAoz6FylOvNbx7obI3XqANbNPG70nMahwZrFBQOBizUjkUSZjqM3VTkgAcGYQSihuXoZR5fQobBAobF6KU9RsmqCJcjlLWb6TguD6YUqaSe3h27plSyrzulDJS9ypB70qZeupGwHc9U0oZcGQQwPqf3dsoZflxFy6UkTZlwrBQ5pkSyoAjgzkFf7ovhLLbb1+/3XWfDGfVCnzubGyYCiPLlGAGPRmEESovZcXMCJAX2pqRZUo5Q1Z30hmpW4DRjXSWdYVDLzgcNcu64gVqaSrZRsotEDIlpkFPfapppH6VyftT03ojD/qqvebLjmZ1ngyWLSjCjFlPG4xEIFOCGvRkDky1TPHEy3+iSooiia2TPOLXeRVw5kqeVWoauKtXAW2oSY1U4LQ1noQ9G4SpuwXsGIRptAqnM2ScoPwzZolz0FBBouMvRTvwOT3WQJ2GywJZEHAzHLrgzIpB54wZ2a0Ys32iOaoHaQDGfHyd+rjQXWld7ZfMqwbaQb+E5Kc6s0mVzeDANsR6LNIy1fCJVDt3CUYXw5lWWWyvYaoRp85Tn8OZA8nbH39+WLCAts2YrtZTnVtuWg9Wem1pysXJTAPcsc8DvAmckPyNHM5z9ZbWo5UOgtvw+UWkzpNBOCFJ/ZKvzv7lJiqtPx8LV3l1lXpNp+VIJTaLv/mWo1b8XT3y8T8=</diagram></mxfile>
|
3
setup.py
3
setup.py
@@ -38,5 +38,8 @@ setup(
|
|||||||
'colorlog',
|
'colorlog',
|
||||||
'voluptuous',
|
'voluptuous',
|
||||||
'gitpython',
|
'gitpython',
|
||||||
|
'pyotp',
|
||||||
|
'pyqrcode',
|
||||||
|
'pytz'
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
@@ -1,7 +1,7 @@
|
|||||||
{
|
{
|
||||||
"hassio": "0.15",
|
"hassio": "0.37",
|
||||||
"homeassistant": "0.43.1",
|
"homeassistant": "0.46.1",
|
||||||
"resinos": "0.4",
|
"resinos": "0.8",
|
||||||
"resinhup": "0.1",
|
"resinhup": "0.1",
|
||||||
"generic": "0.2"
|
"generic": "0.3"
|
||||||
}
|
}
|
||||||
|
@@ -1,7 +0,0 @@
|
|||||||
{
|
|
||||||
"hassio": "0.15",
|
|
||||||
"homeassistant": "0.43.1",
|
|
||||||
"resinos": "0.4",
|
|
||||||
"resinhup": "0.1",
|
|
||||||
"generic": "0.2"
|
|
||||||
}
|
|
Reference in New Issue
Block a user