mirror of
https://github.com/home-assistant/core.git
synced 2025-07-27 15:17:35 +00:00
parent
20f6bd309e
commit
a62619894a
@ -8,7 +8,7 @@ from homeassistant.helpers import config_entry_oauth2_flow
|
|||||||
from .api import OndiloClient
|
from .api import OndiloClient
|
||||||
from .config_flow import OndiloIcoOAuth2FlowHandler
|
from .config_flow import OndiloIcoOAuth2FlowHandler
|
||||||
from .const import DOMAIN
|
from .const import DOMAIN
|
||||||
from .coordinator import OndiloIcoCoordinator
|
from .coordinator import OndiloIcoPoolsCoordinator
|
||||||
from .oauth_impl import OndiloOauth2Implementation
|
from .oauth_impl import OndiloOauth2Implementation
|
||||||
|
|
||||||
PLATFORMS = [Platform.SENSOR]
|
PLATFORMS = [Platform.SENSOR]
|
||||||
@ -28,7 +28,7 @@ async def async_setup_entry(hass: HomeAssistant, entry: ConfigEntry) -> bool:
|
|||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
coordinator = OndiloIcoCoordinator(
|
coordinator = OndiloIcoPoolsCoordinator(
|
||||||
hass, entry, OndiloClient(hass, entry, implementation)
|
hass, entry, OndiloClient(hass, entry, implementation)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -1,7 +1,10 @@
|
|||||||
"""Define an object to coordinate fetching Ondilo ICO data."""
|
"""Define an object to coordinate fetching Ondilo ICO data."""
|
||||||
|
|
||||||
from dataclasses import dataclass
|
from __future__ import annotations
|
||||||
from datetime import timedelta
|
|
||||||
|
import asyncio
|
||||||
|
from dataclasses import dataclass, field
|
||||||
|
from datetime import datetime, timedelta
|
||||||
import logging
|
import logging
|
||||||
from typing import Any
|
from typing import Any
|
||||||
|
|
||||||
@ -9,25 +12,37 @@ from ondilo import OndiloError
|
|||||||
|
|
||||||
from homeassistant.config_entries import ConfigEntry
|
from homeassistant.config_entries import ConfigEntry
|
||||||
from homeassistant.core import HomeAssistant
|
from homeassistant.core import HomeAssistant
|
||||||
|
from homeassistant.helpers import device_registry as dr
|
||||||
from homeassistant.helpers.update_coordinator import DataUpdateCoordinator, UpdateFailed
|
from homeassistant.helpers.update_coordinator import DataUpdateCoordinator, UpdateFailed
|
||||||
|
from homeassistant.util import dt as dt_util
|
||||||
|
|
||||||
from . import DOMAIN
|
from . import DOMAIN
|
||||||
from .api import OndiloClient
|
from .api import OndiloClient
|
||||||
|
|
||||||
_LOGGER = logging.getLogger(__name__)
|
_LOGGER = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
TIME_TO_NEXT_UPDATE = timedelta(hours=1, minutes=5)
|
||||||
|
UPDATE_LOCK = asyncio.Lock()
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
class OndiloIcoData:
|
class OndiloIcoPoolData:
|
||||||
"""Class for storing the data."""
|
"""Store the pools the data."""
|
||||||
|
|
||||||
ico: dict[str, Any]
|
ico: dict[str, Any]
|
||||||
pool: dict[str, Any]
|
pool: dict[str, Any]
|
||||||
|
measures_coordinator: OndiloIcoMeasuresCoordinator = field(init=False)
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass
|
||||||
|
class OndiloIcoMeasurementData:
|
||||||
|
"""Store the measurement data for one pool."""
|
||||||
|
|
||||||
sensors: dict[str, Any]
|
sensors: dict[str, Any]
|
||||||
|
|
||||||
|
|
||||||
class OndiloIcoCoordinator(DataUpdateCoordinator[dict[str, OndiloIcoData]]):
|
class OndiloIcoPoolsCoordinator(DataUpdateCoordinator[dict[str, OndiloIcoPoolData]]):
|
||||||
"""Class to manage fetching Ondilo ICO data from API."""
|
"""Fetch Ondilo ICO pools data from API."""
|
||||||
|
|
||||||
config_entry: ConfigEntry
|
config_entry: ConfigEntry
|
||||||
|
|
||||||
@ -39,45 +54,138 @@ class OndiloIcoCoordinator(DataUpdateCoordinator[dict[str, OndiloIcoData]]):
|
|||||||
hass,
|
hass,
|
||||||
logger=_LOGGER,
|
logger=_LOGGER,
|
||||||
config_entry=config_entry,
|
config_entry=config_entry,
|
||||||
name=DOMAIN,
|
name=f"{DOMAIN}_pools",
|
||||||
update_interval=timedelta(hours=1),
|
update_interval=timedelta(minutes=20),
|
||||||
)
|
)
|
||||||
self.api = api
|
self.api = api
|
||||||
|
self.config_entry = config_entry
|
||||||
|
self._device_registry = dr.async_get(self.hass)
|
||||||
|
|
||||||
async def _async_update_data(self) -> dict[str, OndiloIcoData]:
|
async def _async_update_data(self) -> dict[str, OndiloIcoPoolData]:
|
||||||
"""Fetch data from API endpoint."""
|
"""Fetch pools data from API endpoint and update devices."""
|
||||||
|
known_pools: set[str] = set(self.data) if self.data else set()
|
||||||
try:
|
try:
|
||||||
return await self.hass.async_add_executor_job(self._update_data)
|
async with UPDATE_LOCK:
|
||||||
|
data = await self.hass.async_add_executor_job(self._update_data)
|
||||||
except OndiloError as err:
|
except OndiloError as err:
|
||||||
raise UpdateFailed(f"Error communicating with API: {err}") from err
|
raise UpdateFailed(f"Error communicating with API: {err}") from err
|
||||||
|
|
||||||
def _update_data(self) -> dict[str, OndiloIcoData]:
|
current_pools = set(data)
|
||||||
"""Fetch data from API endpoint."""
|
|
||||||
|
new_pools = current_pools - known_pools
|
||||||
|
for pool_id in new_pools:
|
||||||
|
pool_data = data[pool_id]
|
||||||
|
pool_data.measures_coordinator = OndiloIcoMeasuresCoordinator(
|
||||||
|
self.hass, self.config_entry, self.api, pool_id
|
||||||
|
)
|
||||||
|
self._device_registry.async_get_or_create(
|
||||||
|
config_entry_id=self.config_entry.entry_id,
|
||||||
|
identifiers={(DOMAIN, pool_data.ico["serial_number"])},
|
||||||
|
manufacturer="Ondilo",
|
||||||
|
model="ICO",
|
||||||
|
name=pool_data.pool["name"],
|
||||||
|
sw_version=pool_data.ico["sw_version"],
|
||||||
|
)
|
||||||
|
|
||||||
|
removed_pools = known_pools - current_pools
|
||||||
|
for pool_id in removed_pools:
|
||||||
|
pool_data = self.data.pop(pool_id)
|
||||||
|
await pool_data.measures_coordinator.async_shutdown()
|
||||||
|
device_entry = self._device_registry.async_get_device(
|
||||||
|
identifiers={(DOMAIN, pool_data.ico["serial_number"])}
|
||||||
|
)
|
||||||
|
if device_entry:
|
||||||
|
self._device_registry.async_update_device(
|
||||||
|
device_id=device_entry.id,
|
||||||
|
remove_config_entry_id=self.config_entry.entry_id,
|
||||||
|
)
|
||||||
|
|
||||||
|
for pool_id in current_pools:
|
||||||
|
pool_data = data[pool_id]
|
||||||
|
measures_coordinator = pool_data.measures_coordinator
|
||||||
|
measures_coordinator.set_next_refresh(pool_data)
|
||||||
|
if not measures_coordinator.data:
|
||||||
|
await measures_coordinator.async_refresh()
|
||||||
|
|
||||||
|
return data
|
||||||
|
|
||||||
|
def _update_data(self) -> dict[str, OndiloIcoPoolData]:
|
||||||
|
"""Fetch pools data from API endpoint."""
|
||||||
res = {}
|
res = {}
|
||||||
pools = self.api.get_pools()
|
pools = self.api.get_pools()
|
||||||
_LOGGER.debug("Pools: %s", pools)
|
_LOGGER.debug("Pools: %s", pools)
|
||||||
error: OndiloError | None = None
|
error: OndiloError | None = None
|
||||||
for pool in pools:
|
for pool in pools:
|
||||||
pool_id = pool["id"]
|
pool_id = pool["id"]
|
||||||
|
if (data := self.data) and pool_id in data:
|
||||||
|
pool_data = res[pool_id] = data[pool_id]
|
||||||
|
pool_data.pool = pool
|
||||||
|
# Skip requesting new ICO data for known pools
|
||||||
|
# to avoid unnecessary API calls.
|
||||||
|
continue
|
||||||
try:
|
try:
|
||||||
ico = self.api.get_ICO_details(pool_id)
|
ico = self.api.get_ICO_details(pool_id)
|
||||||
if not ico:
|
|
||||||
_LOGGER.debug(
|
|
||||||
"The pool id %s does not have any ICO attached", pool_id
|
|
||||||
)
|
|
||||||
continue
|
|
||||||
sensors = self.api.get_last_pool_measures(pool_id)
|
|
||||||
except OndiloError as err:
|
except OndiloError as err:
|
||||||
error = err
|
error = err
|
||||||
_LOGGER.debug("Error communicating with API for %s: %s", pool_id, err)
|
_LOGGER.debug("Error communicating with API for %s: %s", pool_id, err)
|
||||||
continue
|
continue
|
||||||
res[pool_id] = OndiloIcoData(
|
|
||||||
ico=ico,
|
if not ico:
|
||||||
pool=pool,
|
_LOGGER.debug("The pool id %s does not have any ICO attached", pool_id)
|
||||||
sensors={sensor["data_type"]: sensor["value"] for sensor in sensors},
|
continue
|
||||||
)
|
|
||||||
|
res[pool_id] = OndiloIcoPoolData(ico=ico, pool=pool)
|
||||||
if not res:
|
if not res:
|
||||||
if error:
|
if error:
|
||||||
raise UpdateFailed(f"Error communicating with API: {error}") from error
|
raise UpdateFailed(f"Error communicating with API: {error}") from error
|
||||||
raise UpdateFailed("No data available")
|
|
||||||
return res
|
return res
|
||||||
|
|
||||||
|
|
||||||
|
class OndiloIcoMeasuresCoordinator(DataUpdateCoordinator[OndiloIcoMeasurementData]):
|
||||||
|
"""Fetch Ondilo ICO measurement data for one pool from API."""
|
||||||
|
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
hass: HomeAssistant,
|
||||||
|
config_entry: ConfigEntry,
|
||||||
|
api: OndiloClient,
|
||||||
|
pool_id: str,
|
||||||
|
) -> None:
|
||||||
|
"""Initialize."""
|
||||||
|
super().__init__(
|
||||||
|
hass,
|
||||||
|
config_entry=config_entry,
|
||||||
|
logger=_LOGGER,
|
||||||
|
name=f"{DOMAIN}_measures_{pool_id}",
|
||||||
|
)
|
||||||
|
self.api = api
|
||||||
|
self._next_refresh: datetime | None = None
|
||||||
|
self._pool_id = pool_id
|
||||||
|
|
||||||
|
async def _async_update_data(self) -> OndiloIcoMeasurementData:
|
||||||
|
"""Fetch measurement data from API endpoint."""
|
||||||
|
async with UPDATE_LOCK:
|
||||||
|
data = await self.hass.async_add_executor_job(self._update_data)
|
||||||
|
if next_refresh := self._next_refresh:
|
||||||
|
now = dt_util.utcnow()
|
||||||
|
# If we've missed the next refresh, schedule a refresh in one hour.
|
||||||
|
if next_refresh <= now:
|
||||||
|
next_refresh = now + timedelta(hours=1)
|
||||||
|
self.update_interval = next_refresh - now
|
||||||
|
|
||||||
|
return data
|
||||||
|
|
||||||
|
def _update_data(self) -> OndiloIcoMeasurementData:
|
||||||
|
"""Fetch measurement data from API endpoint."""
|
||||||
|
try:
|
||||||
|
sensors = self.api.get_last_pool_measures(self._pool_id)
|
||||||
|
except OndiloError as err:
|
||||||
|
raise UpdateFailed(f"Error communicating with API: {err}") from err
|
||||||
|
return OndiloIcoMeasurementData(
|
||||||
|
sensors={sensor["data_type"]: sensor["value"] for sensor in sensors},
|
||||||
|
)
|
||||||
|
|
||||||
|
def set_next_refresh(self, pool_data: OndiloIcoPoolData) -> None:
|
||||||
|
"""Set next refresh of this coordinator."""
|
||||||
|
last_update = datetime.fromisoformat(pool_data.pool["updated_at"])
|
||||||
|
self._next_refresh = last_update + TIME_TO_NEXT_UPDATE
|
||||||
|
@ -15,14 +15,18 @@ from homeassistant.const import (
|
|||||||
UnitOfElectricPotential,
|
UnitOfElectricPotential,
|
||||||
UnitOfTemperature,
|
UnitOfTemperature,
|
||||||
)
|
)
|
||||||
from homeassistant.core import HomeAssistant
|
from homeassistant.core import HomeAssistant, callback
|
||||||
from homeassistant.helpers.device_registry import DeviceInfo
|
from homeassistant.helpers.device_registry import DeviceInfo
|
||||||
from homeassistant.helpers.entity_platform import AddEntitiesCallback
|
from homeassistant.helpers.entity_platform import AddEntitiesCallback
|
||||||
from homeassistant.helpers.typing import StateType
|
from homeassistant.helpers.typing import StateType
|
||||||
from homeassistant.helpers.update_coordinator import CoordinatorEntity
|
from homeassistant.helpers.update_coordinator import CoordinatorEntity
|
||||||
|
|
||||||
from .const import DOMAIN
|
from .const import DOMAIN
|
||||||
from .coordinator import OndiloIcoCoordinator, OndiloIcoData
|
from .coordinator import (
|
||||||
|
OndiloIcoMeasuresCoordinator,
|
||||||
|
OndiloIcoPoolData,
|
||||||
|
OndiloIcoPoolsCoordinator,
|
||||||
|
)
|
||||||
|
|
||||||
SENSOR_TYPES: tuple[SensorEntityDescription, ...] = (
|
SENSOR_TYPES: tuple[SensorEntityDescription, ...] = (
|
||||||
SensorEntityDescription(
|
SensorEntityDescription(
|
||||||
@ -73,50 +77,67 @@ async def async_setup_entry(
|
|||||||
hass: HomeAssistant, entry: ConfigEntry, async_add_entities: AddEntitiesCallback
|
hass: HomeAssistant, entry: ConfigEntry, async_add_entities: AddEntitiesCallback
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Set up the Ondilo ICO sensors."""
|
"""Set up the Ondilo ICO sensors."""
|
||||||
|
pools_coordinator: OndiloIcoPoolsCoordinator = hass.data[DOMAIN][entry.entry_id]
|
||||||
|
known_entities: set[str] = set()
|
||||||
|
|
||||||
coordinator: OndiloIcoCoordinator = hass.data[DOMAIN][entry.entry_id]
|
async_add_entities(get_new_entities(pools_coordinator, known_entities))
|
||||||
|
|
||||||
async_add_entities(
|
@callback
|
||||||
OndiloICO(coordinator, pool_id, description)
|
def add_new_entities():
|
||||||
for pool_id, pool in coordinator.data.items()
|
"""Add any new entities after update of the pools coordinator."""
|
||||||
for description in SENSOR_TYPES
|
async_add_entities(get_new_entities(pools_coordinator, known_entities))
|
||||||
if description.key in pool.sensors
|
|
||||||
)
|
entry.async_on_unload(pools_coordinator.async_add_listener(add_new_entities))
|
||||||
|
|
||||||
|
|
||||||
class OndiloICO(CoordinatorEntity[OndiloIcoCoordinator], SensorEntity):
|
@callback
|
||||||
|
def get_new_entities(
|
||||||
|
pools_coordinator: OndiloIcoPoolsCoordinator,
|
||||||
|
known_entities: set[str],
|
||||||
|
) -> list[OndiloICO]:
|
||||||
|
"""Return new Ondilo ICO sensor entities."""
|
||||||
|
entities = []
|
||||||
|
for pool_id, pool_data in pools_coordinator.data.items():
|
||||||
|
for description in SENSOR_TYPES:
|
||||||
|
measurement_id = f"{pool_id}-{description.key}"
|
||||||
|
if (
|
||||||
|
measurement_id in known_entities
|
||||||
|
or (data := pool_data.measures_coordinator.data) is None
|
||||||
|
or description.key not in data.sensors
|
||||||
|
):
|
||||||
|
continue
|
||||||
|
known_entities.add(measurement_id)
|
||||||
|
entities.append(
|
||||||
|
OndiloICO(
|
||||||
|
pool_data.measures_coordinator, description, pool_id, pool_data
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
return entities
|
||||||
|
|
||||||
|
|
||||||
|
class OndiloICO(CoordinatorEntity[OndiloIcoMeasuresCoordinator], SensorEntity):
|
||||||
"""Representation of a Sensor."""
|
"""Representation of a Sensor."""
|
||||||
|
|
||||||
_attr_has_entity_name = True
|
_attr_has_entity_name = True
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
coordinator: OndiloIcoCoordinator,
|
coordinator: OndiloIcoMeasuresCoordinator,
|
||||||
pool_id: str,
|
|
||||||
description: SensorEntityDescription,
|
description: SensorEntityDescription,
|
||||||
|
pool_id: str,
|
||||||
|
pool_data: OndiloIcoPoolData,
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Initialize sensor entity with data from coordinator."""
|
"""Initialize sensor entity with data from coordinator."""
|
||||||
super().__init__(coordinator)
|
super().__init__(coordinator)
|
||||||
self.entity_description = description
|
self.entity_description = description
|
||||||
|
|
||||||
self._pool_id = pool_id
|
self._pool_id = pool_id
|
||||||
|
self._attr_unique_id = f"{pool_data.ico['serial_number']}-{description.key}"
|
||||||
data = self.pool_data
|
|
||||||
self._attr_unique_id = f"{data.ico['serial_number']}-{description.key}"
|
|
||||||
self._attr_device_info = DeviceInfo(
|
self._attr_device_info = DeviceInfo(
|
||||||
identifiers={(DOMAIN, data.ico["serial_number"])},
|
identifiers={(DOMAIN, pool_data.ico["serial_number"])},
|
||||||
manufacturer="Ondilo",
|
|
||||||
model="ICO",
|
|
||||||
name=data.pool["name"],
|
|
||||||
sw_version=data.ico["sw_version"],
|
|
||||||
)
|
)
|
||||||
|
|
||||||
@property
|
|
||||||
def pool_data(self) -> OndiloIcoData:
|
|
||||||
"""Get pool data."""
|
|
||||||
return self.coordinator.data[self._pool_id]
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def native_value(self) -> StateType:
|
def native_value(self) -> StateType:
|
||||||
"""Last value of the sensor."""
|
"""Last value of the sensor."""
|
||||||
return self.pool_data.sensors[self.entity_description.key]
|
return self.coordinator.data.sensors[self.entity_description.key]
|
||||||
|
@ -15,5 +15,5 @@
|
|||||||
"latitude": 48.861783,
|
"latitude": 48.861783,
|
||||||
"longitude": 2.337421
|
"longitude": 2.337421
|
||||||
},
|
},
|
||||||
"updated_at": "2024-01-01T01:00:00+0000"
|
"updated_at": "2024-01-01T01:05:00+0000"
|
||||||
}
|
}
|
||||||
|
@ -1,8 +1,10 @@
|
|||||||
"""Test Ondilo ICO initialization."""
|
"""Test Ondilo ICO initialization."""
|
||||||
|
|
||||||
|
from datetime import datetime, timedelta
|
||||||
from typing import Any
|
from typing import Any
|
||||||
from unittest.mock import MagicMock
|
from unittest.mock import MagicMock
|
||||||
|
|
||||||
|
from freezegun.api import FrozenDateTimeFactory
|
||||||
from ondilo import OndiloError
|
from ondilo import OndiloError
|
||||||
import pytest
|
import pytest
|
||||||
from syrupy import SnapshotAssertion
|
from syrupy import SnapshotAssertion
|
||||||
@ -13,7 +15,7 @@ from homeassistant.helpers import device_registry as dr
|
|||||||
|
|
||||||
from . import setup_integration
|
from . import setup_integration
|
||||||
|
|
||||||
from tests.common import MockConfigEntry
|
from tests.common import MockConfigEntry, async_fire_time_changed
|
||||||
|
|
||||||
|
|
||||||
async def test_devices(
|
async def test_devices(
|
||||||
@ -63,6 +65,7 @@ async def test_get_pools_error(
|
|||||||
async def test_init_with_no_ico_attached(
|
async def test_init_with_no_ico_attached(
|
||||||
hass: HomeAssistant,
|
hass: HomeAssistant,
|
||||||
mock_ondilo_client: MagicMock,
|
mock_ondilo_client: MagicMock,
|
||||||
|
device_registry: dr.DeviceRegistry,
|
||||||
config_entry: MockConfigEntry,
|
config_entry: MockConfigEntry,
|
||||||
pool1: dict[str, Any],
|
pool1: dict[str, Any],
|
||||||
) -> None:
|
) -> None:
|
||||||
@ -73,14 +76,104 @@ async def test_init_with_no_ico_attached(
|
|||||||
mock_ondilo_client.get_ICO_details.return_value = None
|
mock_ondilo_client.get_ICO_details.return_value = None
|
||||||
await setup_integration(hass, config_entry, mock_ondilo_client)
|
await setup_integration(hass, config_entry, mock_ondilo_client)
|
||||||
|
|
||||||
|
device_entries = dr.async_entries_for_config_entry(
|
||||||
|
device_registry, config_entry.entry_id
|
||||||
|
)
|
||||||
|
# No devices should be created
|
||||||
|
assert len(device_entries) == 0
|
||||||
# No sensor should be created
|
# No sensor should be created
|
||||||
assert len(hass.states.async_all()) == 0
|
assert len(hass.states.async_all()) == 0
|
||||||
# We should not have tried to retrieve pool measures
|
# We should not have tried to retrieve pool measures
|
||||||
mock_ondilo_client.get_last_pool_measures.assert_not_called()
|
mock_ondilo_client.get_last_pool_measures.assert_not_called()
|
||||||
assert config_entry.state is ConfigEntryState.SETUP_RETRY
|
assert config_entry.state is ConfigEntryState.LOADED
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize("api", ["get_ICO_details", "get_last_pool_measures"])
|
async def test_adding_pool_after_setup(
|
||||||
|
hass: HomeAssistant,
|
||||||
|
freezer: FrozenDateTimeFactory,
|
||||||
|
mock_ondilo_client: MagicMock,
|
||||||
|
device_registry: dr.DeviceRegistry,
|
||||||
|
config_entry: MockConfigEntry,
|
||||||
|
pool1: dict[str, Any],
|
||||||
|
two_pools: list[dict[str, Any]],
|
||||||
|
ico_details1: dict[str, Any],
|
||||||
|
ico_details2: dict[str, Any],
|
||||||
|
) -> None:
|
||||||
|
"""Test adding one pool after integration setup."""
|
||||||
|
mock_ondilo_client.get_pools.return_value = pool1
|
||||||
|
mock_ondilo_client.get_ICO_details.return_value = ico_details1
|
||||||
|
|
||||||
|
await setup_integration(hass, config_entry, mock_ondilo_client)
|
||||||
|
|
||||||
|
device_entries = dr.async_entries_for_config_entry(
|
||||||
|
device_registry, config_entry.entry_id
|
||||||
|
)
|
||||||
|
|
||||||
|
# One pool is created with 7 entities.
|
||||||
|
assert len(device_entries) == 1
|
||||||
|
assert len(hass.states.async_all()) == 7
|
||||||
|
|
||||||
|
mock_ondilo_client.get_pools.return_value = two_pools
|
||||||
|
mock_ondilo_client.get_ICO_details.return_value = ico_details2
|
||||||
|
|
||||||
|
# Trigger a refresh of the pools coordinator.
|
||||||
|
freezer.tick(timedelta(minutes=20))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
device_entries = dr.async_entries_for_config_entry(
|
||||||
|
device_registry, config_entry.entry_id
|
||||||
|
)
|
||||||
|
|
||||||
|
# Two pool have been created with 7 entities each.
|
||||||
|
assert len(device_entries) == 2
|
||||||
|
assert len(hass.states.async_all()) == 14
|
||||||
|
|
||||||
|
|
||||||
|
async def test_removing_pool_after_setup(
|
||||||
|
hass: HomeAssistant,
|
||||||
|
freezer: FrozenDateTimeFactory,
|
||||||
|
mock_ondilo_client: MagicMock,
|
||||||
|
device_registry: dr.DeviceRegistry,
|
||||||
|
config_entry: MockConfigEntry,
|
||||||
|
pool1: dict[str, Any],
|
||||||
|
ico_details1: dict[str, Any],
|
||||||
|
) -> None:
|
||||||
|
"""Test removing one pool after integration setup."""
|
||||||
|
await setup_integration(hass, config_entry, mock_ondilo_client)
|
||||||
|
|
||||||
|
device_entries = dr.async_entries_for_config_entry(
|
||||||
|
device_registry, config_entry.entry_id
|
||||||
|
)
|
||||||
|
|
||||||
|
# Two pools are created with 7 entities each.
|
||||||
|
assert len(device_entries) == 2
|
||||||
|
assert len(hass.states.async_all()) == 14
|
||||||
|
|
||||||
|
mock_ondilo_client.get_pools.return_value = pool1
|
||||||
|
mock_ondilo_client.get_ICO_details.return_value = ico_details1
|
||||||
|
|
||||||
|
# Trigger a refresh of the pools coordinator.
|
||||||
|
freezer.tick(timedelta(minutes=20))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
device_entries = dr.async_entries_for_config_entry(
|
||||||
|
device_registry, config_entry.entry_id
|
||||||
|
)
|
||||||
|
|
||||||
|
# One pool is left with 7 entities.
|
||||||
|
assert len(device_entries) == 1
|
||||||
|
assert len(hass.states.async_all()) == 7
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.parametrize(
|
||||||
|
("api", "devices", "config_entry_state"),
|
||||||
|
[
|
||||||
|
("get_ICO_details", 0, ConfigEntryState.SETUP_RETRY),
|
||||||
|
("get_last_pool_measures", 1, ConfigEntryState.LOADED),
|
||||||
|
],
|
||||||
|
)
|
||||||
async def test_details_error_all_pools(
|
async def test_details_error_all_pools(
|
||||||
hass: HomeAssistant,
|
hass: HomeAssistant,
|
||||||
mock_ondilo_client: MagicMock,
|
mock_ondilo_client: MagicMock,
|
||||||
@ -88,6 +181,8 @@ async def test_details_error_all_pools(
|
|||||||
config_entry: MockConfigEntry,
|
config_entry: MockConfigEntry,
|
||||||
pool1: dict[str, Any],
|
pool1: dict[str, Any],
|
||||||
api: str,
|
api: str,
|
||||||
|
devices: int,
|
||||||
|
config_entry_state: ConfigEntryState,
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Test details and measures error for all pools."""
|
"""Test details and measures error for all pools."""
|
||||||
mock_ondilo_client.get_pools.return_value = pool1
|
mock_ondilo_client.get_pools.return_value = pool1
|
||||||
@ -100,8 +195,8 @@ async def test_details_error_all_pools(
|
|||||||
device_registry, config_entry.entry_id
|
device_registry, config_entry.entry_id
|
||||||
)
|
)
|
||||||
|
|
||||||
assert not device_entries
|
assert len(device_entries) == devices
|
||||||
assert config_entry.state is ConfigEntryState.SETUP_RETRY
|
assert config_entry.state is config_entry_state
|
||||||
|
|
||||||
|
|
||||||
async def test_details_error_one_pool(
|
async def test_details_error_one_pool(
|
||||||
@ -131,12 +226,15 @@ async def test_details_error_one_pool(
|
|||||||
|
|
||||||
async def test_measures_error_one_pool(
|
async def test_measures_error_one_pool(
|
||||||
hass: HomeAssistant,
|
hass: HomeAssistant,
|
||||||
|
freezer: FrozenDateTimeFactory,
|
||||||
mock_ondilo_client: MagicMock,
|
mock_ondilo_client: MagicMock,
|
||||||
device_registry: dr.DeviceRegistry,
|
device_registry: dr.DeviceRegistry,
|
||||||
config_entry: MockConfigEntry,
|
config_entry: MockConfigEntry,
|
||||||
last_measures: list[dict[str, Any]],
|
last_measures: list[dict[str, Any]],
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Test measures error for one pool and success for the other."""
|
"""Test measures error for one pool and success for the other."""
|
||||||
|
entity_id_1 = "sensor.pool_1_temperature"
|
||||||
|
entity_id_2 = "sensor.pool_2_temperature"
|
||||||
mock_ondilo_client.get_last_pool_measures.side_effect = [
|
mock_ondilo_client.get_last_pool_measures.side_effect = [
|
||||||
OndiloError(
|
OndiloError(
|
||||||
404,
|
404,
|
||||||
@ -151,4 +249,170 @@ async def test_measures_error_one_pool(
|
|||||||
device_registry, config_entry.entry_id
|
device_registry, config_entry.entry_id
|
||||||
)
|
)
|
||||||
|
|
||||||
assert len(device_entries) == 1
|
assert len(device_entries) == 2
|
||||||
|
# One pool returned an error, the other is ok.
|
||||||
|
# 7 entities are created for the second pool.
|
||||||
|
assert len(hass.states.async_all()) == 7
|
||||||
|
assert hass.states.get(entity_id_1) is None
|
||||||
|
assert hass.states.get(entity_id_2) is not None
|
||||||
|
|
||||||
|
# All pools now return measures.
|
||||||
|
mock_ondilo_client.get_last_pool_measures.side_effect = None
|
||||||
|
|
||||||
|
# Move time to next pools coordinator refresh.
|
||||||
|
freezer.tick(timedelta(minutes=20))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
device_entries = dr.async_entries_for_config_entry(
|
||||||
|
device_registry, config_entry.entry_id
|
||||||
|
)
|
||||||
|
|
||||||
|
assert len(device_entries) == 2
|
||||||
|
# 14 entities in total, 7 entities per pool.
|
||||||
|
assert len(hass.states.async_all()) == 14
|
||||||
|
assert hass.states.get(entity_id_1) is not None
|
||||||
|
assert hass.states.get(entity_id_2) is not None
|
||||||
|
|
||||||
|
|
||||||
|
async def test_measures_scheduling(
|
||||||
|
hass: HomeAssistant,
|
||||||
|
freezer: FrozenDateTimeFactory,
|
||||||
|
mock_ondilo_client: MagicMock,
|
||||||
|
device_registry: dr.DeviceRegistry,
|
||||||
|
config_entry: MockConfigEntry,
|
||||||
|
) -> None:
|
||||||
|
"""Test refresh scheduling of measures coordinator."""
|
||||||
|
# Move time to 10 min after pool 1 was updated and 5 min after pool 2 was updated.
|
||||||
|
freezer.move_to("2024-01-01T01:10:00+00:00")
|
||||||
|
entity_id_1 = "sensor.pool_1_temperature"
|
||||||
|
entity_id_2 = "sensor.pool_2_temperature"
|
||||||
|
await setup_integration(hass, config_entry, mock_ondilo_client)
|
||||||
|
|
||||||
|
device_entries = dr.async_entries_for_config_entry(
|
||||||
|
device_registry, config_entry.entry_id
|
||||||
|
)
|
||||||
|
|
||||||
|
# Two pools are created with 7 entities each.
|
||||||
|
assert len(device_entries) == 2
|
||||||
|
assert len(hass.states.async_all()) == 14
|
||||||
|
|
||||||
|
state = hass.states.get(entity_id_1)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T01:10:00+00:00")
|
||||||
|
state = hass.states.get(entity_id_2)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T01:10:00+00:00")
|
||||||
|
|
||||||
|
# Tick time by 20 min.
|
||||||
|
# The measures coordinators for both pools should not have been refreshed again.
|
||||||
|
freezer.tick(timedelta(minutes=20))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
state = hass.states.get(entity_id_1)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T01:10:00+00:00")
|
||||||
|
state = hass.states.get(entity_id_2)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T01:10:00+00:00")
|
||||||
|
|
||||||
|
# Move time to 65 min after pool 1 was last updated.
|
||||||
|
# This is 5 min after we expect pool 1 to be updated again.
|
||||||
|
# The measures coordinator for pool 1 should refresh at this time.
|
||||||
|
# The measures coordinator for pool 2 should not have been refreshed again.
|
||||||
|
# The pools coordinator has updated the last update time
|
||||||
|
# of the pools to a stale time that is already passed.
|
||||||
|
freezer.move_to("2024-01-01T02:05:00+00:00")
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
state = hass.states.get(entity_id_1)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T02:05:00+00:00")
|
||||||
|
state = hass.states.get(entity_id_2)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T01:10:00+00:00")
|
||||||
|
|
||||||
|
# Tick time by 5 min.
|
||||||
|
# The measures coordinator for pool 1 should not have been refreshed again.
|
||||||
|
# The measures coordinator for pool 2 should refresh at this time.
|
||||||
|
# The pools coordinator has updated the last update time
|
||||||
|
# of the pools to a stale time that is already passed.
|
||||||
|
freezer.tick(timedelta(minutes=5))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
state = hass.states.get(entity_id_1)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T02:05:00+00:00")
|
||||||
|
state = hass.states.get(entity_id_2)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T02:10:00+00:00")
|
||||||
|
|
||||||
|
# Tick time by 55 min.
|
||||||
|
# The measures coordinator for pool 1 should refresh at this time.
|
||||||
|
# This is 1 hour after the last refresh of the measures coordinator for pool 1.
|
||||||
|
freezer.tick(timedelta(minutes=55))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
state = hass.states.get(entity_id_1)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T03:05:00+00:00")
|
||||||
|
state = hass.states.get(entity_id_2)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T02:10:00+00:00")
|
||||||
|
|
||||||
|
# Tick time by 5 min.
|
||||||
|
# The measures coordinator for pool 2 should refresh at this time.
|
||||||
|
# This is 1 hour after the last refresh of the measures coordinator for pool 2.
|
||||||
|
freezer.tick(timedelta(minutes=5))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
state = hass.states.get(entity_id_1)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T03:05:00+00:00")
|
||||||
|
state = hass.states.get(entity_id_2)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T03:10:00+00:00")
|
||||||
|
|
||||||
|
# Set an error on the pools coordinator endpoint.
|
||||||
|
# This will cause the pools coordinator to not update the next refresh.
|
||||||
|
# This should cause the measures coordinators to keep the 1 hour cadence.
|
||||||
|
mock_ondilo_client.get_pools.side_effect = OndiloError(
|
||||||
|
502,
|
||||||
|
(
|
||||||
|
"<html> <head><title>502 Bad Gateway</title></head> "
|
||||||
|
"<body> <center><h1>502 Bad Gateway</h1></center> </body> </html>"
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
|
# Tick time by 55 min.
|
||||||
|
# The measures coordinator for pool 1 should refresh at this time.
|
||||||
|
# This is 1 hour after the last refresh of the measures coordinator for pool 1.
|
||||||
|
freezer.tick(timedelta(minutes=55))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
state = hass.states.get(entity_id_1)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T04:05:00+00:00")
|
||||||
|
state = hass.states.get(entity_id_2)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T03:10:00+00:00")
|
||||||
|
|
||||||
|
# Tick time by 5 min.
|
||||||
|
# The measures coordinator for pool 2 should refresh at this time.
|
||||||
|
# This is 1 hour after the last refresh of the measures coordinator for pool 2.
|
||||||
|
freezer.tick(timedelta(minutes=5))
|
||||||
|
async_fire_time_changed(hass)
|
||||||
|
await hass.async_block_till_done(wait_background_tasks=True)
|
||||||
|
|
||||||
|
state = hass.states.get(entity_id_1)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T04:05:00+00:00")
|
||||||
|
state = hass.states.get(entity_id_2)
|
||||||
|
assert state is not None
|
||||||
|
assert state.last_reported == datetime.fromisoformat("2024-01-01T04:10:00+00:00")
|
||||||
|
Loading…
x
Reference in New Issue
Block a user