mirror of
https://github.com/xiaochao99/fn_nas
synced 2025-12-24 22:07:10 +00:00
Compare commits
12 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
b6fabe3d68 | ||
|
|
035a7e7413 | ||
|
|
41d4e4cc0d | ||
|
|
fcb46f429c | ||
|
|
4ae0b74e78 | ||
|
|
fd079c4ddf | ||
|
|
9e799c8948 | ||
|
|
31ffc24e6a | ||
|
|
905bbf9c96 | ||
|
|
0f691e956f | ||
|
|
25348fff9b | ||
|
|
30b1b7d271 |
@@ -33,7 +33,7 @@
|
||||
1. 进入**HACS商店**
|
||||
2. 添加自定义存储库:
|
||||
```shell
|
||||
https://github.com/anxms/fn_nas
|
||||
https://github.com/xiaochao99/fn_nas
|
||||
```
|
||||
3. 搜索"飞牛NAS",点击下载
|
||||
4. **重启Home Assistant服务**
|
||||
|
||||
72
custom_components/fn_nas/binary_sensor.py
Normal file
72
custom_components/fn_nas/binary_sensor.py
Normal file
@@ -0,0 +1,72 @@
|
||||
import logging
|
||||
from homeassistant.components.binary_sensor import BinarySensorEntity, BinarySensorDeviceClass
|
||||
from homeassistant.helpers.update_coordinator import CoordinatorEntity
|
||||
from .const import (
|
||||
DOMAIN, HDD_HEALTH, DEVICE_ID_NAS, DATA_UPDATE_COORDINATOR
|
||||
)
|
||||
|
||||
_LOGGER = logging.getLogger(__name__)
|
||||
|
||||
async def async_setup_entry(hass, config_entry, async_add_entities):
|
||||
domain_data = hass.data[DOMAIN][config_entry.entry_id]
|
||||
coordinator = domain_data[DATA_UPDATE_COORDINATOR]
|
||||
|
||||
entities = []
|
||||
existing_ids = set()
|
||||
|
||||
# 添加硬盘健康状态二元传感器
|
||||
for disk in coordinator.data.get("disks", []):
|
||||
health_uid = f"{config_entry.entry_id}_{disk['device']}_health_binary"
|
||||
if health_uid not in existing_ids:
|
||||
entities.append(
|
||||
DiskHealthBinarySensor(
|
||||
coordinator,
|
||||
disk["device"],
|
||||
f"硬盘 {disk.get('model', '未知')} 健康状态",
|
||||
health_uid,
|
||||
disk
|
||||
)
|
||||
)
|
||||
existing_ids.add(health_uid)
|
||||
|
||||
async_add_entities(entities)
|
||||
|
||||
|
||||
class DiskHealthBinarySensor(CoordinatorEntity, BinarySensorEntity):
|
||||
def __init__(self, coordinator, device_id, name, unique_id, disk_info):
|
||||
super().__init__(coordinator)
|
||||
self.device_id = device_id
|
||||
self._attr_name = name
|
||||
self._attr_unique_id = unique_id
|
||||
self.disk_info = disk_info
|
||||
self._attr_device_info = {
|
||||
"identifiers": {(DOMAIN, f"disk_{device_id}")},
|
||||
"name": disk_info.get("model", "未知硬盘"),
|
||||
"manufacturer": "硬盘设备",
|
||||
"via_device": (DOMAIN, DEVICE_ID_NAS)
|
||||
}
|
||||
self._attr_device_class = BinarySensorDeviceClass.PROBLEM
|
||||
|
||||
@property
|
||||
def is_on(self):
|
||||
"""返回True表示有问题,False表示正常"""
|
||||
for disk in self.coordinator.data.get("disks", []):
|
||||
if disk["device"] == self.device_id:
|
||||
health = disk.get("health", "未知")
|
||||
# 将健康状态映射为二元状态
|
||||
if health in ["正常", "良好", "OK", "ok", "good", "Good"]:
|
||||
return False # 正常状态
|
||||
elif health in ["警告", "异常", "错误", "warning", "Warning", "error", "Error", "bad", "Bad"]:
|
||||
return True # 有问题状态
|
||||
else:
|
||||
# 未知状态也视为有问题
|
||||
return True
|
||||
return True # 默认视为有问题
|
||||
|
||||
@property
|
||||
def icon(self):
|
||||
"""根据状态返回图标"""
|
||||
if self.is_on:
|
||||
return "mdi:alert-circle" # 有问题时显示警告图标
|
||||
else:
|
||||
return "mdi:check-circle" # 正常时显示对勾图标
|
||||
@@ -3,7 +3,7 @@ from homeassistant.components.button import ButtonEntity
|
||||
from homeassistant.helpers.entity import EntityCategory
|
||||
from homeassistant.helpers.update_coordinator import CoordinatorEntity
|
||||
from .const import (
|
||||
DOMAIN, DATA_UPDATE_COORDINATOR, DEVICE_ID_NAS, CONF_ENABLE_DOCKER
|
||||
DOMAIN, DATA_UPDATE_COORDINATOR, DEVICE_ID_NAS, CONF_ENABLE_DOCKER, DEVICE_ID_ZFS
|
||||
)
|
||||
|
||||
_LOGGER = logging.getLogger(__name__)
|
||||
@@ -18,7 +18,7 @@ async def async_setup_entry(hass, config_entry, async_add_entities):
|
||||
# 1. 添加NAS重启按钮
|
||||
entities.append(RebootButton(coordinator, config_entry.entry_id))
|
||||
|
||||
# 2. 添加虚拟机重启按钮
|
||||
# 2. 添加虚拟机重启按钮和强制关机按钮
|
||||
if "vms" in coordinator.data:
|
||||
for vm in coordinator.data["vms"]:
|
||||
entities.append(
|
||||
@@ -29,6 +29,14 @@ async def async_setup_entry(hass, config_entry, async_add_entities):
|
||||
config_entry.entry_id
|
||||
)
|
||||
)
|
||||
entities.append(
|
||||
VMDestroyButton(
|
||||
coordinator,
|
||||
vm["name"],
|
||||
vm.get("title", vm["name"]),
|
||||
config_entry.entry_id
|
||||
)
|
||||
)
|
||||
|
||||
# 3. 添加Docker容器重启按钮(如果启用了Docker功能)
|
||||
if enable_docker and "docker_containers" in coordinator.data:
|
||||
@@ -44,6 +52,19 @@ async def async_setup_entry(hass, config_entry, async_add_entities):
|
||||
)
|
||||
)
|
||||
|
||||
# 4. 添加ZFS存储池scrub按钮
|
||||
if "zpools" in coordinator.data:
|
||||
for zpool in coordinator.data["zpools"]:
|
||||
safe_name = zpool["name"].replace(" ", "_").replace("/", "_").replace(".", "_")
|
||||
entities.append(
|
||||
ZpoolScrubButton(
|
||||
coordinator,
|
||||
zpool["name"],
|
||||
safe_name,
|
||||
config_entry.entry_id
|
||||
)
|
||||
)
|
||||
|
||||
async_add_entities(entities)
|
||||
|
||||
class RebootButton(CoordinatorEntity, ButtonEntity):
|
||||
@@ -162,4 +183,97 @@ class DockerContainerRestartButton(CoordinatorEntity, ButtonEntity):
|
||||
"容器名称": self.container_name,
|
||||
"操作类型": "重启容器",
|
||||
"提示": "重启操作可能需要一些时间完成"
|
||||
}
|
||||
|
||||
class VMDestroyButton(CoordinatorEntity, ButtonEntity):
|
||||
def __init__(self, coordinator, vm_name, vm_title, entry_id):
|
||||
super().__init__(coordinator)
|
||||
self.vm_name = vm_name
|
||||
self.vm_title = vm_title
|
||||
self._attr_name = f"{vm_title} 强制关机"
|
||||
self._attr_unique_id = f"{entry_id}_flynas_vm_{vm_name}_destroy"
|
||||
self._attr_device_info = {
|
||||
"identifiers": {(DOMAIN, f"vm_{vm_name}")},
|
||||
"name": vm_title,
|
||||
"via_device": (DOMAIN, DEVICE_ID_NAS)
|
||||
}
|
||||
self._attr_icon = "mdi:power-off" # 使用关机图标
|
||||
|
||||
self.vm_manager = coordinator.vm_manager if hasattr(coordinator, 'vm_manager') else None
|
||||
|
||||
async def async_press(self):
|
||||
"""强制关机虚拟机"""
|
||||
if not self.vm_manager:
|
||||
_LOGGER.error("vm_manager不可用,无法强制关机虚拟机 %s", self.vm_name)
|
||||
return
|
||||
|
||||
try:
|
||||
success = await self.vm_manager.control_vm(self.vm_name, "destroy")
|
||||
if success:
|
||||
# 更新状态为"强制关机中"
|
||||
for vm in self.coordinator.data["vms"]:
|
||||
if vm["name"] == self.vm_name:
|
||||
vm["state"] = "destroying"
|
||||
self.async_write_ha_state()
|
||||
|
||||
# 在下次更新时恢复实际状态
|
||||
self.coordinator.async_add_listener(self.async_write_ha_state)
|
||||
except Exception as e:
|
||||
_LOGGER.error("强制关机虚拟机时出错: %s", str(e), exc_info=True)
|
||||
|
||||
@property
|
||||
def extra_state_attributes(self):
|
||||
return {
|
||||
"虚拟机名称": self.vm_name,
|
||||
"操作类型": "强制关机",
|
||||
"警告": "此操作会强制关闭虚拟机,可能导致数据丢失",
|
||||
"提示": "仅在虚拟机无法正常关机时使用此功能"
|
||||
}
|
||||
|
||||
class ZpoolScrubButton(CoordinatorEntity, ButtonEntity):
|
||||
def __init__(self, coordinator, zpool_name, safe_name, entry_id):
|
||||
super().__init__(coordinator)
|
||||
self.zpool_name = zpool_name
|
||||
self.safe_name = safe_name
|
||||
self._attr_name = f"ZFS {zpool_name} 数据检查"
|
||||
self._attr_unique_id = f"{entry_id}_zpool_{safe_name}_scrub"
|
||||
self._attr_device_info = {
|
||||
"identifiers": {(DOMAIN, DEVICE_ID_ZFS)},
|
||||
"name": "ZFS存储池",
|
||||
"via_device": (DOMAIN, DEVICE_ID_NAS)
|
||||
}
|
||||
self._attr_icon = "mdi:harddisk-check"
|
||||
|
||||
@property
|
||||
def available(self):
|
||||
"""检查按钮是否可用(当scrub进行中时不可点击)"""
|
||||
scrub_status = self.coordinator.data.get("scrub_status", {}).get(self.zpool_name, {})
|
||||
return not scrub_status.get("scrub_in_progress", False)
|
||||
|
||||
async def async_press(self):
|
||||
"""执行ZFS存储池数据一致性检查"""
|
||||
try:
|
||||
# 检查是否已经有scrub在进行中
|
||||
scrub_status = self.coordinator.data.get("scrub_status", {}).get(self.zpool_name, {})
|
||||
if scrub_status.get("scrub_in_progress", False):
|
||||
self.coordinator.logger.warning(f"ZFS存储池 {self.zpool_name} 已在进行数据一致性检查")
|
||||
return
|
||||
|
||||
success = await self.coordinator.scrub_zpool(self.zpool_name)
|
||||
if success:
|
||||
self.coordinator.logger.info(f"ZFS存储池 {self.zpool_name} 数据一致性检查启动成功")
|
||||
# 立即刷新状态以更新按钮状态
|
||||
await self.coordinator.async_request_refresh()
|
||||
else:
|
||||
self.coordinator.logger.error(f"ZFS存储池 {self.zpool_name} 数据一致性检查启动失败")
|
||||
except Exception as e:
|
||||
self.coordinator.logger.error(f"启动ZFS存储池 {self.zpool_name} 数据一致性检查时出错: {str(e)}", exc_info=True)
|
||||
|
||||
@property
|
||||
def extra_state_attributes(self):
|
||||
return {
|
||||
"存储池名称": self.zpool_name,
|
||||
"操作类型": "数据一致性检查",
|
||||
"说明": "对ZFS存储池执行数据完整性和一致性验证",
|
||||
"提示": "此操作可能需要较长时间完成,建议在低峰期执行"
|
||||
}
|
||||
@@ -3,6 +3,7 @@ from homeassistant.const import Platform
|
||||
DOMAIN = "fn_nas"
|
||||
PLATFORMS = [
|
||||
Platform.SENSOR,
|
||||
Platform.BINARY_SENSOR,
|
||||
Platform.SWITCH,
|
||||
Platform.BUTTON
|
||||
]
|
||||
@@ -32,7 +33,8 @@ HDD_HEALTH = "health"
|
||||
HDD_STATUS = "status"
|
||||
SYSTEM_INFO = "system"
|
||||
FAN_SPEED = "fan_speed"
|
||||
UPS_INFO = "ups_info"
|
||||
UPS_INFO = "ups_info"
|
||||
ZFS_POOL = "zfs_pool"
|
||||
|
||||
ATTR_DISK_MODEL = "硬盘型号"
|
||||
ATTR_SERIAL_NO = "序列号"
|
||||
@@ -48,8 +50,28 @@ ICON_TEMPERATURE = "mdi:thermometer"
|
||||
ICON_HEALTH = "mdi:heart-pulse"
|
||||
ICON_POWER = "mdi:power"
|
||||
ICON_RESTART = "mdi:restart"
|
||||
ICON_ZFS = "mdi:harddisk-plus"
|
||||
|
||||
# 设备标识符常量
|
||||
DEVICE_ID_NAS = "flynas_nas_system"
|
||||
DEVICE_ID_UPS = "flynas_ups"
|
||||
CONF_NETWORK_MACS = "network_macs"
|
||||
DEVICE_ID_ZFS = "flynas_zfs"
|
||||
CONF_NETWORK_MACS = "network_macs"
|
||||
|
||||
# ZFS相关常量
|
||||
ATTR_ZPOOL_NAME = "存储池名称"
|
||||
ATTR_ZPOOL_HEALTH = "健康状态"
|
||||
ATTR_ZPOOL_SIZE = "总大小"
|
||||
ATTR_ZPOOL_ALLOC = "已使用"
|
||||
ATTR_ZPOOL_FREE = "可用空间"
|
||||
ATTR_ZPOOL_CAPACITY = "使用率"
|
||||
ATTR_ZPOOL_FRAGMENTATION = "碎片率"
|
||||
ATTR_ZPOOL_CKPOINT = "检查点"
|
||||
ATTR_ZPOOL_EXPANDSZ = "扩展大小"
|
||||
ATTR_ZPOOL_DEDUP = "重复数据删除率"
|
||||
ATTR_ZPOOL_SCRUB_STATUS = "检查状态"
|
||||
ATTR_ZPOOL_SCRUB_PROGRESS = "检查进度"
|
||||
ATTR_ZPOOL_SCRUB_SCAN_RATE = "扫描速度"
|
||||
ATTR_ZPOOL_SCRUB_TIME_REMAINING = "剩余时间"
|
||||
ATTR_ZPOOL_SCRUB_ISSUED = "已发出数据"
|
||||
ATTR_ZPOOL_SCRUB_REPAIRED = "已修复数据"
|
||||
@@ -79,7 +79,9 @@ class FlynasCoordinator(DataUpdateCoordinator):
|
||||
},
|
||||
"ups": {},
|
||||
"vms": [],
|
||||
"docker_containers": []
|
||||
"docker_containers": [],
|
||||
"zpools": [],
|
||||
"scrub_status": {}
|
||||
}
|
||||
|
||||
def _debug_log(self, message: str):
|
||||
@@ -313,11 +315,34 @@ class FlynasCoordinator(DataUpdateCoordinator):
|
||||
if connection_id is not None:
|
||||
await self.release_ssh_connection(connection_id)
|
||||
|
||||
async def ping_system(self) -> bool:
|
||||
"""轻量级系统状态检测"""
|
||||
# 对于本地主机直接返回True
|
||||
if self.host in ['localhost', '127.0.0.1']:
|
||||
return True
|
||||
|
||||
try:
|
||||
# 使用异步ping检测,减少超时时间
|
||||
proc = await asyncio.create_subprocess_exec(
|
||||
'ping', '-c', '1', '-W', '1', self.host,
|
||||
stdout=asyncio.subprocess.DEVNULL,
|
||||
stderr=asyncio.subprocess.DEVNULL
|
||||
)
|
||||
await asyncio.wait_for(proc.wait(), timeout=2) # 总超时时间2秒
|
||||
return proc.returncode == 0
|
||||
except Exception:
|
||||
return False
|
||||
|
||||
async def _monitor_system_status(self):
|
||||
"""系统离线时轮询检测状态"""
|
||||
self._debug_log(f"启动系统状态监控,每{self._retry_interval}秒检测一次")
|
||||
|
||||
# 使用指数退避策略,避免频繁检测
|
||||
check_interval = self._retry_interval
|
||||
max_interval = 300 # 最大5分钟检测一次
|
||||
|
||||
while True:
|
||||
await asyncio.sleep(self._retry_interval)
|
||||
await asyncio.sleep(check_interval)
|
||||
|
||||
if await self.ping_system():
|
||||
self._info_log("检测到系统已开机,触发重新加载")
|
||||
@@ -326,24 +351,10 @@ class FlynasCoordinator(DataUpdateCoordinator):
|
||||
self.hass.config_entries.async_reload(self.config_entry.entry_id)
|
||||
)
|
||||
break
|
||||
|
||||
async def ping_system(self) -> bool:
|
||||
"""轻量级系统状态检测"""
|
||||
# 对于本地主机直接返回True
|
||||
if self.host in ['localhost', '127.0.0.1']:
|
||||
return True
|
||||
|
||||
try:
|
||||
# 使用异步ping检测
|
||||
proc = await asyncio.create_subprocess_exec(
|
||||
'ping', '-c', '1', '-W', '1', self.host,
|
||||
stdout=asyncio.subprocess.DEVNULL,
|
||||
stderr=asyncio.subprocess.DEVNULL
|
||||
)
|
||||
await proc.wait()
|
||||
return proc.returncode == 0
|
||||
except Exception:
|
||||
return False
|
||||
else:
|
||||
# 系统仍然离线,增加检测间隔(指数退避)
|
||||
check_interval = min(check_interval * 1.5, max_interval)
|
||||
self._debug_log(f"系统仍离线,下次检测间隔: {check_interval}秒")
|
||||
|
||||
async def _async_update_data(self):
|
||||
"""数据更新入口,优化命令执行频率"""
|
||||
@@ -376,6 +387,18 @@ class FlynasCoordinator(DataUpdateCoordinator):
|
||||
disks = await self.disk_manager.get_disks_info()
|
||||
self._debug_log(f"磁盘信息获取完成,数量: {len(disks)}")
|
||||
|
||||
self._debug_log("开始获取ZFS存储池信息...")
|
||||
zpools = await self.disk_manager.get_zpools()
|
||||
self._debug_log(f"ZFS存储池信息获取完成,数量: {len(zpools)}")
|
||||
|
||||
# 获取所有ZFS存储池的scrub状态
|
||||
scrub_status = {}
|
||||
for zpool in zpools:
|
||||
self._debug_log(f"开始获取存储池 {zpool['name']} 的scrub状态...")
|
||||
scrub_info = await self.disk_manager.get_zpool_status(zpool['name'])
|
||||
scrub_status[zpool['name']] = scrub_info
|
||||
self._debug_log(f"存储池 {zpool['name']} scrub状态获取完成")
|
||||
|
||||
self._debug_log("开始获取UPS信息...")
|
||||
ups_info = await self.ups_manager.get_ups_info()
|
||||
self._debug_log("UPS信息获取完成")
|
||||
@@ -407,7 +430,9 @@ class FlynasCoordinator(DataUpdateCoordinator):
|
||||
"system": {**system, "status": status},
|
||||
"ups": ups_info,
|
||||
"vms": vms,
|
||||
"docker_containers": docker_containers
|
||||
"docker_containers": docker_containers,
|
||||
"zpools": zpools,
|
||||
"scrub_status": scrub_status
|
||||
}
|
||||
|
||||
self._debug_log(f"数据更新完成: disks={len(disks)}, vms={len(vms)}, containers={len(docker_containers)}")
|
||||
@@ -421,6 +446,32 @@ class FlynasCoordinator(DataUpdateCoordinator):
|
||||
|
||||
return self.get_default_data()
|
||||
|
||||
async def shutdown_system(self):
|
||||
"""关闭系统 - 委托给SystemManager"""
|
||||
return await self.system_manager.shutdown_system()
|
||||
|
||||
async def reboot_system(self):
|
||||
"""重启系统 - 委托给SystemManager"""
|
||||
return await self.system_manager.reboot_system()
|
||||
|
||||
async def scrub_zpool(self, pool_name: str) -> bool:
|
||||
"""执行ZFS存储池数据一致性检查"""
|
||||
try:
|
||||
self._debug_log(f"开始对ZFS存储池 {pool_name} 执行scrub操作")
|
||||
command = f"zpool scrub {pool_name}"
|
||||
result = await self.run_command(command)
|
||||
|
||||
if result and not result.lower().startswith("cannot"):
|
||||
self._debug_log(f"ZFS存储池 {pool_name} scrub操作启动成功")
|
||||
return True
|
||||
else:
|
||||
self.logger.error(f"ZFS存储池 {pool_name} scrub操作失败: {result}")
|
||||
return False
|
||||
|
||||
except Exception as e:
|
||||
self.logger.error(f"执行ZFS存储池 {pool_name} scrub操作时出错: {str(e)}", exc_info=True)
|
||||
return False
|
||||
|
||||
class UPSDataUpdateCoordinator(DataUpdateCoordinator):
|
||||
def __init__(self, hass: HomeAssistant, config, main_coordinator):
|
||||
self.config = config
|
||||
|
||||
@@ -14,6 +14,7 @@ class DiskManager:
|
||||
self.disk_full_info_cache = {} # 缓存磁盘完整信息
|
||||
self.first_run = True # 首次运行标志
|
||||
self.initial_detection_done = False # 首次完整检测完成标志
|
||||
self.disk_io_stats_cache = {} # 缓存磁盘I/O统计信息
|
||||
|
||||
def extract_value(self, text: str, patterns, default="未知", format_func=None):
|
||||
if not text:
|
||||
@@ -35,75 +36,247 @@ class DiskManager:
|
||||
self.logger.debug("No match found for patterns: %s", patterns)
|
||||
return default
|
||||
|
||||
async def check_disk_active(self, device: str, window: int = 30) -> bool:
|
||||
def _format_capacity(self, capacity_str: str) -> str:
|
||||
"""将容量字符串格式化为GB或TB格式"""
|
||||
if not capacity_str or capacity_str == "未知":
|
||||
return "未知"
|
||||
|
||||
try:
|
||||
# 处理逗号分隔的数字(如 "1,000,204,886,016 bytes")
|
||||
capacity_str = capacity_str.replace(',', '')
|
||||
|
||||
# 提取数字和单位
|
||||
import re
|
||||
# 匹配数字和单位(如 "500 GB", "1.0 TB", "1000204886016 bytes", "1,000,204,886,016 bytes")
|
||||
match = re.search(r'(\d+(?:\.\d+)?)\s*([KMGT]?B|bytes?)', capacity_str, re.IGNORECASE)
|
||||
if not match:
|
||||
# 如果没有匹配到单位,尝试直接提取数字
|
||||
numbers = re.findall(r'\d+', capacity_str)
|
||||
if numbers:
|
||||
# 取最大的数字(通常是容量值)
|
||||
value = float(max(numbers, key=len))
|
||||
bytes_value = value # 假设为字节
|
||||
else:
|
||||
return capacity_str
|
||||
else:
|
||||
value = float(match.group(1))
|
||||
unit = match.group(2).upper()
|
||||
|
||||
# 转换为字节
|
||||
if unit in ['B', 'BYTE', 'BYTES']:
|
||||
bytes_value = value
|
||||
elif unit in ['KB', 'KIB']:
|
||||
bytes_value = value * 1024
|
||||
elif unit in ['MB', 'MIB']:
|
||||
bytes_value = value * 1024 * 1024
|
||||
elif unit in ['GB', 'GIB']:
|
||||
bytes_value = value * 1024 * 1024 * 1024
|
||||
elif unit in ['TB', 'TIB']:
|
||||
bytes_value = value * 1024 * 1024 * 1024 * 1024
|
||||
else:
|
||||
bytes_value = value # 默认假设为字节
|
||||
|
||||
# 转换为合适的单位
|
||||
if bytes_value >= 1024**4: # 1 TB
|
||||
return f"{bytes_value / (1024**4):.1f} TB"
|
||||
elif bytes_value >= 1024**3: # 1 GB
|
||||
return f"{bytes_value / (1024**3):.1f} GB"
|
||||
elif bytes_value >= 1024**2: # 1 MB
|
||||
return f"{bytes_value / (1024**2):.1f} MB"
|
||||
elif bytes_value >= 1024: # 1 KB
|
||||
return f"{bytes_value / 1024:.1f} KB"
|
||||
else:
|
||||
return f"{bytes_value:.1f} B"
|
||||
|
||||
except Exception as e:
|
||||
self.logger.debug(f"格式化容量失败: {capacity_str}, 错误: {e}")
|
||||
return capacity_str
|
||||
|
||||
async def check_disk_active(self, device: str, window: int = 30, current_status: str = None) -> bool:
|
||||
"""检查硬盘在指定时间窗口内是否有活动"""
|
||||
try:
|
||||
# 正确的路径是 /sys/block/{device}/stat
|
||||
stat_path = f"/sys/block/{device}/stat"
|
||||
# 首先检查硬盘当前状态
|
||||
if current_status is None:
|
||||
current_status = await self.get_disk_activity(device)
|
||||
else:
|
||||
self.logger.debug(f"使用传入的状态: {device} = {current_status}")
|
||||
|
||||
# 读取统计文件
|
||||
stat_output = await self.coordinator.run_command(f"cat {stat_path} 2>/dev/null")
|
||||
if not stat_output:
|
||||
self.logger.debug(f"无法读取 {stat_path},默认返回活跃状态")
|
||||
return True
|
||||
|
||||
# 解析统计信息
|
||||
stats = stat_output.split()
|
||||
if len(stats) < 11:
|
||||
self.logger.debug(f"无效的统计信息格式:{stat_output}")
|
||||
return True
|
||||
|
||||
# 关键字段:当前正在进行的I/O操作数量(第9个字段,索引8)
|
||||
in_flight = int(stats[8])
|
||||
# 如果硬盘处于休眠状态,直接返回非活跃
|
||||
if current_status == "休眠中":
|
||||
self.logger.debug(f"硬盘 {device} 处于休眠状态,不执行详细检测")
|
||||
return False
|
||||
|
||||
# 如果当前有I/O操作,直接返回活跃状态
|
||||
if in_flight > 0:
|
||||
return True
|
||||
# 如果硬盘处于空闲状态,检查是否有近期活动
|
||||
if current_status == "空闲中":
|
||||
# 检查缓存的统计信息来判断近期活动
|
||||
stat_path = f"/sys/block/{device}/stat"
|
||||
stat_output = await self.coordinator.run_command(f"cat {stat_path} 2>/dev/null")
|
||||
|
||||
# 检查I/O操作时间(第10个字段,索引9) - io_ticks(单位毫秒)
|
||||
io_ticks = int(stats[9])
|
||||
|
||||
# 如果设备在窗口时间内有I/O活动,返回活跃状态
|
||||
if io_ticks > window * 1000:
|
||||
return True
|
||||
if stat_output:
|
||||
stats = stat_output.split()
|
||||
if len(stats) >= 11:
|
||||
try:
|
||||
current_read_ios = int(stats[0])
|
||||
current_write_ios = int(stats[4])
|
||||
current_io_ticks = int(stats[9])
|
||||
|
||||
cached_stats = self.disk_io_stats_cache.get(device)
|
||||
if cached_stats:
|
||||
read_diff = current_read_ios - cached_stats.get('read_ios', 0)
|
||||
write_diff = current_write_ios - cached_stats.get('write_ios', 0)
|
||||
io_ticks_diff = current_io_ticks - cached_stats.get('io_ticks', 0)
|
||||
|
||||
# 如果在最近30秒内有I/O活动,认为硬盘活跃
|
||||
if read_diff > 0 or write_diff > 0 or io_ticks_diff > 100:
|
||||
self.logger.debug(f"硬盘 {device} 近期有I/O活动,需要更新信息")
|
||||
return True
|
||||
|
||||
# 更新缓存
|
||||
self.disk_io_stats_cache[device] = {
|
||||
'read_ios': current_read_ios,
|
||||
'write_ios': current_write_ios,
|
||||
'io_ticks': current_io_ticks
|
||||
}
|
||||
|
||||
except (ValueError, IndexError):
|
||||
pass
|
||||
|
||||
# 所有检查都通过,返回非活跃状态
|
||||
return False
|
||||
# 如果硬盘空闲且没有近期活动,使用缓存信息
|
||||
self.logger.debug(f"硬盘 {device} 处于空闲状态且无近期活动,使用缓存信息")
|
||||
return False
|
||||
|
||||
# 如果硬盘处于活动中,返回活跃状态
|
||||
if current_status == "活动中":
|
||||
self.logger.debug(f"硬盘 {device} 处于活动中,执行详细检测")
|
||||
return True
|
||||
|
||||
# 默认情况下返回活跃状态
|
||||
self.logger.debug(f"硬盘 {device} 状态未知,默认执行详细检测")
|
||||
return True
|
||||
|
||||
except Exception as e:
|
||||
self.logger.error(f"检测硬盘活动状态失败: {str(e)}", exc_info=True)
|
||||
self.logger.error(f"检测硬盘活动状态失败: {str(e)}")
|
||||
return True # 出错时默认执行检测
|
||||
|
||||
async def get_disk_activity(self, device: str) -> str:
|
||||
"""获取硬盘活动状态(活动中/空闲中/休眠中)"""
|
||||
async def get_disk_power_state(self, device: str) -> str:
|
||||
"""获取硬盘电源状态"""
|
||||
try:
|
||||
# 检查硬盘是否处于休眠状态
|
||||
# 检查 SCSI 设备状态
|
||||
state_path = f"/sys/block/{device}/device/state"
|
||||
state_output = await self.coordinator.run_command(f"cat {state_path} 2>/dev/null || echo 'unknown'")
|
||||
state = state_output.strip().lower()
|
||||
|
||||
if state in ["standby", "sleep"]:
|
||||
if state in ["running", "active"]:
|
||||
return "active"
|
||||
elif state in ["standby", "sleep"]:
|
||||
return state
|
||||
|
||||
# 对于某些设备,尝试通过hdparm检查状态(非侵入性)
|
||||
hdparm_output = await self.coordinator.run_command(f"hdparm -C /dev/{device} 2>/dev/null || echo 'unknown'")
|
||||
if "standby" in hdparm_output.lower():
|
||||
return "standby"
|
||||
elif "sleeping" in hdparm_output.lower():
|
||||
return "sleep"
|
||||
elif "active/idle" in hdparm_output.lower():
|
||||
return "active"
|
||||
|
||||
return "unknown"
|
||||
|
||||
except Exception as e:
|
||||
self.logger.debug(f"获取磁盘 {device} 电源状态失败: {e}")
|
||||
return "unknown"
|
||||
|
||||
async def get_disk_activity(self, device: str) -> str:
|
||||
"""获取硬盘活动状态(活动中/空闲中/休眠中)"""
|
||||
try:
|
||||
# 先检查电源状态 - 这是最可靠的休眠检测方法
|
||||
power_state = await self.get_disk_power_state(device)
|
||||
if power_state in ["standby", "sleep"]:
|
||||
self.logger.debug(f"硬盘 {device} 电源状态为 {power_state},判定为休眠中")
|
||||
return "休眠中"
|
||||
|
||||
# 检查最近一分钟内的硬盘活动
|
||||
# 检查最近的I/O活动 - 使用非侵入性方式
|
||||
stat_path = f"/sys/block/{device}/stat"
|
||||
stat_output = await self.coordinator.run_command(f"cat {stat_path}")
|
||||
stats = stat_output.split()
|
||||
stat_output = await self.coordinator.run_command(f"cat {stat_path} 2>/dev/null")
|
||||
|
||||
if len(stats) >= 11:
|
||||
# 第9个字段是最近完成的读操作数
|
||||
# 第10个字段是最近完成的写操作数
|
||||
recent_reads = int(stats[8])
|
||||
recent_writes = int(stats[9])
|
||||
|
||||
if recent_reads > 0 or recent_writes > 0:
|
||||
if stat_output:
|
||||
stats = stat_output.split()
|
||||
if len(stats) >= 11:
|
||||
try:
|
||||
in_flight = int(stats[8]) # 当前进行中的I/O
|
||||
io_ticks = int(stats[9]) # I/O活动时间(ms)
|
||||
|
||||
# 如果有正在进行的I/O,返回活动中
|
||||
if in_flight > 0:
|
||||
self.logger.debug(f"硬盘 {device} 有进行中的I/O操作: {in_flight}")
|
||||
return "活动中"
|
||||
|
||||
# 检查缓存的统计信息来判断近期活动
|
||||
cached_stats = self.disk_io_stats_cache.get(device)
|
||||
if cached_stats:
|
||||
current_read_ios = int(stats[0])
|
||||
current_write_ios = int(stats[4])
|
||||
|
||||
read_diff = current_read_ios - cached_stats.get('read_ios', 0)
|
||||
write_diff = current_write_ios - cached_stats.get('write_ios', 0)
|
||||
io_ticks_diff = io_ticks - cached_stats.get('io_ticks', 0)
|
||||
|
||||
# 如果在最近30秒内有I/O活动,认为硬盘活动中
|
||||
if read_diff > 0 or write_diff > 0 or io_ticks_diff > 100: # 100ms内的活动
|
||||
self.logger.debug(f"硬盘 {device} 近期有I/O活动: 读={read_diff}, 写={write_diff}, 活动时间={io_ticks_diff}ms")
|
||||
|
||||
# 更新缓存统计信息
|
||||
self.disk_io_stats_cache[device] = {
|
||||
'read_ios': current_read_ios,
|
||||
'write_ios': current_write_ios,
|
||||
'in_flight': in_flight,
|
||||
'io_ticks': io_ticks
|
||||
}
|
||||
return "活动中"
|
||||
else:
|
||||
# 首次检测,保存当前状态并认为活跃
|
||||
self.logger.debug(f"硬盘 {device} 首次检测,保存统计信息")
|
||||
self.disk_io_stats_cache[device] = {
|
||||
'read_ios': int(stats[0]),
|
||||
'write_ios': int(stats[4]),
|
||||
'in_flight': in_flight,
|
||||
'io_ticks': io_ticks
|
||||
}
|
||||
return "活动中" # 首次检测默认返回活动中
|
||||
|
||||
# 更新缓存统计信息
|
||||
self.disk_io_stats_cache[device] = {
|
||||
'read_ios': int(stats[0]),
|
||||
'write_ios': int(stats[4]),
|
||||
'in_flight': in_flight,
|
||||
'io_ticks': io_ticks
|
||||
}
|
||||
|
||||
# 如果没有活动,返回空闲中
|
||||
self.logger.debug(f"硬盘 {device} 处于空闲状态")
|
||||
return "空闲中"
|
||||
|
||||
except (ValueError, IndexError) as e:
|
||||
self.logger.debug(f"解析硬盘 {device} 统计信息失败: {e}")
|
||||
return "活动中" # 出错时默认返回活动中,避免中断休眠
|
||||
|
||||
# 如果无法获取统计信息,检查硬盘是否可访问
|
||||
try:
|
||||
# 尝试读取设备信息,如果成功说明硬盘可访问
|
||||
test_output = await self.coordinator.run_command(f"ls -la /dev/{device} 2>/dev/null")
|
||||
if test_output and device in test_output:
|
||||
self.logger.debug(f"硬盘 {device} 可访问但无统计信息,默认返回活动中")
|
||||
return "活动中"
|
||||
|
||||
return "空闲中"
|
||||
else:
|
||||
self.logger.debug(f"硬盘 {device} 不可访问,可能处于休眠状态")
|
||||
return "休眠中"
|
||||
except:
|
||||
self.logger.debug(f"硬盘 {device} 检测失败,默认返回活动中")
|
||||
return "活动中"
|
||||
|
||||
except Exception as e:
|
||||
self.logger.error(f"获取硬盘 {device} 状态失败: {str(e)}", exc_info=True)
|
||||
return "未知"
|
||||
return "活动中" # 出错时默认返回活动中,避免中断休眠
|
||||
|
||||
async def get_disks_info(self) -> list[dict]:
|
||||
disks = []
|
||||
@@ -172,31 +345,31 @@ class DiskManager:
|
||||
disks.append(disk_info)
|
||||
continue
|
||||
|
||||
# 检查硬盘是否活跃
|
||||
is_active = await self.check_disk_active(device, window=30)
|
||||
# 检查硬盘是否活跃,传入当前状态确保一致性
|
||||
is_active = await self.check_disk_active(device, window=30, current_status=status)
|
||||
if not is_active:
|
||||
self.logger.debug(f"硬盘 {device} 处于非活跃状态,使用上一次获取的信息")
|
||||
|
||||
# 优先使用缓存的完整信息
|
||||
if cached_info:
|
||||
disk_info.update({
|
||||
"model": cached_info.get("model", "未检测"),
|
||||
"serial": cached_info.get("serial", "未检测"),
|
||||
"capacity": cached_info.get("capacity", "未检测"),
|
||||
"health": cached_info.get("health", "未检测"),
|
||||
"temperature": cached_info.get("temperature", "未检测"),
|
||||
"power_on_hours": cached_info.get("power_on_hours", "未检测"),
|
||||
"model": cached_info.get("model", "未知"),
|
||||
"serial": cached_info.get("serial", "未知"),
|
||||
"capacity": cached_info.get("capacity", "未知"),
|
||||
"health": cached_info.get("health", "未知"),
|
||||
"temperature": cached_info.get("temperature", "未知"),
|
||||
"power_on_hours": cached_info.get("power_on_hours", "未知"),
|
||||
"attributes": cached_info.get("attributes", {})
|
||||
})
|
||||
else:
|
||||
# 如果没有缓存信息,使用默认值
|
||||
disk_info.update({
|
||||
"model": "未检测",
|
||||
"serial": "未检测",
|
||||
"capacity": "未检测",
|
||||
"health": "未检测",
|
||||
"temperature": "未检测",
|
||||
"power_on_hours": "未检测",
|
||||
"model": "未知",
|
||||
"serial": "未知",
|
||||
"capacity": "未知",
|
||||
"health": "未知",
|
||||
"temperature": "未知",
|
||||
"power_on_hours": "未知",
|
||||
"attributes": {}
|
||||
})
|
||||
|
||||
@@ -240,31 +413,45 @@ class DiskManager:
|
||||
|
||||
async def _get_full_disk_info(self, disk_info, device_path):
|
||||
"""获取硬盘的完整信息(模型、序列号、健康状态等)"""
|
||||
# 获取基本信息
|
||||
# 获取基本信息 - 首先尝试NVMe格式
|
||||
info_output = await self.coordinator.run_command(f"smartctl -i {device_path}")
|
||||
self.logger.debug("smartctl -i output for %s: %s", disk_info["device"], info_output[:200] + "..." if len(info_output) > 200 else info_output)
|
||||
|
||||
# 模型
|
||||
# 检查是否为NVMe设备
|
||||
is_nvme = "nvme" in disk_info["device"].lower()
|
||||
|
||||
# 模型 - 增强NVMe支持
|
||||
disk_info["model"] = self.extract_value(
|
||||
info_output,
|
||||
[
|
||||
r"Device Model:\s*(.+)",
|
||||
r"Model(?: Family)?\s*:\s*(.+)",
|
||||
r"Model\s*Number:\s*(.+)"
|
||||
r"Model Number:\s*(.+)",
|
||||
r"Product:\s*(.+)", # NVMe格式
|
||||
r"Model Number:\s*(.+)", # NVMe格式
|
||||
]
|
||||
)
|
||||
|
||||
# 序列号
|
||||
# 序列号 - 增强NVMe支持
|
||||
disk_info["serial"] = self.extract_value(
|
||||
info_output,
|
||||
r"Serial Number\s*:\s*(.+)"
|
||||
[
|
||||
r"Serial Number\s*:\s*(.+)",
|
||||
r"Serial Number:\s*(.+)", # NVMe格式
|
||||
r"Serial\s*:\s*(.+)", # NVMe格式
|
||||
]
|
||||
)
|
||||
|
||||
# 容量
|
||||
disk_info["capacity"] = self.extract_value(
|
||||
info_output,
|
||||
r"User Capacity:\s*([^[]+)"
|
||||
)
|
||||
# 容量 - 增强NVMe支持并转换为GB/TB格式
|
||||
capacity_patterns = [
|
||||
r"User Capacity:\s*([^\[]+)",
|
||||
r"Namespace 1 Size/Capacity:\s*([^\[]+)", # NVMe格式
|
||||
r"Total NVM Capacity:\s*([^\[]+)", # NVMe格式
|
||||
r"Capacity:\s*([^\[]+)", # NVMe格式
|
||||
]
|
||||
|
||||
raw_capacity = self.extract_value(info_output, capacity_patterns)
|
||||
disk_info["capacity"] = self._format_capacity(raw_capacity)
|
||||
|
||||
# 健康状态
|
||||
health_output = await self.coordinator.run_command(f"smartctl -H {device_path}")
|
||||
@@ -341,6 +528,46 @@ class DiskManager:
|
||||
# 改进的通电时间检测逻辑 - 处理特殊格式
|
||||
power_on_hours = "未知"
|
||||
|
||||
# 检查是否为NVMe设备
|
||||
is_nvme = "nvme" in disk_info["device"].lower()
|
||||
|
||||
# 方法0:NVMe设备的通电时间提取(优先处理)
|
||||
if is_nvme:
|
||||
# NVMe格式的通电时间提取 - 支持带逗号的数字格式
|
||||
nvme_patterns = [
|
||||
r"Power On Hours\s*:\s*([\d,]+)", # 支持带逗号的数字格式(如 "6,123")
|
||||
r"Power On Time\s*:\s*([\d,]+)", # NVMe备用格式
|
||||
r"Power on hours\s*:\s*([\d,]+)", # 小写格式
|
||||
r"Power on time\s*:\s*([\d,]+)", # 小写格式
|
||||
]
|
||||
|
||||
for pattern in nvme_patterns:
|
||||
match = re.search(pattern, data_output, re.IGNORECASE)
|
||||
if match:
|
||||
try:
|
||||
# 处理带逗号的数字格式(如 "6,123")
|
||||
hours_str = match.group(1).replace(',', '')
|
||||
hours = int(hours_str)
|
||||
power_on_hours = f"{hours} 小时"
|
||||
self.logger.debug("Found NVMe power_on_hours via pattern %s: %s", pattern, power_on_hours)
|
||||
break
|
||||
except:
|
||||
continue
|
||||
|
||||
# 如果还没找到,尝试在SMART数据部分查找
|
||||
if power_on_hours == "未知":
|
||||
# 查找SMART数据部分中的Power On Hours
|
||||
smart_section_match = re.search(r"SMART/Health Information.*?Power On Hours\s*:\s*([\d,]+)",
|
||||
data_output, re.IGNORECASE | re.DOTALL)
|
||||
if smart_section_match:
|
||||
try:
|
||||
hours_str = smart_section_match.group(1).replace(',', '')
|
||||
hours = int(hours_str)
|
||||
power_on_hours = f"{hours} 小时"
|
||||
self.logger.debug("Found NVMe power_on_hours in SMART section: %s", power_on_hours)
|
||||
except:
|
||||
pass
|
||||
|
||||
# 方法1:提取属性9的RAW_VALUE(处理特殊格式)
|
||||
attr9_match = re.search(
|
||||
r"^\s*9\s+Power_On_Hours\b[^\n]+\s+(\d+)h(?:\+(\d+)m(?:\+(\d+)\.\d+s)?)?",
|
||||
@@ -378,7 +605,7 @@ class DiskManager:
|
||||
[
|
||||
# 精确匹配属性9行
|
||||
r"^\s*9\s+Power_On_Hours\b[^\n]+\s+(\d+)\s*$",
|
||||
|
||||
r"^\s*9\s+Power On Hours\b[^\n]+\s+(\d+)h(?:\+(\d+)m(?:\+(\d+)\.\d+s)?)?",
|
||||
# 通用匹配模式
|
||||
r"9\s+Power_On_Hours\b.*?(\d+)\b",
|
||||
r"Power_On_Hours\b.*?(\d+)\b",
|
||||
@@ -424,7 +651,7 @@ class DiskManager:
|
||||
# 添加额外属性:温度历史记录
|
||||
temp_history = {}
|
||||
# 提取属性194的温度历史
|
||||
temp194_match = re.search(r"194\s+Temperature_Celsius+.*?\(\s*([\d\s]+)$", data_output)
|
||||
temp194_match = re.search(r"194\s+Temperature_Celsius+.*?(\s*[\d\s]+)$", data_output)
|
||||
if temp194_match:
|
||||
try:
|
||||
values = [int(x) for x in temp194_match.group(1).split()]
|
||||
@@ -439,4 +666,199 @@ class DiskManager:
|
||||
pass
|
||||
|
||||
# 保存额外属性
|
||||
disk_info["attributes"] = temp_history
|
||||
disk_info["attributes"] = temp_history
|
||||
|
||||
async def get_zpools(self) -> list[dict]:
|
||||
"""获取ZFS存储池信息"""
|
||||
zpools = []
|
||||
try:
|
||||
self.logger.debug("Fetching ZFS pool list...")
|
||||
# 使用zpool list获取存储池信息(包含所有字段)
|
||||
zpool_output = await self.coordinator.run_command("zpool list 2>/dev/null || echo 'NO_ZPOOL'")
|
||||
self.logger.debug("zpool list output: %s", zpool_output)
|
||||
|
||||
if "NO_ZPOOL" in zpool_output or "command not found" in zpool_output.lower():
|
||||
self.logger.info("系统未安装ZFS或没有ZFS存储池")
|
||||
return []
|
||||
|
||||
# 解析zpool list输出
|
||||
lines = zpool_output.splitlines()
|
||||
# 跳过标题行,从第二行开始解析
|
||||
for line in lines[1:]: # 跳过第一行标题
|
||||
if line.strip():
|
||||
# 分割制表符或连续空格
|
||||
parts = re.split(r'\s+', line.strip())
|
||||
if len(parts) >= 11: # 根据实际输出有11个字段
|
||||
pool_info = {
|
||||
"name": parts[0],
|
||||
"size": parts[1],
|
||||
"alloc": parts[2],
|
||||
"free": parts[3],
|
||||
"ckpoint": parts[4] if parts[4] != "-" else "",
|
||||
"expand_sz": parts[5] if parts[5] != "-" else "",
|
||||
"frag": parts[6] if parts[6] != "-" else "0%",
|
||||
"capacity": parts[7],
|
||||
"dedup": parts[8],
|
||||
"health": parts[9],
|
||||
"altroot": parts[10] if parts[10] != "-" else ""
|
||||
}
|
||||
|
||||
zpools.append(pool_info)
|
||||
self.logger.debug("Found ZFS pool: %s", pool_info["name"])
|
||||
|
||||
self.logger.info("Found %d ZFS pools", len(zpools))
|
||||
return zpools
|
||||
|
||||
except Exception as e:
|
||||
self.logger.error("Failed to get ZFS pool info: %s", str(e), exc_info=True)
|
||||
return []
|
||||
|
||||
async def get_zpool_status(self, pool_name: str) -> dict:
|
||||
"""获取ZFS存储池的详细状态信息,包括scrub进度"""
|
||||
try:
|
||||
self.logger.debug(f"Getting ZFS pool status for {pool_name}")
|
||||
status_output = await self.coordinator.run_command(f"zpool status {pool_name} 2>/dev/null || echo 'NO_POOL'")
|
||||
|
||||
if "NO_POOL" in status_output or "command not found" in status_output.lower():
|
||||
self.logger.debug(f"ZFS pool {pool_name} not found")
|
||||
return {"scrub_in_progress": False}
|
||||
|
||||
# 解析scrub信息
|
||||
scrub_info = self._parse_scrub_info(status_output)
|
||||
return scrub_info
|
||||
|
||||
except Exception as e:
|
||||
self.logger.error(f"Failed to get ZFS pool status for {pool_name}: {str(e)}", exc_info=True)
|
||||
return {"scrub_in_progress": False}
|
||||
|
||||
def _parse_scrub_info(self, status_output: str) -> dict:
|
||||
"""解析zpool status中的scrub信息"""
|
||||
scrub_info = {
|
||||
"scrub_in_progress": False,
|
||||
"scrub_status": "无检查",
|
||||
"scrub_progress": "0%",
|
||||
"scan_rate": "0/s",
|
||||
"time_remaining": "",
|
||||
"scanned": "0",
|
||||
"issued": "0",
|
||||
"repaired": "0",
|
||||
"scrub_start_time": ""
|
||||
}
|
||||
|
||||
lines = status_output.split('\n')
|
||||
has_scan_section = False
|
||||
|
||||
# 首先判断是否有scan段(这是判断scrub进行中的关键)
|
||||
for line in lines:
|
||||
line = line.strip()
|
||||
if line.startswith('scan:'):
|
||||
has_scan_section = True
|
||||
break
|
||||
|
||||
# 如果没有scan段,直接返回无检查状态
|
||||
if not has_scan_section:
|
||||
return scrub_info
|
||||
|
||||
# 解析scan段的内容
|
||||
in_scan_section = False
|
||||
for line in lines:
|
||||
line = line.strip()
|
||||
|
||||
# 检查是否进入scan部分
|
||||
if line.startswith('scan:'):
|
||||
in_scan_section = True
|
||||
scrub_info["scrub_in_progress"] = True # 有scan段就表示在进行中
|
||||
scan_line = line[5:].strip() # 去掉'scan:'
|
||||
|
||||
# 检查scrub具体状态
|
||||
if 'scrub in progress' in scan_line or 'scrub resilvering' in scan_line:
|
||||
scrub_info["scrub_status"] = "检查进行中"
|
||||
scrub_info["scrub_progress"] = "0.1%" # 刚开始,显示微小进度表示进行中
|
||||
|
||||
# 解析开始时间
|
||||
if 'since' in scan_line:
|
||||
time_part = scan_line.split('since')[-1].strip()
|
||||
scrub_info["scrub_start_time"] = time_part
|
||||
|
||||
elif 'scrub repaired' in scan_line or 'scrub completed' in scan_line:
|
||||
scrub_info["scrub_status"] = "检查完成"
|
||||
scrub_info["scrub_in_progress"] = False
|
||||
elif 'scrub canceled' in scan_line:
|
||||
scrub_info["scrub_status"] = "检查已取消"
|
||||
scrub_info["scrub_in_progress"] = False
|
||||
elif 'scrub paused' in scan_line:
|
||||
scrub_info["scrub_status"] = "检查已暂停"
|
||||
scrub_info["scrub_in_progress"] = False
|
||||
else:
|
||||
# 有scan段但没有具体状态说明,默认为进行中
|
||||
scrub_info["scrub_status"] = "检查进行中"
|
||||
scrub_info["scrub_progress"] = "0.1%"
|
||||
|
||||
continue
|
||||
|
||||
# 如果在scan部分,解析详细信息
|
||||
if in_scan_section and line and not line.startswith('config'):
|
||||
# 解析进度信息,例如: "2.10T / 2.10T scanned, 413G / 2.10T issued at 223M/s"
|
||||
if 'scanned' in line and 'issued' in line:
|
||||
parts = line.split(',')
|
||||
|
||||
# 解析扫描进度
|
||||
if len(parts) >= 1:
|
||||
scanned_part = parts[0].strip()
|
||||
if ' / ' in scanned_part:
|
||||
scanned_data = scanned_part.split(' / ')[0].strip()
|
||||
total_data = scanned_part.split(' / ')[1].split()[0].strip()
|
||||
scrub_info["scanned"] = f"{scanned_data}/{total_data}"
|
||||
|
||||
# 解析发出的数据
|
||||
if len(parts) >= 2:
|
||||
issued_part = parts[1].strip()
|
||||
if ' / ' in issued_part:
|
||||
issued_data = issued_part.split(' / ')[0].strip()
|
||||
total_issued = issued_part.split(' / ')[1].split()[0].strip()
|
||||
scrub_info["issued"] = f"{issued_data}/{total_issued}"
|
||||
|
||||
# 解析扫描速度
|
||||
if 'at' in line:
|
||||
speed_part = line.split('at')[-1].strip().split()[0]
|
||||
scrub_info["scan_rate"] = speed_part
|
||||
|
||||
# 解析进度百分比和剩余时间
|
||||
elif '%' in line and 'done' in line:
|
||||
# 例如: "644M repaired, 19.23% done, 02:12:38 to go"
|
||||
if '%' in line:
|
||||
progress_match = re.search(r'(\d+\.?\d*)%', line)
|
||||
if progress_match:
|
||||
scrub_info["scrub_progress"] = f"{progress_match.group(1)}%"
|
||||
|
||||
if 'repaired' in line:
|
||||
repaired_match = re.search(r'([\d.]+[KMGT]?).*repaired', line)
|
||||
if repaired_match:
|
||||
scrub_info["repaired"] = repaired_match.group(1)
|
||||
|
||||
if 'to go' in line:
|
||||
time_match = re.search(r'(\d{2}:\d{2}:\d{2})\s+to\s+go', line)
|
||||
if time_match:
|
||||
scrub_info["time_remaining"] = time_match.group(1)
|
||||
|
||||
# 如果遇到空行或新章节,退出scan部分
|
||||
elif line == '' or line.startswith('config'):
|
||||
break
|
||||
|
||||
return scrub_info
|
||||
|
||||
def _format_bytes(self, bytes_value: int) -> str:
|
||||
"""将字节数格式化为易读的格式"""
|
||||
try:
|
||||
if bytes_value >= 1024**4: # 1 TB
|
||||
return f"{bytes_value / (1024**4):.1f} TB"
|
||||
elif bytes_value >= 1024**3: # 1 GB
|
||||
return f"{bytes_value / (1024**3):.1f} GB"
|
||||
elif bytes_value >= 1024**2: # 1 MB
|
||||
return f"{bytes_value / (1024**2):.1f} MB"
|
||||
elif bytes_value >= 1024: # 1 KB
|
||||
return f"{bytes_value / 1024:.1f} KB"
|
||||
else:
|
||||
return f"{bytes_value} B"
|
||||
except Exception:
|
||||
return f"{bytes_value} B"
|
||||
@@ -1,10 +1,10 @@
|
||||
{
|
||||
"domain": "fn_nas",
|
||||
"name": "飞牛NAS",
|
||||
"version": "1.3.6",
|
||||
"documentation": "https://github.com/anxms/fn_nas",
|
||||
"version": "1.4.1",
|
||||
"documentation": "https://github.com/xiaochao99/fn_nas",
|
||||
"dependencies": [],
|
||||
"codeowners": ["@anxms"],
|
||||
"codeowners": ["@xiaochao99"],
|
||||
"requirements": ["asyncssh>=2.13.1"],
|
||||
"iot_class": "local_polling",
|
||||
"config_flow": true
|
||||
|
||||
@@ -3,10 +3,16 @@ from homeassistant.components.sensor import SensorEntity, SensorDeviceClass, Sen
|
||||
from homeassistant.helpers.update_coordinator import CoordinatorEntity
|
||||
from homeassistant.const import UnitOfTemperature
|
||||
from .const import (
|
||||
DOMAIN, HDD_TEMP, HDD_HEALTH, HDD_STATUS, SYSTEM_INFO, ICON_DISK,
|
||||
ICON_TEMPERATURE, ICON_HEALTH, ATTR_DISK_MODEL, ATTR_SERIAL_NO,
|
||||
DOMAIN, HDD_TEMP, HDD_STATUS, SYSTEM_INFO, ICON_DISK,
|
||||
ICON_TEMPERATURE, ATTR_DISK_MODEL, ATTR_SERIAL_NO,
|
||||
ATTR_POWER_ON_HOURS, ATTR_TOTAL_CAPACITY, ATTR_HEALTH_STATUS,
|
||||
DEVICE_ID_NAS, DATA_UPDATE_COORDINATOR
|
||||
DEVICE_ID_NAS, DATA_UPDATE_COORDINATOR, ZFS_POOL, ICON_ZFS,
|
||||
ATTR_ZPOOL_NAME, ATTR_ZPOOL_HEALTH, ATTR_ZPOOL_SIZE,
|
||||
ATTR_ZPOOL_ALLOC, ATTR_ZPOOL_FREE, ATTR_ZPOOL_CAPACITY,
|
||||
ATTR_ZPOOL_FRAGMENTATION, ATTR_ZPOOL_CKPOINT, ATTR_ZPOOL_EXPANDSZ,
|
||||
ATTR_ZPOOL_DEDUP, ATTR_ZPOOL_SCRUB_STATUS, ATTR_ZPOOL_SCRUB_PROGRESS,
|
||||
ATTR_ZPOOL_SCRUB_SCAN_RATE, ATTR_ZPOOL_SCRUB_TIME_REMAINING,
|
||||
ATTR_ZPOOL_SCRUB_ISSUED, ATTR_ZPOOL_SCRUB_REPAIRED, DEVICE_ID_ZFS
|
||||
)
|
||||
|
||||
_LOGGER = logging.getLogger(__name__)
|
||||
@@ -38,22 +44,7 @@ async def async_setup_entry(hass, config_entry, async_add_entities):
|
||||
)
|
||||
existing_ids.add(temp_uid)
|
||||
|
||||
# 健康状态传感器
|
||||
health_uid = f"{config_entry.entry_id}_{disk['device']}_health"
|
||||
if health_uid not in existing_ids:
|
||||
entities.append(
|
||||
DiskSensor(
|
||||
coordinator,
|
||||
disk["device"],
|
||||
HDD_HEALTH,
|
||||
f"硬盘 {disk.get('model', '未知')} 健康状态",
|
||||
health_uid,
|
||||
None,
|
||||
ICON_HEALTH,
|
||||
disk
|
||||
)
|
||||
)
|
||||
existing_ids.add(health_uid)
|
||||
|
||||
|
||||
# 硬盘状态传感器
|
||||
status_uid = f"{config_entry.entry_id}_{disk['device']}_status"
|
||||
@@ -244,6 +235,77 @@ async def async_setup_entry(hass, config_entry, async_add_entities):
|
||||
)
|
||||
existing_ids.add(sensor_uid)
|
||||
|
||||
# 添加ZFS存储池传感器
|
||||
if "zpools" in coordinator.data:
|
||||
for zpool in coordinator.data["zpools"]:
|
||||
safe_name = zpool["name"].replace(" ", "_").replace("/", "_").replace(".", "_")
|
||||
|
||||
# ZFS存储池健康状态传感器
|
||||
health_uid = f"{config_entry.entry_id}_zpool_{safe_name}_health"
|
||||
if health_uid not in existing_ids:
|
||||
entities.append(
|
||||
ZFSPoolSensor(
|
||||
coordinator,
|
||||
zpool["name"],
|
||||
"health",
|
||||
f"ZFS {zpool['name']} 健康状态",
|
||||
health_uid,
|
||||
None,
|
||||
ICON_ZFS,
|
||||
zpool
|
||||
)
|
||||
)
|
||||
existing_ids.add(health_uid)
|
||||
|
||||
# ZFS存储池容量使用率传感器
|
||||
capacity_uid = f"{config_entry.entry_id}_zpool_{safe_name}_capacity"
|
||||
if capacity_uid not in existing_ids:
|
||||
entities.append(
|
||||
ZFSPoolSensor(
|
||||
coordinator,
|
||||
zpool["name"],
|
||||
"capacity",
|
||||
f"ZFS {zpool['name']} 使用率",
|
||||
capacity_uid,
|
||||
"%",
|
||||
ICON_ZFS,
|
||||
zpool,
|
||||
device_class=SensorDeviceClass.POWER_FACTOR,
|
||||
state_class=SensorStateClass.MEASUREMENT
|
||||
)
|
||||
)
|
||||
existing_ids.add(capacity_uid)
|
||||
|
||||
# ZFS存储池总大小传感器
|
||||
size_uid = f"{config_entry.entry_id}_zpool_{safe_name}_size"
|
||||
if size_uid not in existing_ids:
|
||||
entities.append(
|
||||
ZFSPoolSensor(
|
||||
coordinator,
|
||||
zpool["name"],
|
||||
"size",
|
||||
f"ZFS {zpool['name']} 容量",
|
||||
size_uid,
|
||||
None, # 动态确定单位
|
||||
ICON_ZFS,
|
||||
zpool
|
||||
)
|
||||
)
|
||||
existing_ids.add(size_uid)
|
||||
|
||||
# ZFS存储池scrub进度传感器
|
||||
scrub_uid = f"{config_entry.entry_id}_zpool_{safe_name}_scrub"
|
||||
if scrub_uid not in existing_ids:
|
||||
entities.append(
|
||||
ZFSScrubSensor(
|
||||
coordinator,
|
||||
zpool["name"],
|
||||
f"ZFS {zpool['name']} 检查进度",
|
||||
scrub_uid
|
||||
)
|
||||
)
|
||||
existing_ids.add(scrub_uid)
|
||||
|
||||
# 添加剩余内存传感器
|
||||
mem_available_uid = f"{config_entry.entry_id}_memory_available"
|
||||
if mem_available_uid not in existing_ids:
|
||||
@@ -302,7 +364,7 @@ class DiskSensor(CoordinatorEntity, SensorEntity):
|
||||
if disk["device"] == self.device_id:
|
||||
if self.sensor_type == HDD_TEMP:
|
||||
temp = disk.get("temperature")
|
||||
if temp is None or temp == "未知" or temp == "未检测":
|
||||
if temp is None or temp == "未知":
|
||||
return None
|
||||
if isinstance(temp, str):
|
||||
try:
|
||||
@@ -314,11 +376,7 @@ class DiskSensor(CoordinatorEntity, SensorEntity):
|
||||
elif isinstance(temp, (int, float)):
|
||||
return temp
|
||||
return None
|
||||
elif self.sensor_type == HDD_HEALTH:
|
||||
health = disk.get("health", "未知")
|
||||
if health == "未检测":
|
||||
return "未检测"
|
||||
return health if health != "未知" else "未知状态"
|
||||
|
||||
elif self.sensor_type == HDD_STATUS:
|
||||
return disk.get("status", "未知")
|
||||
return None
|
||||
@@ -329,6 +387,11 @@ class DiskSensor(CoordinatorEntity, SensorEntity):
|
||||
return SensorDeviceClass.TEMPERATURE
|
||||
return None
|
||||
|
||||
@property
|
||||
def native_unit_of_measurement(self):
|
||||
"""返回内存单位"""
|
||||
return self._attr_native_unit_of_measurement
|
||||
|
||||
@property
|
||||
def extra_state_attributes(self):
|
||||
return {
|
||||
@@ -602,6 +665,11 @@ class MemoryAvailableSensor(CoordinatorEntity, SensorEntity):
|
||||
except (TypeError, ValueError):
|
||||
return None
|
||||
|
||||
@property
|
||||
def native_unit_of_measurement(self):
|
||||
"""返回内存单位"""
|
||||
return self._attr_native_unit_of_measurement
|
||||
|
||||
@property
|
||||
def extra_state_attributes(self):
|
||||
"""返回总内存和已用内存(GB)以及原始字节值"""
|
||||
@@ -692,6 +760,141 @@ class VolumeAvailableSensor(CoordinatorEntity, SensorEntity):
|
||||
"已用容量": vol_info.get("used", "未知"),
|
||||
"使用率": vol_info.get("use_percent", "未知")
|
||||
}
|
||||
|
||||
class ZFSPoolSensor(CoordinatorEntity, SensorEntity):
|
||||
"""ZFS存储池传感器"""
|
||||
|
||||
def __init__(self, coordinator, zpool_name, sensor_type, name, unique_id, unit, icon, zpool_info, device_class=None, state_class=None):
|
||||
super().__init__(coordinator)
|
||||
self.zpool_name = zpool_name
|
||||
self.sensor_type = sensor_type
|
||||
self._attr_name = name
|
||||
self._attr_unique_id = unique_id
|
||||
self._attr_native_unit_of_measurement = unit
|
||||
self._attr_icon = icon
|
||||
self.zpool_info = zpool_info
|
||||
self._attr_device_info = {
|
||||
"identifiers": {(DOMAIN, DEVICE_ID_ZFS)},
|
||||
"name": "ZFS存储池",
|
||||
"via_device": (DOMAIN, DEVICE_ID_NAS)
|
||||
}
|
||||
|
||||
# 设置设备类和状态类(如果提供)
|
||||
if device_class:
|
||||
self._attr_device_class = device_class
|
||||
if state_class:
|
||||
self._attr_state_class = state_class
|
||||
|
||||
@property
|
||||
def native_value(self):
|
||||
"""返回传感器的值"""
|
||||
for zpool in self.coordinator.data.get("zpools", []):
|
||||
if zpool["name"] == self.zpool_name:
|
||||
if self.sensor_type == "health":
|
||||
# 健康状态中英文映射
|
||||
health_map = {
|
||||
"ONLINE": "在线",
|
||||
"DEGRADED": "降级",
|
||||
"FAULTED": "故障",
|
||||
"OFFLINE": "离线",
|
||||
"REMOVED": "已移除",
|
||||
"UNAVAIL": "不可用"
|
||||
}
|
||||
return health_map.get(zpool.get("health", "UNKNOWN"), zpool.get("health", "未知"))
|
||||
elif self.sensor_type == "capacity":
|
||||
# 返回使用率数值(去掉百分号)
|
||||
capacity = zpool.get("capacity", "0%")
|
||||
try:
|
||||
return float(capacity.replace("%", ""))
|
||||
except ValueError:
|
||||
return None
|
||||
elif self.sensor_type == "size":
|
||||
# 返回总大小的数值部分
|
||||
size = zpool.get("size", "0")
|
||||
try:
|
||||
# 提取数字部分
|
||||
import re
|
||||
match = re.search(r'([\d.]+)', size)
|
||||
if match:
|
||||
return float(match.group(1))
|
||||
return None
|
||||
except (ValueError, AttributeError):
|
||||
return None
|
||||
return None
|
||||
|
||||
@property
|
||||
def native_unit_of_measurement(self):
|
||||
"""返回内存单位"""
|
||||
return self._attr_native_unit_of_measurement
|
||||
|
||||
@property
|
||||
def extra_state_attributes(self):
|
||||
"""返回额外的状态属性"""
|
||||
for zpool in self.coordinator.data.get("zpools", []):
|
||||
if zpool["name"] == self.zpool_name:
|
||||
return {
|
||||
ATTR_ZPOOL_NAME: zpool.get("name", "未知"),
|
||||
ATTR_ZPOOL_HEALTH: zpool.get("health", "未知"),
|
||||
ATTR_ZPOOL_SIZE: zpool.get("size", "未知"),
|
||||
ATTR_ZPOOL_ALLOC: zpool.get("alloc", "未知"),
|
||||
ATTR_ZPOOL_FREE: zpool.get("free", "未知"),
|
||||
ATTR_ZPOOL_CAPACITY: zpool.get("capacity", "未知"),
|
||||
ATTR_ZPOOL_FRAGMENTATION: zpool.get("frag", "未知"),
|
||||
ATTR_ZPOOL_CKPOINT: zpool.get("ckpoint", "") if zpool.get("ckpoint") != "" else "无",
|
||||
ATTR_ZPOOL_EXPANDSZ: zpool.get("expand_sz", "") if zpool.get("expand_sz") != "" else "无",
|
||||
ATTR_ZPOOL_DEDUP: zpool.get("dedup", "未知"),
|
||||
"根路径": zpool.get("altroot", "") if zpool.get("altroot") != "" else "默认"
|
||||
}
|
||||
return {}
|
||||
|
||||
class ZFSScrubSensor(CoordinatorEntity, SensorEntity):
|
||||
"""ZFS存储池scrub进度传感器"""
|
||||
|
||||
def __init__(self, coordinator, zpool_name, name, unique_id):
|
||||
super().__init__(coordinator)
|
||||
self.zpool_name = zpool_name
|
||||
self._attr_name = name
|
||||
self._attr_unique_id = unique_id
|
||||
self._attr_native_unit_of_measurement = "%"
|
||||
self._attr_icon = "mdi:progress-check"
|
||||
self._attr_device_info = {
|
||||
"identifiers": {(DOMAIN, DEVICE_ID_ZFS)},
|
||||
"name": "ZFS存储池",
|
||||
"via_device": (DOMAIN, DEVICE_ID_NAS)
|
||||
}
|
||||
self._attr_device_class = SensorDeviceClass.POWER_FACTOR
|
||||
self._attr_state_class = SensorStateClass.MEASUREMENT
|
||||
self.scrub_cache = {}
|
||||
|
||||
@property
|
||||
def native_value(self):
|
||||
"""返回scrub进度百分比"""
|
||||
# 获取scrub状态信息
|
||||
scrub_info = self.coordinator.data.get("scrub_status", {}).get(self.zpool_name, {})
|
||||
progress_str = scrub_info.get("scrub_progress", "0%")
|
||||
|
||||
return attributes
|
||||
try:
|
||||
# 提取数字部分
|
||||
if progress_str and progress_str != "0%":
|
||||
return float(progress_str.replace("%", ""))
|
||||
return 0.0
|
||||
except (ValueError, AttributeError):
|
||||
return 0.0
|
||||
|
||||
@property
|
||||
def extra_state_attributes(self):
|
||||
"""返回scrub详细状态信息"""
|
||||
scrub_info = self.coordinator.data.get("scrub_status", {}).get(self.zpool_name, {})
|
||||
|
||||
return {
|
||||
ATTR_ZPOOL_NAME: self.zpool_name,
|
||||
ATTR_ZPOOL_SCRUB_STATUS: scrub_info.get("scrub_status", "无检查"),
|
||||
ATTR_ZPOOL_SCRUB_PROGRESS: scrub_info.get("scrub_progress", "0%"),
|
||||
ATTR_ZPOOL_SCRUB_SCAN_RATE: scrub_info.get("scan_rate", "0/s"),
|
||||
ATTR_ZPOOL_SCRUB_TIME_REMAINING: scrub_info.get("time_remaining", ""),
|
||||
ATTR_ZPOOL_SCRUB_ISSUED: scrub_info.get("issued", "0"),
|
||||
ATTR_ZPOOL_SCRUB_REPAIRED: scrub_info.get("repaired", "0"),
|
||||
"开始时间": scrub_info.get("scrub_start_time", ""),
|
||||
"扫描数据": scrub_info.get("scanned", "0"),
|
||||
"检查进行中": scrub_info.get("scrub_in_progress", False)
|
||||
}
|
||||
@@ -213,29 +213,84 @@ class SystemManager:
|
||||
for i, line in enumerate(lines):
|
||||
line_lower = line.lower().strip()
|
||||
|
||||
# 主板温度关键词
|
||||
# 主板温度关键词 - 扩展关键词列表
|
||||
if any(keyword in line_lower for keyword in [
|
||||
"motherboard", "mobo", "mb", "system", "chipset",
|
||||
"ambient", "temp1:", "temp2:", "temp3:", "systin"
|
||||
"ambient", "temp1:", "temp2:", "temp3:", "systin",
|
||||
"acpitz", "thermal", "pch", "platform", "board",
|
||||
"sys", "thermal zone", "acpi", "isa"
|
||||
]) and not any(cpu_keyword in line_lower for cpu_keyword in [
|
||||
"cpu", "core", "package", "processor", "tctl", "tdie"
|
||||
]) and not any(exclude in line_lower for exclude in ["fan", "rpm"]):
|
||||
|
||||
self._debug_log(f"找到可能的主板温度行: {line}")
|
||||
|
||||
# 多种温度格式匹配
|
||||
temp_value = None
|
||||
|
||||
# 格式1: +45.0°C (high = +80.0°C, crit = +95.0°C)
|
||||
if '+' in line and '°c' in line_lower:
|
||||
try:
|
||||
temp_match = line.split('+')[1].split('°')[0].strip()
|
||||
temp = float(temp_match)
|
||||
# 主板温度通常在15-70度之间
|
||||
if 15 <= temp <= 70:
|
||||
self._info_log(f"从sensors提取主板温度: {temp:.1f}°C")
|
||||
return f"{temp:.1f} °C"
|
||||
else:
|
||||
self._debug_log(f"主板温度值超出合理范围: {temp:.1f}°C")
|
||||
except (ValueError, IndexError) as e:
|
||||
self._debug_log(f"解析主板温度失败: {e}")
|
||||
continue
|
||||
temp_value = float(temp_match)
|
||||
except (ValueError, IndexError):
|
||||
pass
|
||||
|
||||
# 格式2: 45.0°C
|
||||
if temp_value is None and '°c' in line_lower:
|
||||
try:
|
||||
# 查找数字后跟°C的模式
|
||||
import re
|
||||
temp_match = re.search(r'(\d+\.?\d*)\s*°c', line_lower)
|
||||
if temp_match:
|
||||
temp_value = float(temp_match.group(1))
|
||||
except (ValueError, AttributeError):
|
||||
pass
|
||||
|
||||
# 格式3: 45.0 C (没有°符号)
|
||||
if temp_value is None and (' c' in line_lower or 'c ' in line_lower):
|
||||
try:
|
||||
# 查找数字后跟C的模式
|
||||
import re
|
||||
temp_match = re.search(r'(\d+\.?\d*)\s*c', line_lower)
|
||||
if temp_match:
|
||||
temp_value = float(temp_match.group(1))
|
||||
except (ValueError, AttributeError):
|
||||
pass
|
||||
|
||||
if temp_value is not None:
|
||||
# 主板温度通常在15-70度之间,但放宽范围到10-80度
|
||||
if 10 <= temp_value <= 80:
|
||||
# 存储候选值,不立即返回
|
||||
import re
|
||||
if not hasattr(self, '_temp_candidates'):
|
||||
self._temp_candidates = []
|
||||
self._temp_candidates.append((temp_value, line))
|
||||
self._debug_log(f"找到有效主板温度候选: {temp_value:.1f}°C")
|
||||
else:
|
||||
self._debug_log(f"主板温度值超出合理范围: {temp_value:.1f}°C")
|
||||
continue
|
||||
|
||||
# 处理候选值
|
||||
if hasattr(self, '_temp_candidates') and self._temp_candidates:
|
||||
# 优先选择温度在25-45度之间的值(典型主板温度)
|
||||
ideal_candidates = [t for t in self._temp_candidates if 25 <= t[0] <= 45]
|
||||
if ideal_candidates:
|
||||
best_temp = ideal_candidates[0][0] # 取第一个理想候选值
|
||||
else:
|
||||
# 如果没有理想值,取第一个候选值
|
||||
best_temp = self._temp_candidates[0][0]
|
||||
|
||||
self._info_log(f"从sensors提取主板温度: {best_temp:.1f}°C")
|
||||
# 清理候选值
|
||||
delattr(self, '_temp_candidates')
|
||||
return f"{best_temp:.1f} °C"
|
||||
|
||||
# 如果没有找到主板温度,尝试备用方法
|
||||
self._debug_log("尝试备用方法获取主板温度")
|
||||
mobo_temp = self._extract_mobo_temp_fallback(sensors_output)
|
||||
if mobo_temp != "未知":
|
||||
return mobo_temp
|
||||
|
||||
self._warning_log("未在sensors输出中找到主板温度")
|
||||
return "未知"
|
||||
@@ -244,9 +299,58 @@ class SystemManager:
|
||||
self._error_log(f"解析sensors主板温度输出失败: {e}")
|
||||
return "未知"
|
||||
|
||||
def _extract_mobo_temp_fallback(self, sensors_output: str) -> str:
|
||||
"""备用方法获取主板温度"""
|
||||
try:
|
||||
lines = sensors_output.split('\n')
|
||||
|
||||
# 方法1: 查找非CPU的温度传感器
|
||||
for line in lines:
|
||||
line_lower = line.lower().strip()
|
||||
|
||||
# 跳过明显的CPU相关行
|
||||
if any(cpu_keyword in line_lower for cpu_keyword in [
|
||||
"cpu", "core", "package", "processor", "tctl", "tdie"
|
||||
]):
|
||||
continue
|
||||
|
||||
# 查找温度值
|
||||
if '°c' in line_lower or ' c' in line_lower:
|
||||
# 尝试提取温度值
|
||||
import re
|
||||
temp_match = re.search(r'(\d+\.?\d*)\s*[°]?\s*c', line_lower)
|
||||
if temp_match:
|
||||
temp_value = float(temp_match.group(1))
|
||||
if 15 <= temp_value <= 60: # 主板温度合理范围
|
||||
self._info_log(f"通过备用方法获取主板温度: {temp_value:.1f}°C")
|
||||
return f"{temp_value:.1f} °C"
|
||||
|
||||
# 方法2: 查找hwmon设备中的主板温度
|
||||
for i, line in enumerate(lines):
|
||||
line_lower = line.lower()
|
||||
if "hwmon" in line_lower and "temp" in line_lower:
|
||||
# 检查接下来的几行是否有温度值
|
||||
for j in range(i+1, min(i+5, len(lines))):
|
||||
next_line = lines[j].lower()
|
||||
if '°c' in next_line or ' c' in next_line:
|
||||
import re
|
||||
temp_match = re.search(r'(\d+\.?\d*)\s*[°]?\s*c', next_line)
|
||||
if temp_match:
|
||||
temp_value = float(temp_match.group(1))
|
||||
if 15 <= temp_value <= 60:
|
||||
self._info_log(f"通过hwmon获取主板温度: {temp_value:.1f}°C")
|
||||
return f"{temp_value:.1f} °C"
|
||||
|
||||
return "未知"
|
||||
|
||||
except Exception as e:
|
||||
self._debug_log(f"备用方法获取主板温度失败: {e}")
|
||||
return "未知"
|
||||
|
||||
def format_uptime(self, seconds: float) -> str:
|
||||
"""格式化运行时间为易读格式"""
|
||||
try:
|
||||
days, remainder = divmod(seconds, 86400)
|
||||
days, remainder = divmod(seconds, 86400)
|
||||
hours, remainder = divmod(remainder, 3600)
|
||||
minutes, seconds = divmod(remainder, 60)
|
||||
@@ -293,90 +397,231 @@ class SystemManager:
|
||||
return {}
|
||||
|
||||
async def get_vol_usage(self) -> dict:
|
||||
"""获取 /vol* 开头的存储卷使用信息"""
|
||||
"""获取 /vol* 开头的存储卷使用信息,避免唤醒休眠磁盘"""
|
||||
try:
|
||||
# 优先使用字节单位
|
||||
df_output = await self.coordinator.run_command("df -B 1 /vol* 2>/dev/null")
|
||||
if df_output:
|
||||
return self.parse_df_bytes(df_output)
|
||||
# 首先尝试智能检测活跃卷
|
||||
active_vols = await self.check_active_volumes()
|
||||
|
||||
df_output = await self.coordinator.run_command("df -h /vol*")
|
||||
if df_output:
|
||||
return self.parse_df_human_readable(df_output)
|
||||
if active_vols:
|
||||
# 只查询活跃的卷,避免使用通配符可能唤醒所有磁盘
|
||||
vol_list = " ".join(active_vols)
|
||||
df_output = await self.coordinator.run_command(f"df -B 1 {vol_list} 2>/dev/null")
|
||||
if df_output:
|
||||
result = self.parse_df_bytes(df_output)
|
||||
if result: # 确保有数据返回
|
||||
return result
|
||||
|
||||
df_output = await self.coordinator.run_command(f"df -h {vol_list} 2>/dev/null")
|
||||
if df_output:
|
||||
result = self.parse_df_human_readable(df_output)
|
||||
if result: # 确保有数据返回
|
||||
return result
|
||||
|
||||
# 如果智能检测失败,回退到传统方法(仅在必要时)
|
||||
self._debug_log("智能卷检测无结果,回退到传统检测方法")
|
||||
|
||||
# 优先使用字节单位,但添加错误处理
|
||||
df_output = await self.coordinator.run_command("df -B 1 /vol* 2>/dev/null || true")
|
||||
if df_output and "No such file or directory" not in df_output:
|
||||
result = self.parse_df_bytes(df_output)
|
||||
if result:
|
||||
return result
|
||||
|
||||
df_output = await self.coordinator.run_command("df -h /vol* 2>/dev/null || true")
|
||||
if df_output and "No such file or directory" not in df_output:
|
||||
result = self.parse_df_human_readable(df_output)
|
||||
if result:
|
||||
return result
|
||||
|
||||
# 最后的回退:尝试检测任何挂载的卷
|
||||
mount_output = await self.coordinator.run_command("mount | grep '/vol' || true")
|
||||
if mount_output:
|
||||
vol_points = []
|
||||
for line in mount_output.splitlines():
|
||||
parts = line.split()
|
||||
for part in parts:
|
||||
if part.startswith('/vol') and part not in vol_points:
|
||||
vol_points.append(part)
|
||||
|
||||
if vol_points:
|
||||
self._debug_log(f"从mount输出检测到卷: {vol_points}")
|
||||
vol_list = " ".join(vol_points)
|
||||
df_output = await self.coordinator.run_command(f"df -h {vol_list} 2>/dev/null || true")
|
||||
if df_output:
|
||||
return self.parse_df_human_readable(df_output)
|
||||
|
||||
self._debug_log("所有存储卷检测方法都失败,返回空字典")
|
||||
return {}
|
||||
|
||||
except Exception as e:
|
||||
self.logger.error("获取存储卷信息失败: %s", str(e))
|
||||
self._error_log(f"获取存储卷信息失败: {str(e)}")
|
||||
return {}
|
||||
|
||||
def parse_df_bytes(self, df_output: str) -> dict:
|
||||
volumes = {}
|
||||
for line in df_output.splitlines()[1:]:
|
||||
parts = line.split()
|
||||
if len(parts) < 6:
|
||||
continue
|
||||
|
||||
mount_point = parts[-1]
|
||||
# 只处理 /vol 开头的挂载点
|
||||
if not mount_point.startswith("/vol"):
|
||||
continue
|
||||
|
||||
try:
|
||||
size_bytes = int(parts[1])
|
||||
used_bytes = int(parts[2])
|
||||
avail_bytes = int(parts[3])
|
||||
use_percent = parts[4]
|
||||
|
||||
def bytes_to_human(b):
|
||||
for unit in ['', 'K', 'M', 'G', 'T']:
|
||||
if abs(b) < 1024.0:
|
||||
return f"{b:.1f}{unit}"
|
||||
b /= 1024.0
|
||||
return f"{b:.1f}P"
|
||||
|
||||
volumes[mount_point] = {
|
||||
"filesystem": parts[0],
|
||||
"size": bytes_to_human(size_bytes),
|
||||
"used": bytes_to_human(used_bytes),
|
||||
"available": bytes_to_human(avail_bytes),
|
||||
"use_percent": use_percent
|
||||
}
|
||||
except (ValueError, IndexError) as e:
|
||||
self.logger.debug("解析存储卷行失败: %s - %s", line, str(e))
|
||||
continue
|
||||
async def check_active_volumes(self) -> list:
|
||||
"""检查当前活跃的存储卷,避免唤醒休眠磁盘"""
|
||||
try:
|
||||
# 获取所有挂载点,这个操作不会访问磁盘内容
|
||||
mount_output = await self.coordinator.run_command("mount | grep '/vol' 2>/dev/null || true")
|
||||
if not mount_output:
|
||||
self._debug_log("未找到任何/vol挂载点")
|
||||
return []
|
||||
|
||||
active_vols = []
|
||||
|
||||
for line in mount_output.splitlines():
|
||||
if '/vol' in line:
|
||||
# 提取挂载点
|
||||
parts = line.split()
|
||||
mount_point = None
|
||||
|
||||
# 查找挂载点(通常在 'on' 关键词之后)
|
||||
try:
|
||||
on_index = parts.index('on')
|
||||
if on_index + 1 < len(parts):
|
||||
candidate = parts[on_index + 1]
|
||||
# 严格检查是否以/vol开头
|
||||
if candidate.startswith('/vol'):
|
||||
mount_point = candidate
|
||||
except ValueError:
|
||||
# 如果没有 'on' 关键词,查找以/vol开头的部分
|
||||
for part in parts:
|
||||
if part.startswith('/vol'):
|
||||
mount_point = part
|
||||
break
|
||||
|
||||
# 过滤挂载点:只保留根级别的/vol*挂载点
|
||||
if mount_point and self.is_root_vol_mount(mount_point):
|
||||
# 检查这个卷对应的磁盘是否活跃
|
||||
is_active = await self.is_volume_disk_active(mount_point)
|
||||
if is_active:
|
||||
active_vols.append(mount_point)
|
||||
self._debug_log(f"添加活跃卷: {mount_point}")
|
||||
else:
|
||||
# 即使磁盘不活跃,也添加到列表中,但标记为可能休眠
|
||||
# 这样可以保证有基本的存储信息
|
||||
active_vols.append(mount_point)
|
||||
self._debug_log(f"卷 {mount_point} 对应磁盘可能休眠,但仍包含在检测中")
|
||||
else:
|
||||
self._debug_log(f"跳过非根级别vol挂载点: {mount_point}")
|
||||
|
||||
# 去重并排序
|
||||
active_vols = sorted(list(set(active_vols)))
|
||||
self._debug_log(f"最终检测到的根级别/vol存储卷: {active_vols}")
|
||||
return active_vols
|
||||
|
||||
except Exception as e:
|
||||
self._debug_log(f"检查活跃存储卷失败: {e}")
|
||||
return []
|
||||
|
||||
def is_root_vol_mount(self, mount_point: str) -> bool:
|
||||
"""检查是否为根级别的/vol挂载点"""
|
||||
if not mount_point or not mount_point.startswith('/vol'):
|
||||
return False
|
||||
|
||||
# 移除开头的/vol部分进行分析
|
||||
remainder = mount_point[4:] # 去掉'/vol'
|
||||
|
||||
# 如果remainder为空,说明是/vol,这是根级别
|
||||
if not remainder:
|
||||
return True
|
||||
|
||||
# 如果remainder只是数字(如/vol1, /vol2),这是根级别
|
||||
if remainder.isdigit():
|
||||
return True
|
||||
|
||||
# 如果remainder是单个字母或字母数字组合且没有斜杠,也认为是根级别
|
||||
# 例如:/vola, /volb, /vol1a 等
|
||||
if '/' not in remainder and len(remainder) <= 3:
|
||||
return True
|
||||
|
||||
# 其他情况都认为是子目录,如:
|
||||
# /vol1/docker/overlay2/...
|
||||
# /vol1/data/...
|
||||
# /vol1/config/...
|
||||
self._debug_log(f"检测到子目录挂载点: {mount_point}")
|
||||
return False
|
||||
|
||||
def parse_df_bytes(self, df_output: str) -> dict:
|
||||
"""解析df命令的字节输出"""
|
||||
volumes = {}
|
||||
try:
|
||||
for line in df_output.splitlines()[1:]: # 跳过标题行
|
||||
parts = line.split()
|
||||
if len(parts) < 6:
|
||||
continue
|
||||
|
||||
mount_point = parts[-1]
|
||||
# 严格检查只处理根级别的 /vol 挂载点
|
||||
if not self.is_root_vol_mount(mount_point):
|
||||
self._debug_log(f"跳过非根级别vol挂载点: {mount_point}")
|
||||
continue
|
||||
|
||||
try:
|
||||
size_bytes = int(parts[1])
|
||||
used_bytes = int(parts[2])
|
||||
avail_bytes = int(parts[3])
|
||||
use_percent = parts[4]
|
||||
|
||||
def bytes_to_human(b):
|
||||
for unit in ['', 'K', 'M', 'G', 'T']:
|
||||
if abs(b) < 1024.0:
|
||||
return f"{b:.1f}{unit}"
|
||||
b /= 1024.0
|
||||
return f"{b:.1f}P"
|
||||
|
||||
volumes[mount_point] = {
|
||||
"filesystem": parts[0],
|
||||
"size": bytes_to_human(size_bytes),
|
||||
"used": bytes_to_human(used_bytes),
|
||||
"available": bytes_to_human(avail_bytes),
|
||||
"use_percent": use_percent
|
||||
}
|
||||
self._debug_log(f"添加根级别/vol存储卷信息: {mount_point}")
|
||||
except (ValueError, IndexError) as e:
|
||||
self._debug_log(f"解析存储卷行失败: {line} - {str(e)}")
|
||||
continue
|
||||
except Exception as e:
|
||||
self._error_log(f"解析df字节输出失败: {e}")
|
||||
|
||||
return volumes
|
||||
|
||||
def parse_df_human_readable(self, df_output: str) -> dict:
|
||||
"""解析df命令输出"""
|
||||
volumes = {}
|
||||
for line in df_output.splitlines()[1:]:
|
||||
parts = line.split()
|
||||
if len(parts) < 6:
|
||||
continue
|
||||
try:
|
||||
for line in df_output.splitlines()[1:]: # 跳过标题行
|
||||
parts = line.split()
|
||||
if len(parts) < 6:
|
||||
continue
|
||||
|
||||
mount_point = parts[-1]
|
||||
# 严格检查只处理根级别的 /vol 挂载点
|
||||
if not self.is_root_vol_mount(mount_point):
|
||||
self._debug_log(f"跳过非根级别vol挂载点: {mount_point}")
|
||||
continue
|
||||
|
||||
try:
|
||||
size = parts[1]
|
||||
used = parts[2]
|
||||
avail = parts[3]
|
||||
use_percent = parts[4]
|
||||
|
||||
volumes[mount_point] = {
|
||||
"filesystem": parts[0],
|
||||
"size": size,
|
||||
"used": used,
|
||||
"available": avail,
|
||||
"use_percent": use_percent
|
||||
}
|
||||
self._debug_log(f"添加根级别/vol存储卷信息: {mount_point}")
|
||||
except (ValueError, IndexError) as e:
|
||||
self._debug_log(f"解析存储卷行失败: {line} - {str(e)}")
|
||||
continue
|
||||
except Exception as e:
|
||||
self._error_log(f"解析df输出失败: {e}")
|
||||
|
||||
mount_point = parts[-1]
|
||||
if not mount_point.startswith("/vol"):
|
||||
continue
|
||||
|
||||
try:
|
||||
size = parts[1]
|
||||
used = parts[2]
|
||||
avail = parts[3]
|
||||
use_percent = parts[4]
|
||||
|
||||
volumes[mount_point] = {
|
||||
"filesystem": parts[0],
|
||||
"size": size,
|
||||
"used": used,
|
||||
"available": avail,
|
||||
"use_percent": use_percent
|
||||
}
|
||||
except (ValueError, IndexError) as e:
|
||||
self.logger.debug("解析存储卷行失败: %s - %s", line, str(e))
|
||||
continue
|
||||
|
||||
return volumes
|
||||
|
||||
return volumes
|
||||
|
||||
async def reboot_system(self):
|
||||
"""重启系统"""
|
||||
self._info_log("Initiating system reboot...")
|
||||
|
||||
@@ -84,7 +84,7 @@ class VMManager:
|
||||
|
||||
async def control_vm(self, vm_name, action):
|
||||
"""控制虚拟机操作"""
|
||||
valid_actions = ["start", "shutdown", "reboot"]
|
||||
valid_actions = ["start", "shutdown", "reboot", "destroy"]
|
||||
if action not in valid_actions:
|
||||
raise ValueError(f"无效操作: {action}")
|
||||
|
||||
|
||||
Reference in New Issue
Block a user