clan_vm_manager: Extracted VMObject to own component
This commit is contained in:
@@ -125,7 +125,7 @@ class Machine:
|
|||||||
if hasattr(self, "flake_path"):
|
if hasattr(self, "flake_path"):
|
||||||
return Path(self.flake_path)
|
return Path(self.flake_path)
|
||||||
|
|
||||||
self.flake_path = nix_metadata(self.flake)["path"]
|
self.flake_path: str = nix_metadata(self.flake)["path"]
|
||||||
return Path(self.flake_path)
|
return Path(self.flake_path)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
|
|||||||
@@ -19,7 +19,6 @@ V = TypeVar(
|
|||||||
|
|
||||||
|
|
||||||
class GKVStore(GObject.GObject, Gio.ListModel, Generic[K, V]):
|
class GKVStore(GObject.GObject, Gio.ListModel, Generic[K, V]):
|
||||||
__gtype_name__ = "MyGKVStore"
|
|
||||||
"""
|
"""
|
||||||
A simple key-value store that implements the Gio.ListModel interface, with generic types for keys and values.
|
A simple key-value store that implements the Gio.ListModel interface, with generic types for keys and values.
|
||||||
Only use self[key] and del self[key] for accessing the items for better performance.
|
Only use self[key] and del self[key] for accessing the items for better performance.
|
||||||
|
|||||||
310
pkgs/clan-vm-manager/clan_vm_manager/components/vmobj.py
Normal file
310
pkgs/clan-vm-manager/clan_vm_manager/components/vmobj.py
Normal file
@@ -0,0 +1,310 @@
|
|||||||
|
import logging
|
||||||
|
import multiprocessing as mp
|
||||||
|
import os
|
||||||
|
import tempfile
|
||||||
|
import threading
|
||||||
|
import time
|
||||||
|
import weakref
|
||||||
|
from collections.abc import Generator
|
||||||
|
from contextlib import contextmanager
|
||||||
|
from datetime import datetime
|
||||||
|
from pathlib import Path
|
||||||
|
from typing import IO, ClassVar
|
||||||
|
|
||||||
|
import gi
|
||||||
|
from clan_cli import vms
|
||||||
|
from clan_cli.clan_uri import ClanScheme, ClanURI
|
||||||
|
from clan_cli.errors import ClanError
|
||||||
|
from clan_cli.history.add import HistoryEntry
|
||||||
|
from clan_cli.machines.machines import Machine
|
||||||
|
|
||||||
|
from clan_vm_manager.components.executor import MPProcess, spawn
|
||||||
|
|
||||||
|
gi.require_version("GObject", "2.0")
|
||||||
|
gi.require_version("Gtk", "4.0")
|
||||||
|
from gi.repository import GLib, GObject, Gtk
|
||||||
|
|
||||||
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class VMObject(GObject.Object):
|
||||||
|
# Define a custom signal with the name "vm_stopped" and a string argument for the message
|
||||||
|
__gsignals__: ClassVar = {
|
||||||
|
"vm_status_changed": (GObject.SignalFlags.RUN_FIRST, None, [GObject.Object])
|
||||||
|
}
|
||||||
|
|
||||||
|
def vm_status_changed_task(self) -> bool:
|
||||||
|
self.emit("vm_status_changed", self)
|
||||||
|
return GLib.SOURCE_REMOVE
|
||||||
|
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
icon: Path,
|
||||||
|
data: HistoryEntry,
|
||||||
|
) -> None:
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
# Store the data from the history entry
|
||||||
|
self.data = data
|
||||||
|
|
||||||
|
# Create a process object to store the VM process
|
||||||
|
self.vm_process = MPProcess("vm_dummy", mp.Process(), Path("./dummy"))
|
||||||
|
self.build_process = MPProcess("build_dummy", mp.Process(), Path("./dummy"))
|
||||||
|
self._start_thread: threading.Thread = threading.Thread()
|
||||||
|
self.machine: Machine | None = None
|
||||||
|
|
||||||
|
# Watcher to stop the VM
|
||||||
|
self.KILL_TIMEOUT = 20 # seconds
|
||||||
|
self._stop_thread: threading.Thread = threading.Thread()
|
||||||
|
|
||||||
|
# Build progress bar vars
|
||||||
|
self.progress_bar: Gtk.ProgressBar = Gtk.ProgressBar()
|
||||||
|
self.progress_bar.hide()
|
||||||
|
self.progress_bar.set_hexpand(True) # Horizontally expand
|
||||||
|
self.prog_bar_id: int = 0
|
||||||
|
|
||||||
|
# Create a temporary directory to store the logs
|
||||||
|
self.log_dir = tempfile.TemporaryDirectory(
|
||||||
|
prefix="clan_vm-", suffix=f"-{self.data.flake.flake_attr}"
|
||||||
|
)
|
||||||
|
self._logs_id: int = 0
|
||||||
|
self._log_file: IO[str] | None = None
|
||||||
|
|
||||||
|
# To be able to set the switch state programmatically
|
||||||
|
# we need to store the handler id returned by the connect method
|
||||||
|
# and block the signal while we change the state. This is cursed.
|
||||||
|
self.switch = Gtk.Switch()
|
||||||
|
self.switch_handler_id: int = self.switch.connect(
|
||||||
|
"notify::active", self.on_switch_toggle
|
||||||
|
)
|
||||||
|
self.connect("vm_status_changed", self.on_vm_status_changed)
|
||||||
|
|
||||||
|
# Make sure the VM is killed when the reference to this object is dropped
|
||||||
|
self._finalizer = weakref.finalize(self, self.kill_ref_drop)
|
||||||
|
|
||||||
|
def on_vm_status_changed(self, vm: "VMObject", _vm: "VMObject") -> None:
|
||||||
|
self.switch.set_state(self.is_running() and not self.is_building())
|
||||||
|
if self.switch.get_sensitive() is False and not self.is_building():
|
||||||
|
self.switch.set_sensitive(True)
|
||||||
|
|
||||||
|
exit_vm = self.vm_process.proc.exitcode
|
||||||
|
exit_build = self.build_process.proc.exitcode
|
||||||
|
exitc = exit_vm or exit_build
|
||||||
|
if not self.is_running() and exitc != 0:
|
||||||
|
self.switch.handler_block(self.switch_handler_id)
|
||||||
|
self.switch.set_active(False)
|
||||||
|
self.switch.handler_unblock(self.switch_handler_id)
|
||||||
|
log.error(f"VM exited with error. Exitcode: {exitc}")
|
||||||
|
|
||||||
|
def on_switch_toggle(self, switch: Gtk.Switch, user_state: bool) -> None:
|
||||||
|
if switch.get_active():
|
||||||
|
switch.set_state(False)
|
||||||
|
self.start()
|
||||||
|
else:
|
||||||
|
switch.set_state(True)
|
||||||
|
self.shutdown()
|
||||||
|
switch.set_sensitive(False)
|
||||||
|
|
||||||
|
# We use a context manager to create the machine object
|
||||||
|
# and make sure it is destroyed when the context is exited
|
||||||
|
@contextmanager
|
||||||
|
def create_machine(self) -> Generator[Machine, None, None]:
|
||||||
|
uri = ClanURI.from_str(
|
||||||
|
url=self.data.flake.flake_url, flake_attr=self.data.flake.flake_attr
|
||||||
|
)
|
||||||
|
match uri.scheme:
|
||||||
|
case ClanScheme.LOCAL.value(path):
|
||||||
|
self.machine = Machine(
|
||||||
|
name=self.data.flake.flake_attr,
|
||||||
|
flake=path, # type: ignore
|
||||||
|
)
|
||||||
|
case ClanScheme.REMOTE.value(url):
|
||||||
|
self.machine = Machine(
|
||||||
|
name=self.data.flake.flake_attr,
|
||||||
|
flake=url, # type: ignore
|
||||||
|
)
|
||||||
|
yield self.machine
|
||||||
|
self.machine = None
|
||||||
|
|
||||||
|
def _pulse_progress_bar_task(self) -> bool:
|
||||||
|
if self.progress_bar.is_visible():
|
||||||
|
self.progress_bar.pulse()
|
||||||
|
return GLib.SOURCE_CONTINUE
|
||||||
|
else:
|
||||||
|
return GLib.SOURCE_REMOVE
|
||||||
|
|
||||||
|
def __start(self) -> None:
|
||||||
|
with self.create_machine() as machine:
|
||||||
|
# Start building VM
|
||||||
|
tstart = datetime.now()
|
||||||
|
log.info(f"Building VM {self.get_id()}")
|
||||||
|
log_dir = Path(str(self.log_dir.name))
|
||||||
|
self.build_process = spawn(
|
||||||
|
on_except=None,
|
||||||
|
out_file=log_dir / "build.log",
|
||||||
|
func=vms.run.build_vm,
|
||||||
|
machine=machine,
|
||||||
|
tmpdir=log_dir,
|
||||||
|
)
|
||||||
|
GLib.idle_add(self.vm_status_changed_task)
|
||||||
|
|
||||||
|
# Start the logs watcher
|
||||||
|
self._logs_id = GLib.timeout_add(
|
||||||
|
50, self._get_logs_task, self.build_process
|
||||||
|
)
|
||||||
|
if self._logs_id == 0:
|
||||||
|
log.error("Failed to start VM log watcher")
|
||||||
|
log.debug(f"Starting logs watcher on file: {self.build_process.out_file}")
|
||||||
|
|
||||||
|
# Start the progress bar and show it
|
||||||
|
self.progress_bar.show()
|
||||||
|
self.prog_bar_id = GLib.timeout_add(100, self._pulse_progress_bar_task)
|
||||||
|
if self.prog_bar_id == 0:
|
||||||
|
log.error("Couldn't spawn a progess bar task")
|
||||||
|
|
||||||
|
# Wait for the build to finish then hide the progress bar
|
||||||
|
self.build_process.proc.join()
|
||||||
|
tend = datetime.now()
|
||||||
|
log.info(f"VM {self.get_id()} build took {tend - tstart}s")
|
||||||
|
self.progress_bar.hide()
|
||||||
|
|
||||||
|
# Check if the VM was built successfully
|
||||||
|
if self.build_process.proc.exitcode != 0:
|
||||||
|
log.error(f"Failed to build VM {self.get_id()}")
|
||||||
|
GLib.idle_add(self.vm_status_changed_task)
|
||||||
|
return
|
||||||
|
log.info(f"Successfully built VM {self.get_id()}")
|
||||||
|
|
||||||
|
# Start the VM
|
||||||
|
self.vm_process = spawn(
|
||||||
|
on_except=None,
|
||||||
|
out_file=Path(str(self.log_dir.name)) / "vm.log",
|
||||||
|
func=vms.run.run_vm,
|
||||||
|
vm=self.data.flake.vm,
|
||||||
|
)
|
||||||
|
log.debug(f"Started VM {self.get_id()}")
|
||||||
|
GLib.idle_add(self.vm_status_changed_task)
|
||||||
|
|
||||||
|
# Start the logs watcher
|
||||||
|
self._logs_id = GLib.timeout_add(50, self._get_logs_task, self.vm_process)
|
||||||
|
if self._logs_id == 0:
|
||||||
|
log.error("Failed to start VM log watcher")
|
||||||
|
log.debug(f"Starting logs watcher on file: {self.vm_process.out_file}")
|
||||||
|
|
||||||
|
# Wait for the VM to stop
|
||||||
|
self.vm_process.proc.join()
|
||||||
|
log.debug(f"VM {self.get_id()} has stopped")
|
||||||
|
GLib.idle_add(self.vm_status_changed_task)
|
||||||
|
|
||||||
|
def start(self) -> None:
|
||||||
|
if self.is_running():
|
||||||
|
log.warn("VM is already running. Ignoring start request")
|
||||||
|
self.emit("vm_status_changed", self)
|
||||||
|
return
|
||||||
|
log.debug(f"VM state dir {self.log_dir.name}")
|
||||||
|
self._start_thread = threading.Thread(target=self.__start)
|
||||||
|
self._start_thread.start()
|
||||||
|
|
||||||
|
def _get_logs_task(self, proc: MPProcess) -> bool:
|
||||||
|
if not proc.out_file.exists():
|
||||||
|
return GLib.SOURCE_CONTINUE
|
||||||
|
|
||||||
|
if not self._log_file:
|
||||||
|
try:
|
||||||
|
self._log_file = open(proc.out_file)
|
||||||
|
except Exception as ex:
|
||||||
|
log.exception(ex)
|
||||||
|
self._log_file = None
|
||||||
|
return GLib.SOURCE_REMOVE
|
||||||
|
|
||||||
|
line = os.read(self._log_file.fileno(), 4096)
|
||||||
|
if len(line) != 0:
|
||||||
|
print(line.decode("utf-8"), end="", flush=True)
|
||||||
|
|
||||||
|
if not proc.proc.is_alive():
|
||||||
|
log.debug("Removing logs watcher")
|
||||||
|
self._log_file = None
|
||||||
|
return GLib.SOURCE_REMOVE
|
||||||
|
|
||||||
|
return GLib.SOURCE_CONTINUE
|
||||||
|
|
||||||
|
def is_running(self) -> bool:
|
||||||
|
return self._start_thread.is_alive()
|
||||||
|
|
||||||
|
def is_building(self) -> bool:
|
||||||
|
return self.build_process.proc.is_alive()
|
||||||
|
|
||||||
|
def is_shutting_down(self) -> bool:
|
||||||
|
return self._stop_thread.is_alive()
|
||||||
|
|
||||||
|
def get_id(self) -> str:
|
||||||
|
return f"{self.data.flake.flake_url}#{self.data.flake.flake_attr}"
|
||||||
|
|
||||||
|
def __stop(self) -> None:
|
||||||
|
log.info(f"Stopping VM {self.get_id()}")
|
||||||
|
|
||||||
|
start_time = datetime.now()
|
||||||
|
while self.is_running():
|
||||||
|
diff = datetime.now() - start_time
|
||||||
|
if diff.seconds > self.KILL_TIMEOUT:
|
||||||
|
log.error(
|
||||||
|
f"VM {self.get_id()} has not stopped after {self.KILL_TIMEOUT}s. Killing it"
|
||||||
|
)
|
||||||
|
self.vm_process.kill_group()
|
||||||
|
return
|
||||||
|
if self.is_building():
|
||||||
|
log.info(f"VM {self.get_id()} is still building. Killing it")
|
||||||
|
self.build_process.kill_group()
|
||||||
|
return
|
||||||
|
if not self.machine:
|
||||||
|
log.error(f"Machine object is None. Killing VM {self.get_id()}")
|
||||||
|
self.vm_process.kill_group()
|
||||||
|
return
|
||||||
|
|
||||||
|
# Try to shutdown the VM gracefully using QMP
|
||||||
|
try:
|
||||||
|
with self.machine.vm.qmp_ctx() as qmp:
|
||||||
|
qmp.command("system_powerdown")
|
||||||
|
except (OSError, ClanError) as ex:
|
||||||
|
log.debug(f"QMP command 'system_powerdown' ignored. Error: {ex}")
|
||||||
|
|
||||||
|
# Try 20 times to stop the VM
|
||||||
|
time.sleep(self.KILL_TIMEOUT / 20)
|
||||||
|
GLib.idle_add(self.vm_status_changed_task)
|
||||||
|
log.debug(f"VM {self.get_id()} has stopped")
|
||||||
|
|
||||||
|
def shutdown(self) -> None:
|
||||||
|
if not self.is_running():
|
||||||
|
log.warning("VM not running. Ignoring shutdown request.")
|
||||||
|
self.emit("vm_status_changed", self)
|
||||||
|
return
|
||||||
|
if self.is_shutting_down():
|
||||||
|
log.warning("Shutdown already in progress")
|
||||||
|
self.emit("vm_status_changed", self)
|
||||||
|
return
|
||||||
|
self._stop_thread = threading.Thread(target=self.__stop)
|
||||||
|
self._stop_thread.start()
|
||||||
|
|
||||||
|
def kill_ref_drop(self) -> None:
|
||||||
|
if self.is_running():
|
||||||
|
log.warning("Killing VM due to reference drop")
|
||||||
|
self.kill()
|
||||||
|
|
||||||
|
def kill(self) -> None:
|
||||||
|
if not self.is_running():
|
||||||
|
log.warning(f"Tried to kill VM {self.get_id()} is not running")
|
||||||
|
return
|
||||||
|
log.info(f"Killing VM {self.get_id()} now")
|
||||||
|
self.vm_process.kill_group()
|
||||||
|
|
||||||
|
def read_whole_log(self) -> str:
|
||||||
|
if not self.vm_process.out_file.exists():
|
||||||
|
log.error(f"Log file {self.vm_process.out_file} does not exist")
|
||||||
|
return ""
|
||||||
|
return self.vm_process.out_file.read_text()
|
||||||
|
|
||||||
|
def __str__(self) -> str:
|
||||||
|
return f"VM({self.get_id()})"
|
||||||
|
|
||||||
|
def __repr__(self) -> str:
|
||||||
|
return self.__str__()
|
||||||
@@ -1,321 +1,22 @@
|
|||||||
import logging
|
import logging
|
||||||
import multiprocessing as mp
|
|
||||||
import os
|
|
||||||
import tempfile
|
|
||||||
import threading
|
|
||||||
import time
|
|
||||||
import weakref
|
|
||||||
from collections.abc import Generator
|
|
||||||
from contextlib import contextmanager
|
|
||||||
from datetime import datetime
|
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import IO, Any, ClassVar
|
from typing import Any
|
||||||
|
|
||||||
import gi
|
import gi
|
||||||
from clan_cli import vms
|
|
||||||
from clan_cli.clan_uri import ClanScheme, ClanURI
|
|
||||||
from clan_cli.errors import ClanError
|
|
||||||
from clan_cli.history.add import HistoryEntry
|
from clan_cli.history.add import HistoryEntry
|
||||||
from clan_cli.machines.machines import Machine
|
|
||||||
|
|
||||||
from clan_vm_manager import assets
|
from clan_vm_manager import assets
|
||||||
from clan_vm_manager.components.executor import MPProcess, spawn
|
|
||||||
from clan_vm_manager.components.gkvstore import GKVStore
|
from clan_vm_manager.components.gkvstore import GKVStore
|
||||||
|
from clan_vm_manager.components.vmobj import VMObject
|
||||||
|
|
||||||
gi.require_version("GObject", "2.0")
|
gi.require_version("GObject", "2.0")
|
||||||
gi.require_version("Gtk", "4.0")
|
gi.require_version("Gtk", "4.0")
|
||||||
from gi.repository import GLib, GObject, Gtk
|
from gi.repository import GLib
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class VMObject(GObject.Object):
|
|
||||||
__gtype_name__: ClassVar = "VMObject"
|
|
||||||
# Define a custom signal with the name "vm_stopped" and a string argument for the message
|
|
||||||
__gsignals__: ClassVar = {
|
|
||||||
"vm_status_changed": (GObject.SignalFlags.RUN_FIRST, None, [GObject.Object])
|
|
||||||
}
|
|
||||||
|
|
||||||
def vm_status_changed_task(self) -> bool:
|
|
||||||
self.emit("vm_status_changed", self)
|
|
||||||
return GLib.SOURCE_REMOVE
|
|
||||||
|
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
icon: Path,
|
|
||||||
data: HistoryEntry,
|
|
||||||
) -> None:
|
|
||||||
super().__init__()
|
|
||||||
|
|
||||||
# Store the data from the history entry
|
|
||||||
self.data = data
|
|
||||||
|
|
||||||
# Create a process object to store the VM process
|
|
||||||
self.vm_process = MPProcess("vm_dummy", mp.Process(), Path("./dummy"))
|
|
||||||
self.build_process = MPProcess("build_dummy", mp.Process(), Path("./dummy"))
|
|
||||||
self._start_thread: threading.Thread = threading.Thread()
|
|
||||||
self.machine: Machine | None = None
|
|
||||||
|
|
||||||
# Watcher to stop the VM
|
|
||||||
self.KILL_TIMEOUT = 20 # seconds
|
|
||||||
self._stop_thread: threading.Thread = threading.Thread()
|
|
||||||
|
|
||||||
# Build progress bar vars
|
|
||||||
self.progress_bar: Gtk.ProgressBar = Gtk.ProgressBar()
|
|
||||||
self.progress_bar.hide()
|
|
||||||
self.progress_bar.set_hexpand(True) # Horizontally expand
|
|
||||||
self.prog_bar_id: int = 0
|
|
||||||
|
|
||||||
# Create a temporary directory to store the logs
|
|
||||||
self.log_dir = tempfile.TemporaryDirectory(
|
|
||||||
prefix="clan_vm-", suffix=f"-{self.data.flake.flake_attr}"
|
|
||||||
)
|
|
||||||
self._logs_id: int = 0
|
|
||||||
self._log_file: IO[str] | None = None
|
|
||||||
|
|
||||||
# To be able to set the switch state programmatically
|
|
||||||
# we need to store the handler id returned by the connect method
|
|
||||||
# and block the signal while we change the state. This is cursed.
|
|
||||||
self.switch = Gtk.Switch()
|
|
||||||
self.switch_handler_id: int = self.switch.connect(
|
|
||||||
"notify::active", self.on_switch_toggle
|
|
||||||
)
|
|
||||||
self.connect("vm_status_changed", self.on_vm_status_changed)
|
|
||||||
|
|
||||||
# Make sure the VM is killed when the reference to this object is dropped
|
|
||||||
self._finalizer = weakref.finalize(self, self.kill_ref_drop)
|
|
||||||
|
|
||||||
def on_vm_status_changed(self, vm: "VMObject", _vm: "VMObject") -> None:
|
|
||||||
self.switch.set_state(self.is_running() and not self.is_building())
|
|
||||||
if self.switch.get_sensitive() is False and not self.is_building():
|
|
||||||
self.switch.set_sensitive(True)
|
|
||||||
|
|
||||||
exit_vm = self.vm_process.proc.exitcode
|
|
||||||
exit_build = self.build_process.proc.exitcode
|
|
||||||
exitc = exit_vm or exit_build
|
|
||||||
if not self.is_running() and exitc != 0:
|
|
||||||
self.switch.handler_block(self.switch_handler_id)
|
|
||||||
self.switch.set_active(False)
|
|
||||||
self.switch.handler_unblock(self.switch_handler_id)
|
|
||||||
log.error(f"VM exited with error. Exitcode: {exitc}")
|
|
||||||
|
|
||||||
def on_switch_toggle(self, switch: Gtk.Switch, user_state: bool) -> None:
|
|
||||||
if switch.get_active():
|
|
||||||
switch.set_state(False)
|
|
||||||
self.start()
|
|
||||||
else:
|
|
||||||
switch.set_state(True)
|
|
||||||
self.shutdown()
|
|
||||||
switch.set_sensitive(False)
|
|
||||||
|
|
||||||
# We use a context manager to create the machine object
|
|
||||||
# and make sure it is destroyed when the context is exited
|
|
||||||
@contextmanager
|
|
||||||
def create_machine(self) -> Generator[Machine, None, None]:
|
|
||||||
uri = ClanURI.from_str(
|
|
||||||
url=self.data.flake.flake_url, flake_attr=self.data.flake.flake_attr
|
|
||||||
)
|
|
||||||
match uri.scheme:
|
|
||||||
case ClanScheme.LOCAL.value(path):
|
|
||||||
self.machine = Machine(
|
|
||||||
name=self.data.flake.flake_attr,
|
|
||||||
flake=path, # type: ignore
|
|
||||||
)
|
|
||||||
case ClanScheme.REMOTE.value(url):
|
|
||||||
self.machine = Machine(
|
|
||||||
name=self.data.flake.flake_attr,
|
|
||||||
flake=url, # type: ignore
|
|
||||||
)
|
|
||||||
yield self.machine
|
|
||||||
self.machine = None
|
|
||||||
|
|
||||||
def _pulse_progress_bar_task(self) -> bool:
|
|
||||||
if self.progress_bar.is_visible():
|
|
||||||
self.progress_bar.pulse()
|
|
||||||
return GLib.SOURCE_CONTINUE
|
|
||||||
else:
|
|
||||||
return GLib.SOURCE_REMOVE
|
|
||||||
|
|
||||||
def __start(self) -> None:
|
|
||||||
with self.create_machine() as machine:
|
|
||||||
# Start building VM
|
|
||||||
tstart = datetime.now()
|
|
||||||
log.info(f"Building VM {self.get_id()}")
|
|
||||||
log_dir = Path(str(self.log_dir.name))
|
|
||||||
self.build_process = spawn(
|
|
||||||
on_except=None,
|
|
||||||
out_file=log_dir / "build.log",
|
|
||||||
func=vms.run.build_vm,
|
|
||||||
machine=machine,
|
|
||||||
tmpdir=log_dir,
|
|
||||||
)
|
|
||||||
GLib.idle_add(self.vm_status_changed_task)
|
|
||||||
|
|
||||||
# Start the logs watcher
|
|
||||||
self._logs_id = GLib.timeout_add(
|
|
||||||
50, self._get_logs_task, self.build_process
|
|
||||||
)
|
|
||||||
if self._logs_id == 0:
|
|
||||||
log.error("Failed to start VM log watcher")
|
|
||||||
log.debug(f"Starting logs watcher on file: {self.build_process.out_file}")
|
|
||||||
|
|
||||||
# Start the progress bar and show it
|
|
||||||
self.progress_bar.show()
|
|
||||||
self.prog_bar_id = GLib.timeout_add(100, self._pulse_progress_bar_task)
|
|
||||||
if self.prog_bar_id == 0:
|
|
||||||
log.error("Couldn't spawn a progess bar task")
|
|
||||||
|
|
||||||
# Wait for the build to finish then hide the progress bar
|
|
||||||
self.build_process.proc.join()
|
|
||||||
tend = datetime.now()
|
|
||||||
log.info(f"VM {self.get_id()} build took {tend - tstart}s")
|
|
||||||
self.progress_bar.hide()
|
|
||||||
|
|
||||||
# Check if the VM was built successfully
|
|
||||||
if self.build_process.proc.exitcode != 0:
|
|
||||||
log.error(f"Failed to build VM {self.get_id()}")
|
|
||||||
GLib.idle_add(self.vm_status_changed_task)
|
|
||||||
return
|
|
||||||
log.info(f"Successfully built VM {self.get_id()}")
|
|
||||||
|
|
||||||
# Start the VM
|
|
||||||
self.vm_process = spawn(
|
|
||||||
on_except=None,
|
|
||||||
out_file=Path(str(self.log_dir.name)) / "vm.log",
|
|
||||||
func=vms.run.run_vm,
|
|
||||||
vm=self.data.flake.vm,
|
|
||||||
)
|
|
||||||
log.debug(f"Started VM {self.get_id()}")
|
|
||||||
GLib.idle_add(self.vm_status_changed_task)
|
|
||||||
|
|
||||||
# Start the logs watcher
|
|
||||||
self._logs_id = GLib.timeout_add(50, self._get_logs_task, self.vm_process)
|
|
||||||
if self._logs_id == 0:
|
|
||||||
log.error("Failed to start VM log watcher")
|
|
||||||
log.debug(f"Starting logs watcher on file: {self.vm_process.out_file}")
|
|
||||||
|
|
||||||
# Wait for the VM to stop
|
|
||||||
self.vm_process.proc.join()
|
|
||||||
log.debug(f"VM {self.get_id()} has stopped")
|
|
||||||
GLib.idle_add(self.vm_status_changed_task)
|
|
||||||
|
|
||||||
def start(self) -> None:
|
|
||||||
if self.is_running():
|
|
||||||
log.warn("VM is already running. Ignoring start request")
|
|
||||||
self.emit("vm_status_changed", self)
|
|
||||||
return
|
|
||||||
log.debug(f"VM state dir {self.log_dir.name}")
|
|
||||||
self._start_thread = threading.Thread(target=self.__start)
|
|
||||||
self._start_thread.start()
|
|
||||||
|
|
||||||
def _get_logs_task(self, proc: MPProcess) -> bool:
|
|
||||||
if not proc.out_file.exists():
|
|
||||||
return GLib.SOURCE_CONTINUE
|
|
||||||
|
|
||||||
if not self._log_file:
|
|
||||||
try:
|
|
||||||
self._log_file = open(proc.out_file)
|
|
||||||
except Exception as ex:
|
|
||||||
log.exception(ex)
|
|
||||||
self._log_file = None
|
|
||||||
return GLib.SOURCE_REMOVE
|
|
||||||
|
|
||||||
line = os.read(self._log_file.fileno(), 4096)
|
|
||||||
if len(line) != 0:
|
|
||||||
print(line.decode("utf-8"), end="", flush=True)
|
|
||||||
|
|
||||||
if not proc.proc.is_alive():
|
|
||||||
log.debug("Removing logs watcher")
|
|
||||||
self._log_file = None
|
|
||||||
return GLib.SOURCE_REMOVE
|
|
||||||
|
|
||||||
return GLib.SOURCE_CONTINUE
|
|
||||||
|
|
||||||
def is_running(self) -> bool:
|
|
||||||
return self._start_thread.is_alive()
|
|
||||||
|
|
||||||
def is_building(self) -> bool:
|
|
||||||
return self.build_process.proc.is_alive()
|
|
||||||
|
|
||||||
def is_shutting_down(self) -> bool:
|
|
||||||
return self._stop_thread.is_alive()
|
|
||||||
|
|
||||||
def get_id(self) -> str:
|
|
||||||
return f"{self.data.flake.flake_url}#{self.data.flake.flake_attr}"
|
|
||||||
|
|
||||||
def __stop(self) -> None:
|
|
||||||
log.info(f"Stopping VM {self.get_id()}")
|
|
||||||
|
|
||||||
start_time = datetime.now()
|
|
||||||
while self.is_running():
|
|
||||||
diff = datetime.now() - start_time
|
|
||||||
if diff.seconds > self.KILL_TIMEOUT:
|
|
||||||
log.error(
|
|
||||||
f"VM {self.get_id()} has not stopped after {self.KILL_TIMEOUT}s. Killing it"
|
|
||||||
)
|
|
||||||
self.vm_process.kill_group()
|
|
||||||
return
|
|
||||||
if self.is_building():
|
|
||||||
log.info(f"VM {self.get_id()} is still building. Killing it")
|
|
||||||
self.build_process.kill_group()
|
|
||||||
return
|
|
||||||
if not self.machine:
|
|
||||||
log.error(f"Machine object is None. Killing VM {self.get_id()}")
|
|
||||||
self.vm_process.kill_group()
|
|
||||||
return
|
|
||||||
|
|
||||||
# Try to shutdown the VM gracefully using QMP
|
|
||||||
try:
|
|
||||||
with self.machine.vm.qmp_ctx() as qmp:
|
|
||||||
qmp.command("system_powerdown")
|
|
||||||
except (OSError, ClanError) as ex:
|
|
||||||
log.debug(f"QMP command 'system_powerdown' ignored. Error: {ex}")
|
|
||||||
|
|
||||||
# Try 20 times to stop the VM
|
|
||||||
time.sleep(self.KILL_TIMEOUT / 20)
|
|
||||||
GLib.idle_add(self.vm_status_changed_task)
|
|
||||||
log.debug(f"VM {self.get_id()} has stopped")
|
|
||||||
|
|
||||||
def shutdown(self) -> None:
|
|
||||||
if not self.is_running():
|
|
||||||
log.warning("VM not running. Ignoring shutdown request.")
|
|
||||||
self.emit("vm_status_changed", self)
|
|
||||||
return
|
|
||||||
if self.is_shutting_down():
|
|
||||||
log.warning("Shutdown already in progress")
|
|
||||||
self.emit("vm_status_changed", self)
|
|
||||||
return
|
|
||||||
self._stop_thread = threading.Thread(target=self.__stop)
|
|
||||||
self._stop_thread.start()
|
|
||||||
|
|
||||||
def kill_ref_drop(self) -> None:
|
|
||||||
if self.is_running():
|
|
||||||
log.warning("Killing VM due to reference drop")
|
|
||||||
self.kill()
|
|
||||||
|
|
||||||
def kill(self) -> None:
|
|
||||||
if not self.is_running():
|
|
||||||
log.warning(f"Tried to kill VM {self.get_id()} is not running")
|
|
||||||
return
|
|
||||||
log.info(f"Killing VM {self.get_id()} now")
|
|
||||||
self.vm_process.kill_group()
|
|
||||||
|
|
||||||
def read_whole_log(self) -> str:
|
|
||||||
if not self.vm_process.out_file.exists():
|
|
||||||
log.error(f"Log file {self.vm_process.out_file} does not exist")
|
|
||||||
return ""
|
|
||||||
return self.vm_process.out_file.read_text()
|
|
||||||
|
|
||||||
def __str__(self) -> str:
|
|
||||||
return f"VM({self.get_id()})"
|
|
||||||
|
|
||||||
def __repr__(self) -> str:
|
|
||||||
return self.__str__()
|
|
||||||
|
|
||||||
|
|
||||||
class VMStore(GKVStore):
|
class VMStore(GKVStore):
|
||||||
__gtype_name__ = "MyVMStore"
|
|
||||||
|
|
||||||
def __init__(self) -> None:
|
def __init__(self) -> None:
|
||||||
super().__init__(VMObject, lambda vm: vm.data.flake.flake_attr)
|
super().__init__(VMObject, lambda vm: vm.data.flake.flake_attr)
|
||||||
|
|
||||||
|
|||||||
@@ -8,8 +8,9 @@ from clan_cli import history, machines
|
|||||||
from clan_cli.clan_uri import ClanURI
|
from clan_cli.clan_uri import ClanURI
|
||||||
|
|
||||||
from clan_vm_manager.components.interfaces import ClanConfig
|
from clan_vm_manager.components.interfaces import ClanConfig
|
||||||
|
from clan_vm_manager.components.vmobj import VMObject
|
||||||
from clan_vm_manager.singletons.use_join import JoinList, JoinValue
|
from clan_vm_manager.singletons.use_join import JoinList, JoinValue
|
||||||
from clan_vm_manager.singletons.use_vms import ClanStore, VMObject, VMStore
|
from clan_vm_manager.singletons.use_vms import ClanStore, VMStore
|
||||||
|
|
||||||
gi.require_version("Adw", "1")
|
gi.require_version("Adw", "1")
|
||||||
from gi.repository import Adw, Gdk, Gio, GLib, GObject, Gtk
|
from gi.repository import Adw, Gdk, Gio, GLib, GObject, Gtk
|
||||||
|
|||||||
Reference in New Issue
Block a user