From 2bac7b3aecdba37b7c1e0b036f7850e69ec4345d Mon Sep 17 00:00:00 2001 From: "Alexander V. Nikolaev" Date: Fri, 18 Oct 2024 18:55:36 +0300 Subject: [PATCH 1/2] Fix managed endpoints resolving Signed-off-by: Alexander V. Nikolaev --- src/admin/server.rs | 31 ++++++++++++------------------- 1 file changed, 12 insertions(+), 19 deletions(-) diff --git a/src/admin/server.rs b/src/admin/server.rs index 2b32fda..02d5c73 100644 --- a/src/admin/server.rs +++ b/src/admin/server.rs @@ -107,8 +107,17 @@ impl AdminServiceImpl { self.endpoint(&host_mgr).context("Resolving host agent") } - pub fn endpoint(&self, reentry: &RegistryEntry) -> anyhow::Result { - let transport = reentry.agent()?.to_owned(); + pub fn endpoint(&self, entry: &RegistryEntry) -> anyhow::Result { + let transport = match &entry.placement { + Placement::Managed(parent) => { + let parent = self.registry.by_name(&parent)?; + parent + .agent() + .with_context(|| "When get_remote_status()")? + .to_owned() // Fail, if parent also `Managed` + } + Placement::Endpoint(endpoint) => endpoint.clone(), // FIXME: avoid clone! + }; let tls_name = transport.tls_name.clone(); Ok(EndpointConfig { transport, @@ -136,23 +145,7 @@ impl AdminServiceImpl { &self, entry: &RegistryEntry, ) -> anyhow::Result { - let transport = match &entry.placement { - Placement::Managed(parent) => { - let parent = self.registry.by_name(parent)?; - parent.agent()?.to_owned() // Fail, if parent also `Managed` - } - Placement::Endpoint(endpoint) => endpoint.clone(), // FIXME: avoid clone! - }; - let tls_name = transport.tls_name.clone(); - info!("Get remote status for {tls_name}!"); - let endpoint = EndpointConfig { - transport: transport.to_owned(), - tls: self.tls_config.clone().map(|mut tls| { - tls.tls_name = Some(tls_name); - tls - }), - }; - + let endpoint = self.endpoint(entry)?; let client = SystemDClient::new(endpoint); client.get_remote_status(entry.name.clone()).await } From cf9df3cee6fa694bf8a5e65328aaa009ce7721b0 Mon Sep 17 00:00:00 2001 From: "Alexander V. Nikolaev" Date: Mon, 21 Oct 2024 12:34:25 +0300 Subject: [PATCH 2/2] Implement stop/pause/resume in client code, add test Signed-off-by: Alexander V. Nikolaev --- client/src/client.rs | 30 ++++++++++++++++++++++++------ nixos/tests/admin.nix | 5 +++++ 2 files changed, 29 insertions(+), 6 deletions(-) diff --git a/client/src/client.rs b/client/src/client.rs index 0f245cc..cbe1727 100644 --- a/client/src/client.rs +++ b/client/src/client.rs @@ -97,14 +97,32 @@ impl AdminClient { let _response = self.connect_to().await?.start_application(request).await?; Ok(()) } - pub async fn stop(&self, _app: String) -> anyhow::Result<()> { - todo!(); + pub async fn stop(&self, app_name: String) -> anyhow::Result<()> { + let request = pb::admin::ApplicationRequest { + app_name, + vm_name: None, + args: Vec::new(), + }; + let _response = self.connect_to().await?.stop_application(request).await?; + Ok(()) } - pub async fn pause(&self, _app: String) -> anyhow::Result<()> { - todo!(); + pub async fn pause(&self, app_name: String) -> anyhow::Result<()> { + let request = pb::admin::ApplicationRequest { + app_name, + vm_name: None, + args: Vec::new(), + }; + let _response = self.connect_to().await?.pause_application(request).await?; + Ok(()) } - pub async fn resume(&self, _app: String) -> anyhow::Result<()> { - todo!(); + pub async fn resume(&self, app_name: String) -> anyhow::Result<()> { + let request = pb::admin::ApplicationRequest { + app_name, + vm_name: None, + args: Vec::new(), + }; + let _response = self.connect_to().await?.resume_application(request).await?; + Ok(()) } pub async fn reboot(&self) -> anyhow::Result<()> { let request = pb::admin::Empty {}; diff --git a/nixos/tests/admin.nix b/nixos/tests/admin.nix index 21a9b98..d4ca70f 100644 --- a/nixos/tests/admin.nix +++ b/nixos/tests/admin.nix @@ -352,6 +352,11 @@ in print(hostvm.succeed("${cli} --addr ${nodes.adminvm.config.givc.admin.addr} --port ${nodes.adminvm.config.givc.admin.port} --cacert ${nodes.hostvm.givc.host.tls.caCertPath} --cert ${nodes.hostvm.givc.host.tls.certPath} --key ${nodes.hostvm.givc.host.tls.keyPath} ${if tls then "" else "--notls"} --name ${nodes.adminvm.config.givc.admin.name} start --vm chromium-vm foot")) wait_for_window("ghaf@appvm") + with subtest("stop application"): + appvm.succeed("pgrep foot") + print(hostvm.succeed("${cli} --addr ${nodes.adminvm.config.givc.admin.addr} --port ${nodes.adminvm.config.givc.admin.port} --cacert ${nodes.hostvm.givc.host.tls.caCertPath} --cert ${nodes.hostvm.givc.host.tls.certPath} --key ${nodes.hostvm.givc.host.tls.keyPath} ${if tls then "" else "--notls"} --name ${nodes.adminvm.config.givc.admin.name} stop foot@1.service")) + appvm.fail("pgrep foot") + with subtest("clear exit and restart"): print(hostvm.succeed("${cli} --addr ${nodes.adminvm.config.givc.admin.addr} --port ${nodes.adminvm.config.givc.admin.port} --cacert ${nodes.hostvm.givc.host.tls.caCertPath} --cert ${nodes.hostvm.givc.host.tls.certPath} --key ${nodes.hostvm.givc.host.tls.keyPath} ${if tls then "" else "--notls"} --name ${nodes.adminvm.config.givc.admin.name} start --vm chromium-vm clearexit")) time.sleep(20) # Give few seconds to application to spin up, exit, then start it again