mirror of
https://github.com/aljazceru/kata-containers.git
synced 2025-12-30 04:24:29 +01:00
Firecracker provides a jailer to constrain the VMM. Use this jailer to launch the firecracker VMM instead of launching it directly from the kata-runtime. The jailer will ensure that the firecracker VMM will run in its own network and mount namespace. All assets required by the VMM have to be present within these namespaces. The assets need to be copied or bind mounted into the chroot location setup by jailer in order for firecracker to access these resouces. This includes files, device nodes and all other assets. Jailer automatically sets up the jail to have access to kvm and vhost-vsock. If a jailer is not available (i.e. not setup in the toml) for a given hypervisor the runtime will act as the jailer. Also enhance the hypervisor interface and unit tests to include the network namespace. This allows the hypervisor to choose how and where to lauch the VMM process, vs virtcontainers directly launching the VMM process. Fixes: #1129 Signed-off-by: Manohar Castelino <manohar.r.castelino@intel.com>
483 lines
11 KiB
Go
483 lines
11 KiB
Go
// Copyright (c) 2018 HyperHQ Inc.
|
|
//
|
|
// SPDX-License-Identifier: Apache-2.0
|
|
//
|
|
|
|
package virtcontainers
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"os"
|
|
"path/filepath"
|
|
"time"
|
|
|
|
pb "github.com/kata-containers/runtime/protocols/cache"
|
|
"github.com/kata-containers/runtime/virtcontainers/pkg/uuid"
|
|
"github.com/kata-containers/runtime/virtcontainers/store"
|
|
"github.com/sirupsen/logrus"
|
|
)
|
|
|
|
// VM is abstraction of a virtual machine.
|
|
type VM struct {
|
|
id string
|
|
|
|
hypervisor hypervisor
|
|
agent agent
|
|
|
|
proxy proxy
|
|
proxyPid int
|
|
proxyURL string
|
|
|
|
cpu uint32
|
|
memory uint32
|
|
|
|
cpuDelta uint32
|
|
|
|
store *store.VCStore
|
|
}
|
|
|
|
// VMConfig is a collection of all info that a new blackbox VM needs.
|
|
type VMConfig struct {
|
|
HypervisorType HypervisorType
|
|
HypervisorConfig HypervisorConfig
|
|
|
|
AgentType AgentType
|
|
AgentConfig interface{}
|
|
|
|
ProxyType ProxyType
|
|
ProxyConfig ProxyConfig
|
|
}
|
|
|
|
// Valid check VMConfig validity.
|
|
func (c *VMConfig) Valid() error {
|
|
return c.HypervisorConfig.valid()
|
|
}
|
|
|
|
// ToGrpc convert VMConfig struct to grpc format pb.GrpcVMConfig.
|
|
func (c *VMConfig) ToGrpc() (*pb.GrpcVMConfig, error) {
|
|
data, err := json.Marshal(&c)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
aconf, ok := c.AgentConfig.(KataAgentConfig)
|
|
if !ok {
|
|
return nil, fmt.Errorf("agent type is not supported by VM cache")
|
|
}
|
|
|
|
agentConfig, err := json.Marshal(&aconf)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &pb.GrpcVMConfig{
|
|
Data: data,
|
|
AgentConfig: agentConfig,
|
|
}, nil
|
|
}
|
|
|
|
// GrpcToVMConfig convert grpc format pb.GrpcVMConfig to VMConfig struct.
|
|
func GrpcToVMConfig(j *pb.GrpcVMConfig) (*VMConfig, error) {
|
|
var config VMConfig
|
|
err := json.Unmarshal(j.Data, &config)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if config.AgentType != KataContainersAgent {
|
|
return nil, fmt.Errorf("agent type %s is not supported by VM cache", config.AgentType)
|
|
}
|
|
|
|
var kataConfig KataAgentConfig
|
|
err = json.Unmarshal(j.AgentConfig, &kataConfig)
|
|
if err == nil {
|
|
config.AgentConfig = kataConfig
|
|
}
|
|
|
|
return &config, nil
|
|
}
|
|
|
|
func setupProxy(h hypervisor, agent agent, config VMConfig, id string) (int, string, proxy, error) {
|
|
consoleURL, err := h.getSandboxConsole(id)
|
|
if err != nil {
|
|
return -1, "", nil, err
|
|
}
|
|
agentURL, err := agent.getAgentURL()
|
|
if err != nil {
|
|
return -1, "", nil, err
|
|
}
|
|
|
|
proxy, err := newProxy(config.ProxyType)
|
|
if err != nil {
|
|
return -1, "", nil, err
|
|
}
|
|
|
|
proxyParams := proxyParams{
|
|
id: id,
|
|
path: config.ProxyConfig.Path,
|
|
agentURL: agentURL,
|
|
consoleURL: consoleURL,
|
|
logger: virtLog.WithField("vm", id),
|
|
debug: config.ProxyConfig.Debug,
|
|
}
|
|
pid, url, err := proxy.start(proxyParams)
|
|
if err != nil {
|
|
virtLog.WithFields(logrus.Fields{
|
|
"vm": id,
|
|
"proxy type": config.ProxyType,
|
|
"params": proxyParams,
|
|
}).WithError(err).Error("failed to start proxy")
|
|
return -1, "", nil, err
|
|
}
|
|
|
|
return pid, url, proxy, nil
|
|
}
|
|
|
|
// NewVM creates a new VM based on provided VMConfig.
|
|
func NewVM(ctx context.Context, config VMConfig) (*VM, error) {
|
|
var (
|
|
proxy proxy
|
|
pid int
|
|
url string
|
|
)
|
|
|
|
// 1. setup hypervisor
|
|
hypervisor, err := newHypervisor(config.HypervisorType)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if err = config.Valid(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
id := uuid.Generate().String()
|
|
|
|
virtLog.WithField("vm", id).WithField("config", config).Info("create new vm")
|
|
|
|
vcStore, err := store.NewVCStore(ctx,
|
|
store.SandboxConfigurationRoot(id),
|
|
store.SandboxRuntimeRoot(id))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
defer func() {
|
|
if err != nil {
|
|
virtLog.WithField("vm", id).WithError(err).Error("failed to create new vm")
|
|
virtLog.WithField("vm", id).Errorf("Deleting store for %s", id)
|
|
vcStore.Delete()
|
|
}
|
|
}()
|
|
|
|
if err = hypervisor.createSandbox(ctx, id, NetworkNamespace{}, &config.HypervisorConfig, vcStore); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// 2. setup agent
|
|
agent := newAgent(config.AgentType)
|
|
vmSharePath := buildVMSharePath(id)
|
|
err = agent.configure(hypervisor, id, vmSharePath, isProxyBuiltIn(config.ProxyType), config.AgentConfig)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// 3. boot up guest vm
|
|
if err = hypervisor.startSandbox(vmStartTimeout); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
defer func() {
|
|
if err != nil {
|
|
virtLog.WithField("vm", id).WithError(err).Info("clean up vm")
|
|
hypervisor.stopSandbox()
|
|
}
|
|
}()
|
|
|
|
// 4. setup proxy
|
|
pid, url, proxy, err = setupProxy(hypervisor, agent, config, id)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer func() {
|
|
if err != nil {
|
|
virtLog.WithField("vm", id).WithError(err).Info("clean up proxy")
|
|
proxy.stop(pid)
|
|
}
|
|
}()
|
|
if err = agent.setProxy(nil, proxy, pid, url); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// 5. check agent aliveness
|
|
// VMs booted from template are paused, do not check
|
|
if !config.HypervisorConfig.BootFromTemplate {
|
|
virtLog.WithField("vm", id).Info("check agent status")
|
|
err = agent.check()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
return &VM{
|
|
id: id,
|
|
hypervisor: hypervisor,
|
|
agent: agent,
|
|
proxy: proxy,
|
|
proxyPid: pid,
|
|
proxyURL: url,
|
|
cpu: config.HypervisorConfig.NumVCPUs,
|
|
memory: config.HypervisorConfig.MemorySize,
|
|
store: vcStore,
|
|
}, nil
|
|
}
|
|
|
|
// NewVMFromGrpc creates a new VM based on provided pb.GrpcVM and VMConfig.
|
|
func NewVMFromGrpc(ctx context.Context, v *pb.GrpcVM, config VMConfig) (*VM, error) {
|
|
virtLog.WithField("GrpcVM", v).WithField("config", config).Info("create new vm from Grpc")
|
|
|
|
hypervisor, err := newHypervisor(config.HypervisorType)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
vcStore, err := store.NewVCStore(ctx,
|
|
store.SandboxConfigurationRoot(v.Id),
|
|
store.SandboxRuntimeRoot(v.Id))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
defer func() {
|
|
if err != nil {
|
|
virtLog.WithField("vm", v.Id).WithError(err).Error("failed to create new vm from Grpc")
|
|
virtLog.WithField("vm", v.Id).Errorf("Deleting store for %s", v.Id)
|
|
vcStore.Delete()
|
|
}
|
|
}()
|
|
|
|
err = hypervisor.fromGrpc(ctx, &config.HypervisorConfig, vcStore, v.Hypervisor)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
agent := newAgent(config.AgentType)
|
|
agent.configureFromGrpc(v.Id, isProxyBuiltIn(config.ProxyType), config.AgentConfig)
|
|
|
|
proxy, err := newProxy(config.ProxyType)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
agent.setProxyFromGrpc(proxy, int(v.ProxyPid), v.ProxyURL)
|
|
|
|
return &VM{
|
|
id: v.Id,
|
|
hypervisor: hypervisor,
|
|
agent: agent,
|
|
proxy: proxy,
|
|
proxyPid: int(v.ProxyPid),
|
|
proxyURL: v.ProxyURL,
|
|
cpu: v.Cpu,
|
|
memory: v.Memory,
|
|
cpuDelta: v.CpuDelta,
|
|
}, nil
|
|
}
|
|
|
|
func buildVMSharePath(id string) string {
|
|
return filepath.Join(store.RunVMStoragePath, id, "shared")
|
|
}
|
|
|
|
func (v *VM) logger() logrus.FieldLogger {
|
|
return virtLog.WithField("vm", v.id)
|
|
}
|
|
|
|
// Pause pauses a VM.
|
|
func (v *VM) Pause() error {
|
|
v.logger().Info("pause vm")
|
|
return v.hypervisor.pauseSandbox()
|
|
}
|
|
|
|
// Save saves a VM to persistent disk.
|
|
func (v *VM) Save() error {
|
|
v.logger().Info("save vm")
|
|
return v.hypervisor.saveSandbox()
|
|
}
|
|
|
|
// Resume resumes a paused VM.
|
|
func (v *VM) Resume() error {
|
|
v.logger().Info("resume vm")
|
|
return v.hypervisor.resumeSandbox()
|
|
}
|
|
|
|
// Start kicks off a configured VM.
|
|
func (v *VM) Start() error {
|
|
v.logger().Info("start vm")
|
|
return v.hypervisor.startSandbox(vmStartTimeout)
|
|
}
|
|
|
|
// Disconnect agent and proxy connections to a VM
|
|
func (v *VM) Disconnect() error {
|
|
v.logger().Info("kill vm")
|
|
|
|
if err := v.agent.disconnect(); err != nil {
|
|
v.logger().WithError(err).Error("failed to disconnect agent")
|
|
}
|
|
if err := v.proxy.stop(v.proxyPid); err != nil {
|
|
v.logger().WithError(err).Error("failed to stop proxy")
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Stop stops a VM process.
|
|
func (v *VM) Stop() error {
|
|
v.logger().Info("stop vm")
|
|
|
|
if err := v.hypervisor.stopSandbox(); err != nil {
|
|
return err
|
|
}
|
|
|
|
return v.store.Delete()
|
|
}
|
|
|
|
// AddCPUs adds num of CPUs to the VM.
|
|
func (v *VM) AddCPUs(num uint32) error {
|
|
if num > 0 {
|
|
v.logger().Infof("hot adding %d vCPUs", num)
|
|
if _, err := v.hypervisor.hotplugAddDevice(num, cpuDev); err != nil {
|
|
return err
|
|
}
|
|
v.cpuDelta += num
|
|
v.cpu += num
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// AddMemory adds numMB of memory to the VM.
|
|
func (v *VM) AddMemory(numMB uint32) error {
|
|
if numMB > 0 {
|
|
v.logger().Infof("hot adding %d MB memory", numMB)
|
|
dev := &memoryDevice{1, int(numMB), 0, false}
|
|
if _, err := v.hypervisor.hotplugAddDevice(dev, memoryDev); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// OnlineCPUMemory puts the hotplugged CPU and memory online.
|
|
func (v *VM) OnlineCPUMemory() error {
|
|
v.logger().Infof("online CPU %d and memory", v.cpuDelta)
|
|
err := v.agent.onlineCPUMem(v.cpuDelta, false)
|
|
if err == nil {
|
|
v.cpuDelta = 0
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// ReseedRNG adds random entropy to guest random number generator
|
|
// and reseeds it.
|
|
func (v *VM) ReseedRNG() error {
|
|
v.logger().Infof("reseed guest random number generator")
|
|
urandomDev := "/dev/urandom"
|
|
data := make([]byte, 512)
|
|
f, err := os.OpenFile(urandomDev, os.O_RDONLY, 0)
|
|
if err != nil {
|
|
v.logger().WithError(err).Warnf("fail to open %s", urandomDev)
|
|
return err
|
|
}
|
|
defer f.Close()
|
|
if _, err = f.Read(data); err != nil {
|
|
v.logger().WithError(err).Warnf("fail to read %s", urandomDev)
|
|
return err
|
|
}
|
|
|
|
return v.agent.reseedRNG(data)
|
|
}
|
|
|
|
// SyncTime syncs guest time with host time.
|
|
func (v *VM) SyncTime() error {
|
|
now := time.Now()
|
|
v.logger().WithField("time", now).Infof("sync guest time")
|
|
return v.agent.setGuestDateTime(now)
|
|
}
|
|
|
|
func (v *VM) assignSandbox(s *Sandbox) error {
|
|
// add vm symlinks
|
|
// - link vm socket from sandbox dir (/run/vc/vm/sbid/<kata.sock>) to vm dir (/run/vc/vm/vmid/<kata.sock>)
|
|
// - link 9pfs share path from sandbox dir (/run/kata-containers/shared/sandboxes/sbid/) to vm dir (/run/vc/vm/vmid/shared/)
|
|
|
|
vmSharePath := buildVMSharePath(v.id)
|
|
vmSockDir := v.agent.getVMPath(v.id)
|
|
sbSharePath := s.agent.getSharePath(s.id)
|
|
sbSockDir := s.agent.getVMPath(s.id)
|
|
|
|
v.logger().WithFields(logrus.Fields{
|
|
"vmSharePath": vmSharePath,
|
|
"vmSockDir": vmSockDir,
|
|
"sbSharePath": sbSharePath,
|
|
"sbSockDir": sbSockDir,
|
|
"proxy-pid": v.proxyPid,
|
|
"proxy-url": v.proxyURL,
|
|
}).Infof("assign vm to sandbox %s", s.id)
|
|
|
|
if err := s.agent.setProxy(s, v.proxy, v.proxyPid, v.proxyURL); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := s.agent.reuseAgent(v.agent); err != nil {
|
|
return err
|
|
}
|
|
|
|
// First make sure the symlinks do not exist
|
|
os.RemoveAll(sbSharePath)
|
|
os.RemoveAll(sbSockDir)
|
|
|
|
if err := os.Symlink(vmSharePath, sbSharePath); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := os.Symlink(vmSockDir, sbSockDir); err != nil {
|
|
os.Remove(sbSharePath)
|
|
return err
|
|
}
|
|
|
|
s.hypervisor = v.hypervisor
|
|
s.config.HypervisorConfig.VMid = v.id
|
|
|
|
return nil
|
|
}
|
|
|
|
// ToGrpc convert VM struct to Grpc format pb.GrpcVM.
|
|
func (v *VM) ToGrpc(config VMConfig) (*pb.GrpcVM, error) {
|
|
hJSON, err := v.hypervisor.toGrpc()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &pb.GrpcVM{
|
|
Id: v.id,
|
|
Hypervisor: hJSON,
|
|
|
|
ProxyPid: int64(v.proxyPid),
|
|
ProxyURL: v.proxyURL,
|
|
|
|
Cpu: v.cpu,
|
|
Memory: v.memory,
|
|
CpuDelta: v.cpuDelta,
|
|
}, nil
|
|
}
|
|
|
|
func (v *VM) GetVMStatus() *pb.GrpcVMStatus {
|
|
return &pb.GrpcVMStatus{
|
|
Pid: int64(v.hypervisor.pid()),
|
|
Cpu: v.cpu,
|
|
Memory: v.memory,
|
|
}
|
|
}
|