This repository has been archived by the owner on May 12, 2021. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 373
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #299 from jshachm/implement-events-command
cli :Implement events command
- Loading branch information
Showing
22 changed files
with
5,246 additions
and
1,188 deletions.
There are no files selected for viewing
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,270 @@ | ||
// Copyright (c) 2014,2015,2016,2017 Docker, Inc. | ||
// Copyright (c) 2018 Huawei Corporation. | ||
// | ||
// SPDX-License-Identifier: Apache-2.0 | ||
// | ||
|
||
package main | ||
|
||
import ( | ||
"encoding/json" | ||
"fmt" | ||
"os" | ||
"sync" | ||
"time" | ||
|
||
vc "github.com/kata-containers/runtime/virtcontainers" | ||
|
||
"github.com/sirupsen/logrus" | ||
"github.com/urfave/cli" | ||
) | ||
|
||
type event struct { | ||
Type string `json:"type"` | ||
ID string `json:"id"` | ||
Data interface{} `json:"data,omitempty"` | ||
} | ||
|
||
// stats is the runc specific stats structure for stability when encoding and decoding stats. | ||
type stats struct { | ||
CPU cpu `json:"cpu"` | ||
Memory memory `json:"memory"` | ||
Pids pids `json:"pids"` | ||
Blkio blkio `json:"blkio"` | ||
Hugetlb map[string]hugetlb `json:"hugetlb"` | ||
IntelRdt intelRdt `json:"intel_rdt"` | ||
} | ||
|
||
type hugetlb struct { | ||
Usage uint64 `json:"usage,omitempty"` | ||
Max uint64 `json:"max,omitempty"` | ||
Failcnt uint64 `json:"failcnt"` | ||
} | ||
|
||
type blkioEntry struct { | ||
Major uint64 `json:"major,omitempty"` | ||
Minor uint64 `json:"minor,omitempty"` | ||
Op string `json:"op,omitempty"` | ||
Value uint64 `json:"value,omitempty"` | ||
} | ||
|
||
type blkio struct { | ||
IoServiceBytesRecursive []blkioEntry `json:"ioServiceBytesRecursive,omitempty"` | ||
IoServicedRecursive []blkioEntry `json:"ioServicedRecursive,omitempty"` | ||
IoQueuedRecursive []blkioEntry `json:"ioQueueRecursive,omitempty"` | ||
IoServiceTimeRecursive []blkioEntry `json:"ioServiceTimeRecursive,omitempty"` | ||
IoWaitTimeRecursive []blkioEntry `json:"ioWaitTimeRecursive,omitempty"` | ||
IoMergedRecursive []blkioEntry `json:"ioMergedRecursive,omitempty"` | ||
IoTimeRecursive []blkioEntry `json:"ioTimeRecursive,omitempty"` | ||
SectorsRecursive []blkioEntry `json:"sectorsRecursive,omitempty"` | ||
} | ||
|
||
type pids struct { | ||
Current uint64 `json:"current,omitempty"` | ||
Limit uint64 `json:"limit,omitempty"` | ||
} | ||
|
||
type throttling struct { | ||
Periods uint64 `json:"periods,omitempty"` | ||
ThrottledPeriods uint64 `json:"throttledPeriods,omitempty"` | ||
ThrottledTime uint64 `json:"throttledTime,omitempty"` | ||
} | ||
|
||
type cpuUsage struct { | ||
// Units: nanoseconds. | ||
Total uint64 `json:"total,omitempty"` | ||
Percpu []uint64 `json:"percpu,omitempty"` | ||
Kernel uint64 `json:"kernel"` | ||
User uint64 `json:"user"` | ||
} | ||
|
||
type cpu struct { | ||
Usage cpuUsage `json:"usage,omitempty"` | ||
Throttling throttling `json:"throttling,omitempty"` | ||
} | ||
|
||
type memoryEntry struct { | ||
Limit uint64 `json:"limit"` | ||
Usage uint64 `json:"usage,omitempty"` | ||
Max uint64 `json:"max,omitempty"` | ||
Failcnt uint64 `json:"failcnt"` | ||
} | ||
|
||
type memory struct { | ||
Cache uint64 `json:"cache,omitempty"` | ||
Usage memoryEntry `json:"usage,omitempty"` | ||
Swap memoryEntry `json:"swap,omitempty"` | ||
Kernel memoryEntry `json:"kernel,omitempty"` | ||
KernelTCP memoryEntry `json:"kernelTCP,omitempty"` | ||
Raw map[string]uint64 `json:"raw,omitempty"` | ||
} | ||
|
||
type l3CacheInfo struct { | ||
CbmMask string `json:"cbm_mask,omitempty"` | ||
MinCbmBits uint64 `json:"min_cbm_bits,omitempty"` | ||
NumClosids uint64 `json:"num_closids,omitempty"` | ||
} | ||
|
||
type intelRdt struct { | ||
// The read-only L3 cache information | ||
L3CacheInfo *l3CacheInfo `json:"l3_cache_info,omitempty"` | ||
|
||
// The read-only L3 cache schema in root | ||
L3CacheSchemaRoot string `json:"l3_cache_schema_root,omitempty"` | ||
|
||
// The L3 cache schema in 'container_id' group | ||
L3CacheSchema string `json:"l3_cache_schema,omitempty"` | ||
} | ||
|
||
var eventsCLICommand = cli.Command{ | ||
Name: "events", | ||
Usage: "display container events such as OOM notifications, cpu, memory, and IO usage statistics", | ||
ArgsUsage: `<container-id> | ||
Where "<container-id>" is the name for the instance of the container.`, | ||
Description: `The events command displays information about the container. By default the | ||
information is displayed once every 5 seconds.`, | ||
Flags: []cli.Flag{ | ||
cli.DurationFlag{ | ||
Name: "interval", | ||
Value: 5 * time.Second, | ||
Usage: "set the stats collection interval", | ||
}, | ||
cli.BoolFlag{ | ||
Name: "stats", | ||
Usage: "display the container's stats then exit", | ||
}, | ||
}, | ||
Action: func(context *cli.Context) error { | ||
containerID := context.Args().First() | ||
if containerID == "" { | ||
return fmt.Errorf("container id cannot be empty") | ||
} | ||
|
||
duration := context.Duration("interval") | ||
if duration <= 0 { | ||
return fmt.Errorf("duration interval must be greater than 0") | ||
} | ||
|
||
status, sandboxID, err := getExistingContainerInfo(containerID) | ||
if err != nil { | ||
return err | ||
} | ||
|
||
if status.State.State == vc.StateStopped { | ||
return fmt.Errorf("container with id %s is not running", status.ID) | ||
} | ||
|
||
var ( | ||
events = make(chan *event, 1024) | ||
group = &sync.WaitGroup{} | ||
) | ||
group.Add(1) | ||
|
||
go func() { | ||
defer group.Done() | ||
enc := json.NewEncoder(os.Stdout) | ||
for e := range events { | ||
if err := enc.Encode(e); err != nil { | ||
logrus.Error(err) | ||
} | ||
} | ||
}() | ||
|
||
if context.Bool("stats") { | ||
s, err := vci.StatsContainer(sandboxID, containerID) | ||
if err != nil { | ||
return err | ||
} | ||
events <- &event{Type: "stats", ID: status.ID, Data: convertVirtcontainerStats(&s)} | ||
close(events) | ||
group.Wait() | ||
return nil | ||
} | ||
|
||
go func() { | ||
for range time.Tick(context.Duration("interval")) { | ||
s, err := vci.StatsContainer(sandboxID, containerID) | ||
if err != nil { | ||
logrus.Error(err) | ||
continue | ||
} | ||
events <- &event{Type: "stats", ID: status.ID, Data: convertVirtcontainerStats(&s)} | ||
} | ||
}() | ||
|
||
group.Wait() | ||
return nil | ||
}, | ||
} | ||
|
||
func convertVirtcontainerStats(containerStats *vc.ContainerStats) *stats { | ||
cg := containerStats.CgroupStats | ||
if cg == nil { | ||
return nil | ||
} | ||
var s stats | ||
s.Pids.Current = cg.PidsStats.Current | ||
s.Pids.Limit = cg.PidsStats.Limit | ||
|
||
s.CPU.Usage.Kernel = cg.CPUStats.CPUUsage.UsageInKernelmode | ||
s.CPU.Usage.User = cg.CPUStats.CPUUsage.UsageInUsermode | ||
s.CPU.Usage.Total = cg.CPUStats.CPUUsage.TotalUsage | ||
s.CPU.Usage.Percpu = cg.CPUStats.CPUUsage.PercpuUsage | ||
s.CPU.Throttling.Periods = cg.CPUStats.ThrottlingData.Periods | ||
s.CPU.Throttling.ThrottledPeriods = cg.CPUStats.ThrottlingData.ThrottledPeriods | ||
s.CPU.Throttling.ThrottledTime = cg.CPUStats.ThrottlingData.ThrottledTime | ||
|
||
s.Memory.Cache = cg.MemoryStats.Cache | ||
s.Memory.Kernel = convertMemoryEntry(cg.MemoryStats.KernelUsage) | ||
s.Memory.KernelTCP = convertMemoryEntry(cg.MemoryStats.KernelTCPUsage) | ||
s.Memory.Swap = convertMemoryEntry(cg.MemoryStats.SwapUsage) | ||
s.Memory.Usage = convertMemoryEntry(cg.MemoryStats.Usage) | ||
s.Memory.Raw = cg.MemoryStats.Stats | ||
|
||
s.Blkio.IoServiceBytesRecursive = convertBlkioEntry(cg.BlkioStats.IoServiceBytesRecursive) | ||
s.Blkio.IoServicedRecursive = convertBlkioEntry(cg.BlkioStats.IoServicedRecursive) | ||
s.Blkio.IoQueuedRecursive = convertBlkioEntry(cg.BlkioStats.IoQueuedRecursive) | ||
s.Blkio.IoServiceTimeRecursive = convertBlkioEntry(cg.BlkioStats.IoServiceTimeRecursive) | ||
s.Blkio.IoWaitTimeRecursive = convertBlkioEntry(cg.BlkioStats.IoWaitTimeRecursive) | ||
s.Blkio.IoMergedRecursive = convertBlkioEntry(cg.BlkioStats.IoMergedRecursive) | ||
s.Blkio.IoTimeRecursive = convertBlkioEntry(cg.BlkioStats.IoTimeRecursive) | ||
s.Blkio.SectorsRecursive = convertBlkioEntry(cg.BlkioStats.SectorsRecursive) | ||
|
||
s.Hugetlb = make(map[string]hugetlb) | ||
for k, v := range cg.HugetlbStats { | ||
s.Hugetlb[k] = convertHugtlb(v) | ||
} | ||
|
||
return &s | ||
} | ||
|
||
func convertHugtlb(c vc.HugetlbStats) hugetlb { | ||
return hugetlb{ | ||
Usage: c.Usage, | ||
Max: c.MaxUsage, | ||
Failcnt: c.Failcnt, | ||
} | ||
} | ||
|
||
func convertMemoryEntry(c vc.MemoryData) memoryEntry { | ||
return memoryEntry{ | ||
Limit: c.Limit, | ||
Usage: c.Usage, | ||
Max: c.MaxUsage, | ||
Failcnt: c.Failcnt, | ||
} | ||
} | ||
|
||
func convertBlkioEntry(c []vc.BlkioStatEntry) []blkioEntry { | ||
var out []blkioEntry | ||
for _, e := range c { | ||
out = append(out, blkioEntry{ | ||
Major: e.Major, | ||
Minor: e.Minor, | ||
Op: e.Op, | ||
Value: e.Value, | ||
}) | ||
} | ||
return out | ||
} |
Oops, something went wrong.