214 lines
5.2 KiB
Go
214 lines
5.2 KiB
Go
// +build linux
|
|
|
|
package docker
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"os"
|
|
"os/exec"
|
|
"path"
|
|
"strconv"
|
|
"strings"
|
|
|
|
cpu "github.com/shirou/gopsutil/cpu"
|
|
"github.com/shirou/gopsutil/internal/common"
|
|
)
|
|
|
|
// GetDockerIDList returnes a list of DockerID.
|
|
// This requires certain permission.
|
|
func GetDockerIDList() ([]string, error) {
|
|
path, err := exec.LookPath("docker")
|
|
if err != nil {
|
|
return nil, ErrDockerNotAvailable
|
|
}
|
|
|
|
out, err := exec.Command(path, "ps", "-q", "--no-trunc").Output()
|
|
if err != nil {
|
|
return []string{}, err
|
|
}
|
|
lines := strings.Split(string(out), "\n")
|
|
ret := make([]string, 0, len(lines))
|
|
|
|
for _, l := range lines {
|
|
if l == "" {
|
|
continue
|
|
}
|
|
ret = append(ret, l)
|
|
}
|
|
|
|
return ret, nil
|
|
}
|
|
|
|
// CgroupCPU returnes specified cgroup id CPU status.
|
|
// containerId is same as docker id if you use docker.
|
|
// If you use container via systemd.slice, you could use
|
|
// containerId = docker-<container id>.scope and base=/sys/fs/cgroup/cpuacct/system.slice/
|
|
func CgroupCPU(containerId string, base string) (*cpu.CPUTimesStat, error) {
|
|
statfile := getCgroupFilePath(containerId, base, "cpuacct", "cpuacct.stat")
|
|
lines, err := common.ReadLines(statfile)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
// empty containerId means all cgroup
|
|
if len(containerId) == 0 {
|
|
containerId = "all"
|
|
}
|
|
ret := &cpu.CPUTimesStat{CPU: containerId}
|
|
for _, line := range lines {
|
|
fields := strings.Split(line, " ")
|
|
if fields[0] == "user" {
|
|
user, err := strconv.ParseFloat(fields[1], 64)
|
|
if err == nil {
|
|
ret.User = float64(user)
|
|
}
|
|
}
|
|
if fields[0] == "system" {
|
|
system, err := strconv.ParseFloat(fields[1], 64)
|
|
if err == nil {
|
|
ret.System = float64(system)
|
|
}
|
|
}
|
|
}
|
|
|
|
return ret, nil
|
|
}
|
|
|
|
func CgroupCPUDocker(containerid string) (*cpu.CPUTimesStat, error) {
|
|
return CgroupCPU(containerid, common.HostSys("fs/cgroup/cpuacct/docker"))
|
|
}
|
|
|
|
func CgroupMem(containerId string, base string) (*CgroupMemStat, error) {
|
|
statfile := getCgroupFilePath(containerId, base, "memory", "memory.stat")
|
|
|
|
// empty containerId means all cgroup
|
|
if len(containerId) == 0 {
|
|
containerId = "all"
|
|
}
|
|
lines, err := common.ReadLines(statfile)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
ret := &CgroupMemStat{ContainerID: containerId}
|
|
for _, line := range lines {
|
|
fields := strings.Split(line, " ")
|
|
v, err := strconv.ParseUint(fields[1], 10, 64)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
switch fields[0] {
|
|
case "cache":
|
|
ret.Cache = v
|
|
case "rss":
|
|
ret.RSS = v
|
|
case "rss_huge":
|
|
ret.RSSHuge = v
|
|
case "mapped_file":
|
|
ret.MappedFile = v
|
|
case "pgpgin":
|
|
ret.Pgpgin = v
|
|
case "pgpgout":
|
|
ret.Pgpgout = v
|
|
case "pgfault":
|
|
ret.Pgfault = v
|
|
case "pgmajfault":
|
|
ret.Pgmajfault = v
|
|
case "inactive_anon":
|
|
ret.InactiveAnon = v
|
|
case "active_anon":
|
|
ret.ActiveAnon = v
|
|
case "inactive_file":
|
|
ret.InactiveFile = v
|
|
case "active_file":
|
|
ret.ActiveFile = v
|
|
case "unevictable":
|
|
ret.Unevictable = v
|
|
case "hierarchical_memory_limit":
|
|
ret.HierarchicalMemoryLimit = v
|
|
case "total_cache":
|
|
ret.TotalCache = v
|
|
case "total_rss":
|
|
ret.TotalRSS = v
|
|
case "total_rss_huge":
|
|
ret.TotalRSSHuge = v
|
|
case "total_mapped_file":
|
|
ret.TotalMappedFile = v
|
|
case "total_pgpgin":
|
|
ret.TotalPgpgIn = v
|
|
case "total_pgpgout":
|
|
ret.TotalPgpgOut = v
|
|
case "total_pgfault":
|
|
ret.TotalPgFault = v
|
|
case "total_pgmajfault":
|
|
ret.TotalPgMajFault = v
|
|
case "total_inactive_anon":
|
|
ret.TotalInactiveAnon = v
|
|
case "total_active_anon":
|
|
ret.TotalActiveAnon = v
|
|
case "total_inactive_file":
|
|
ret.TotalInactiveFile = v
|
|
case "total_active_file":
|
|
ret.TotalActiveFile = v
|
|
case "total_unevictable":
|
|
ret.TotalUnevictable = v
|
|
}
|
|
}
|
|
|
|
r, err := getCgroupMemFile(containerId, base, "memory.usage_in_bytes")
|
|
if err == nil {
|
|
ret.MemUsageInBytes = r
|
|
}
|
|
r, err = getCgroupMemFile(containerId, base, "memory.max_usage_in_bytes")
|
|
if err == nil {
|
|
ret.MemMaxUsageInBytes = r
|
|
}
|
|
r, err = getCgroupMemFile(containerId, base, "memory.limit_in_bytes")
|
|
if err == nil {
|
|
ret.MemLimitInBytes = r
|
|
}
|
|
r, err = getCgroupMemFile(containerId, base, "memory.failcnt")
|
|
if err == nil {
|
|
ret.MemFailCnt = r
|
|
}
|
|
|
|
return ret, nil
|
|
}
|
|
|
|
func CgroupMemDocker(containerId string) (*CgroupMemStat, error) {
|
|
return CgroupMem(containerId, common.HostSys("fs/cgroup/memory/docker"))
|
|
}
|
|
|
|
func (m CgroupMemStat) String() string {
|
|
s, _ := json.Marshal(m)
|
|
return string(s)
|
|
}
|
|
|
|
// getCgroupFilePath constructs file path to get targetted stats file.
|
|
func getCgroupFilePath(containerId, base, target, file string) string {
|
|
if len(base) == 0 {
|
|
base = common.HostSys(fmt.Sprintf("fs/cgroup/%s/docker", target))
|
|
}
|
|
statfile := path.Join(base, containerId, file)
|
|
|
|
if _, err := os.Stat(statfile); os.IsNotExist(err) {
|
|
statfile = path.Join(
|
|
common.HostSys(fmt.Sprintf("fs/cgroup/%s/system.slice", target)), "docker-"+containerId+".scope", file)
|
|
}
|
|
|
|
return statfile
|
|
}
|
|
|
|
// getCgroupMemFile reads a cgroup file and return the contents as uint64.
|
|
func getCgroupMemFile(containerId, base, file string) (uint64, error) {
|
|
|
|
statfile := getCgroupFilePath(containerId, base, "memory", file)
|
|
lines, err := common.ReadLines(statfile)
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
if len(lines) != 1 {
|
|
return 0, fmt.Errorf("wrong format file: %s", statfile)
|
|
}
|
|
return strconv.ParseUint(lines[0], 10, 64)
|
|
}
|