Rainbond/worker/master/master.go
2019-09-04 14:50:34 +08:00

177 lines
6.0 KiB
Go

// RAINBOND, Application Management Platform
// Copyright (C) 2014-2017 Goodrain Co., Ltd.
// This program is free software: you can redistribute it and/or modify
// it under the terms of the GNU General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version. For any non-GPL usage of Rainbond,
// one or multiple Commercial Licenses authorized by Goodrain Co., Ltd.
// must be obtained first.
// This program is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU General Public License for more details.
// You should have received a copy of the GNU General Public License
// along with this program. If not, see <http://www.gnu.org/licenses/>.
package master
import (
"context"
"fmt"
"strings"
"time"
"github.com/Sirupsen/logrus"
"github.com/goodrain/rainbond/cmd/worker/option"
"github.com/goodrain/rainbond/db"
"github.com/goodrain/rainbond/db/model"
"github.com/goodrain/rainbond/util/leader"
"github.com/goodrain/rainbond/worker/appm/store"
"github.com/goodrain/rainbond/worker/master/podevent"
"github.com/goodrain/rainbond/worker/master/volumes/provider"
"github.com/goodrain/rainbond/worker/master/volumes/provider/lib/controller"
"github.com/goodrain/rainbond/worker/master/volumes/statistical"
"github.com/prometheus/client_golang/prometheus"
corev1 "k8s.io/api/core/v1"
)
//Controller app runtime master controller
type Controller struct {
ctx context.Context
cancel context.CancelFunc
conf option.Config
store store.Storer
dbmanager db.Manager
memoryUse *prometheus.GaugeVec
fsUse *prometheus.GaugeVec
diskCache *statistical.DiskCache
pc *controller.ProvisionController
isLeader bool
stopCh chan struct{}
podEventChs []chan *corev1.Pod
podEvent *podevent.PodEvent
}
//NewMasterController new master controller
func NewMasterController(conf option.Config, store store.Storer) (*Controller, error) {
ctx, cancel := context.WithCancel(context.Background())
// The controller needs to know what the server version is because out-of-tree
// provisioners aren't officially supported until 1.5
serverVersion, err := conf.KubeClient.Discovery().ServerVersion()
if err != nil {
logrus.Errorf("Error getting server version: %v", err)
cancel()
return nil, err
}
// Create the provisioner: it implements the Provisioner interface expected by
// the controller
//statefulset share controller
rainbondssscProvisioner := provider.NewRainbondssscProvisioner()
//statefulset local controller
rainbondsslcProvisioner := provider.NewRainbondsslcProvisioner(conf.KubeClient, store)
// Start the provision controller which will dynamically provision hostPath
// PVs
pc := controller.NewProvisionController(conf.KubeClient, map[string]controller.Provisioner{
rainbondssscProvisioner.Name(): rainbondssscProvisioner,
rainbondsslcProvisioner.Name(): rainbondsslcProvisioner,
}, serverVersion.GitVersion)
stopCh := make(chan struct{})
return &Controller{
conf: conf,
pc: pc,
store: store,
stopCh: stopCh,
cancel: cancel,
ctx: ctx,
dbmanager: db.GetManager(),
memoryUse: prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: "app_resource",
Name: "appmemory",
Help: "tenant service memory used.",
}, []string{"tenant_id", "service_id", "service_status"}),
fsUse: prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: "app_resource",
Name: "appfs",
Help: "tenant service fs used.",
}, []string{"tenant_id", "service_id", "volume_type"}),
diskCache: statistical.CreatDiskCache(ctx),
podEvent: podevent.New(conf.KubeClient, stopCh),
}, nil
}
//IsLeader is leader
func (m *Controller) IsLeader() bool {
return m.isLeader
}
//Start start
func (m *Controller) Start() error {
start := func(stop <-chan struct{}) {
m.isLeader = true
defer func() {
m.isLeader = false
}()
go m.diskCache.Start()
defer m.diskCache.Stop()
go m.pc.Run(stop)
m.store.RegistPodUpdateListener("podEvent", m.podEvent.GetChan())
defer m.store.UnRegistPodUpdateListener("podEvent")
go m.podEvent.Handle()
<-stop
}
// Leader election was requested.
if m.conf.LeaderElectionNamespace == "" {
return fmt.Errorf("-leader-election-namespace must not be empty")
}
if m.conf.LeaderElectionIdentity == "" {
m.conf.LeaderElectionIdentity = m.conf.NodeName
}
if m.conf.LeaderElectionIdentity == "" {
return fmt.Errorf("-leader-election-identity must not be empty")
}
// Name of config map with leader election lock
lockName := "rainbond-appruntime-worker-leader"
go leader.RunAsLeader(m.conf.KubeClient, m.conf.LeaderElectionNamespace, m.conf.LeaderElectionIdentity, lockName, start, func() {})
return nil
}
//Stop stop
func (m *Controller) Stop() {
close(m.stopCh)
}
//Scrape scrape app runtime
func (m *Controller) Scrape(ch chan<- prometheus.Metric, scrapeDurationDesc *prometheus.Desc) {
if !m.isLeader {
return
}
scrapeTime := time.Now()
services := m.store.GetAllAppServices()
status := m.store.GetNeedBillingStatus(nil)
//获取内存使用情况
for _, service := range services {
if _, ok := status[service.ServiceID]; ok {
m.memoryUse.WithLabelValues(service.TenantID, service.ServiceID, "running").Set(float64(service.ContainerMemory * service.Replicas))
}
}
ch <- prometheus.MustNewConstMetric(scrapeDurationDesc, prometheus.GaugeValue, time.Since(scrapeTime).Seconds(), "collect.memory")
scrapeTime = time.Now()
diskcache := m.diskCache.Get()
for k, v := range diskcache {
key := strings.Split(k, "_")
if len(key) == 2 {
m.fsUse.WithLabelValues(key[1], key[0], string(model.ShareFileVolumeType)).Set(v)
}
}
ch <- prometheus.MustNewConstMetric(scrapeDurationDesc, prometheus.GaugeValue, time.Since(scrapeTime).Seconds(), "collect.fs")
m.fsUse.Collect(ch)
m.memoryUse.Collect(ch)
logrus.Infof("success collect app disk and memory used metric")
}