code cleanup (#301)

This commit is contained in:
Sergey Lanzman 2018-02-08 01:51:15 +02:00 committed by Murali Reddy
parent 336989088a
commit 617c773655
9 changed files with 37 additions and 37 deletions

View File

@ -11,16 +11,16 @@ import (
"golang.org/x/net/context" "golang.org/x/net/context"
) )
//ControllerHeartbeat is the structure to hold the heartbeats sent by controlers //ControllerHeartbeat is the structure to hold the heartbeats sent by controllers
type ControllerHeartbeat struct { type ControllerHeartbeat struct {
Component string Component string
Lastheartbeat time.Time LastHeartBeat time.Time
} }
//HealthController reports the health of the controller loops as a http endpoint //HealthController reports the health of the controller loops as a http endpoint
type HealthController struct { type HealthController struct {
HealthPort uint16 HealthPort uint16
HTTPenabled bool HTTPEnabled bool
Status HealthStats Status HealthStats
Config *options.KubeRouterConfig Config *options.KubeRouterConfig
} }
@ -39,7 +39,7 @@ type HealthStats struct {
func sendHeartBeat(channel chan<- *ControllerHeartbeat, controller string) { func sendHeartBeat(channel chan<- *ControllerHeartbeat, controller string) {
heartbeat := ControllerHeartbeat{ heartbeat := ControllerHeartbeat{
Component: controller, Component: controller,
Lastheartbeat: time.Now(), LastHeartBeat: time.Now(),
} }
channel <- &heartbeat channel <- &heartbeat
} }
@ -66,7 +66,7 @@ func (hc *HealthController) Handler(w http.ResponseWriter, req *http.Request) {
} }
} }
//HandleHeartbeat handles recevied heartbeats onthe health channel //HandleHeartbeat handles received heartbeats on the health channel
func (hc *HealthController) HandleHeartbeat(beat *ControllerHeartbeat) { func (hc *HealthController) HandleHeartbeat(beat *ControllerHeartbeat) {
glog.V(3).Infof("Received heartbeat from %s", beat.Component) glog.V(3).Infof("Received heartbeat from %s", beat.Component)
@ -131,7 +131,7 @@ func (hc *HealthController) Run(healthChan <-chan *ControllerHeartbeat, stopCh <
http.HandleFunc("/healthz", hc.Handler) http.HandleFunc("/healthz", hc.Handler)
if (hc.Config.HealthPort > 0) && (hc.Config.HealthPort <= 65535) { if (hc.Config.HealthPort > 0) && (hc.Config.HealthPort <= 65535) {
hc.HTTPenabled = true hc.HTTPEnabled = true
go func() { go func() {
if err := srv.ListenAndServe(); err != nil { if err := srv.ListenAndServe(); err != nil {
// cannot panic, because this probably is an intentional close // cannot panic, because this probably is an intentional close
@ -141,7 +141,7 @@ func (hc *HealthController) Run(healthChan <-chan *ControllerHeartbeat, stopCh <
} else if hc.Config.MetricsPort > 65535 { } else if hc.Config.MetricsPort > 65535 {
glog.Errorf("Metrics port must be over 0 and under 65535, given port: %d", hc.Config.MetricsPort) glog.Errorf("Metrics port must be over 0 and under 65535, given port: %d", hc.Config.MetricsPort)
} else { } else {
hc.HTTPenabled = false hc.HTTPEnabled = false
} }
for { for {
@ -152,7 +152,7 @@ func (hc *HealthController) Run(healthChan <-chan *ControllerHeartbeat, stopCh <
select { select {
case <-stopCh: case <-stopCh:
glog.Infof("Shutting down health controller") glog.Infof("Shutting down health controller")
if hc.HTTPenabled { if hc.HTTPEnabled {
if err := srv.Shutdown(context.Background()); err != nil { if err := srv.Shutdown(context.Background()); err != nil {
glog.Errorf("could not shutdown: %v", err) glog.Errorf("could not shutdown: %v", err)
} }
@ -167,7 +167,7 @@ func (hc *HealthController) Run(healthChan <-chan *ControllerHeartbeat, stopCh <
} }
//NewHealthController creates a new healh controller and returns a reference to it //NewHealthController creates a new health controller and returns a reference to it
func NewHealthController(config *options.KubeRouterConfig) (*HealthController, error) { func NewHealthController(config *options.KubeRouterConfig) (*HealthController, error) {
hc := HealthController{ hc := HealthController{
Config: config, Config: config,

View File

@ -19,7 +19,7 @@ var (
serviceTotalConn = prometheus.NewGaugeVec(prometheus.GaugeOpts{ serviceTotalConn = prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: namespace, Namespace: namespace,
Name: "service_total_connections", Name: "service_total_connections",
Help: "Total incoming conntections made", Help: "Total incoming connections made",
}, []string{"namespace", "service_name", "service_vip", "protocol", "port"}) }, []string{"namespace", "service_name", "service_vip", "protocol", "port"})
servicePacketsIn = prometheus.NewGaugeVec(prometheus.GaugeOpts{ servicePacketsIn = prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: namespace, Namespace: namespace,
@ -29,7 +29,7 @@ var (
servicePacketsOut = prometheus.NewGaugeVec(prometheus.GaugeOpts{ servicePacketsOut = prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: namespace, Namespace: namespace,
Name: "service_packets_out", Name: "service_packets_out",
Help: "Total outoging packets", Help: "Total outgoing packets",
}, []string{"namespace", "service_name", "service_vip", "protocol", "port"}) }, []string{"namespace", "service_name", "service_vip", "protocol", "port"})
serviceBytesIn = prometheus.NewGaugeVec(prometheus.GaugeOpts{ serviceBytesIn = prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: namespace, Namespace: namespace,
@ -49,7 +49,7 @@ var (
servicePpsOut = prometheus.NewGaugeVec(prometheus.GaugeOpts{ servicePpsOut = prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: namespace, Namespace: namespace,
Name: "service_pps_out", Name: "service_pps_out",
Help: "Outoging packets per second", Help: "Outgoing packets per second",
}, []string{"namespace", "service_name", "service_vip", "protocol", "port"}) }, []string{"namespace", "service_name", "service_vip", "protocol", "port"})
serviceCPS = prometheus.NewGaugeVec(prometheus.GaugeOpts{ serviceCPS = prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: namespace, Namespace: namespace,
@ -64,7 +64,7 @@ var (
serviceBpsOut = prometheus.NewGaugeVec(prometheus.GaugeOpts{ serviceBpsOut = prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: namespace, Namespace: namespace,
Name: "service_bps_out", Name: "service_bps_out",
Help: "Outoging bytes per second", Help: "Outgoing bytes per second",
}, []string{"namespace", "service_name", "service_vip", "protocol", "port"}) }, []string{"namespace", "service_name", "service_vip", "protocol", "port"})
controllerIpvsServices = prometheus.NewGaugeVec(prometheus.GaugeOpts{ controllerIpvsServices = prometheus.NewGaugeVec(prometheus.GaugeOpts{
Namespace: namespace, Namespace: namespace,

View File

@ -871,7 +871,7 @@ func (nrc *NetworkRoutingController) Cleanup() {
} }
func (nrc *NetworkRoutingController) disableSourceDestinationCheck() { func (nrc *NetworkRoutingController) disableSourceDestinationCheck() {
nodes, err := nrc.clientset.Core().Nodes().List(metav1.ListOptions{}) nodes, err := nrc.clientset.CoreV1().Nodes().List(metav1.ListOptions{})
if err != nil { if err != nil {
glog.Errorf("Failed to list nodes from API server due to: %s. Can not perform BGP peer sync", err.Error()) glog.Errorf("Failed to list nodes from API server due to: %s. Can not perform BGP peer sync", err.Error())
return return
@ -882,8 +882,8 @@ func (nrc *NetworkRoutingController) disableSourceDestinationCheck() {
return return
} }
providerID := strings.Replace(node.Spec.ProviderID, "///", "//", 1) providerID := strings.Replace(node.Spec.ProviderID, "///", "//", 1)
url, err := url.Parse(providerID) URL, err := url.Parse(providerID)
instanceID := url.Path instanceID := URL.Path
instanceID = strings.Trim(instanceID, "/") instanceID = strings.Trim(instanceID, "/")
sess, _ := session.NewSession(aws.NewConfig().WithMaxRetries(5)) sess, _ := session.NewSession(aws.NewConfig().WithMaxRetries(5))
@ -913,7 +913,7 @@ func (nrc *NetworkRoutingController) disableSourceDestinationCheck() {
func (nrc *NetworkRoutingController) syncNodeIPSets() error { func (nrc *NetworkRoutingController) syncNodeIPSets() error {
// Get the current list of the nodes from API server // Get the current list of the nodes from API server
nodes, err := nrc.clientset.Core().Nodes().List(metav1.ListOptions{}) nodes, err := nrc.clientset.CoreV1().Nodes().List(metav1.ListOptions{})
if err != nil { if err != nil {
return errors.New("Failed to list nodes from API server: " + err.Error()) return errors.New("Failed to list nodes from API server: " + err.Error())
} }
@ -979,7 +979,7 @@ func (nrc *NetworkRoutingController) syncInternalPeers() {
}() }()
// get the current list of the nodes from API server // get the current list of the nodes from API server
nodes, err := nrc.clientset.Core().Nodes().List(metav1.ListOptions{}) nodes, err := nrc.clientset.CoreV1().Nodes().List(metav1.ListOptions{})
if err != nil { if err != nil {
glog.Errorf("Failed to list nodes from API server due to: %s. Can not perform BGP peer sync", err.Error()) glog.Errorf("Failed to list nodes from API server due to: %s. Can not perform BGP peer sync", err.Error())
return return
@ -1323,7 +1323,7 @@ func (nrc *NetworkRoutingController) startBgpServer() error {
} }
// Get Global Peer Router Password configs // Get Global Peer Router Password configs
peerPasswords := []string{} var peerPasswords []string
nodeBGPPasswordsAnnotation, ok := node.ObjectMeta.Annotations["kube-router.io/peer.passwords"] nodeBGPPasswordsAnnotation, ok := node.ObjectMeta.Annotations["kube-router.io/peer.passwords"]
if !ok { if !ok {
glog.Infof("Could not find BGP peer password info in the node's annotations. Assuming no passwords.") glog.Infof("Could not find BGP peer password info in the node's annotations. Assuming no passwords.")

View File

@ -336,7 +336,7 @@ func (nsc *NetworkServicesController) syncIpvsServices(serviceInfoMap serviceInf
for k, svc := range serviceInfoMap { for k, svc := range serviceInfoMap {
var protocol uint16 var protocol uint16
switch aProtocol := svc.protocol; aProtocol { switch svc.protocol {
case "tcp": case "tcp":
protocol = syscall.IPPROTO_TCP protocol = syscall.IPPROTO_TCP
case "udp": case "udp":
@ -663,12 +663,12 @@ func prepareEndpointForDsr(containerId string, endpointIP string, vip string) er
glog.V(1).Infof("Current network namespace before netns.Set: " + activeNetworkNamespaceHandle.String()) glog.V(1).Infof("Current network namespace before netns.Set: " + activeNetworkNamespaceHandle.String())
activeNetworkNamespaceHandle.Close() activeNetworkNamespaceHandle.Close()
client, err := client.NewEnvClient() dockerClient, err := client.NewEnvClient()
if err != nil { if err != nil {
return errors.New("Failed to get docker client due to " + err.Error()) return errors.New("Failed to get docker client due to " + err.Error())
} }
containerSpec, err := client.ContainerInspect(context.Background(), containerId) containerSpec, err := dockerClient.ContainerInspect(context.Background(), containerId)
if err != nil { if err != nil {
return errors.New("Failed to get docker container spec due to " + err.Error()) return errors.New("Failed to get docker container spec due to " + err.Error())
} }
@ -842,7 +842,7 @@ func buildServicesInfo() serviceInfoMap {
} }
} }
copy(svcInfo.externalIPs, svc.Spec.ExternalIPs) copy(svcInfo.externalIPs, svc.Spec.ExternalIPs)
svcInfo.sessionAffinity = (svc.Spec.SessionAffinity == "ClientIP") svcInfo.sessionAffinity = svc.Spec.SessionAffinity == "ClientIP"
_, svcInfo.hairpin = svc.ObjectMeta.Annotations["kube-router.io/service.hairpin"] _, svcInfo.hairpin = svc.ObjectMeta.Annotations["kube-router.io/service.hairpin"]
_, svcInfo.local = svc.ObjectMeta.Annotations["kube-router.io/service.local"] _, svcInfo.local = svc.ObjectMeta.Annotations["kube-router.io/service.local"]
@ -1487,12 +1487,12 @@ func setupRoutesForExternalIPForDSR(serviceInfoMap serviceInfoMap) error {
return nil return nil
} }
// unique identfier for a load-balanced service (namespace + name + portname) // unique identifier for a load-balanced service (namespace + name + portname)
func generateServiceId(namespace, svcName, port string) string { func generateServiceId(namespace, svcName, port string) string {
return namespace + "-" + svcName + "-" + port return namespace + "-" + svcName + "-" + port
} }
// unique identfier for a load-balanced service (namespace + name + portname) // unique identifier for a load-balanced service (namespace + name + portname)
func generateIpPortId(ip, protocol, port string) string { func generateIpPortId(ip, protocol, port string) string {
return ip + "-" + protocol + "-" + port return ip + "-" + protocol + "-" + port
} }
@ -1532,7 +1532,7 @@ func getKubeDummyInterface() (netlink.Link, error) {
dummyVipInterface, err := netlink.LinkByName(KUBE_DUMMY_IF) dummyVipInterface, err := netlink.LinkByName(KUBE_DUMMY_IF)
if err != nil && err.Error() == IFACE_NOT_FOUND { if err != nil && err.Error() == IFACE_NOT_FOUND {
glog.V(1).Infof("Could not find dummy interface: " + KUBE_DUMMY_IF + " to assign cluster ip's, creating one") glog.V(1).Infof("Could not find dummy interface: " + KUBE_DUMMY_IF + " to assign cluster ip's, creating one")
err = netlink.LinkAdd(&netlink.Dummy{netlink.LinkAttrs{Name: KUBE_DUMMY_IF}}) err = netlink.LinkAdd(&netlink.Dummy{LinkAttrs: netlink.LinkAttrs{Name: KUBE_DUMMY_IF}})
if err != nil { if err != nil {
return nil, errors.New("Failed to add dummy interface: " + err.Error()) return nil, errors.New("Failed to add dummy interface: " + err.Error())
} }

View File

@ -9,7 +9,7 @@ import (
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/fields" "k8s.io/apimachinery/pkg/fields"
"k8s.io/client-go/kubernetes" "k8s.io/client-go/kubernetes"
cache "k8s.io/client-go/tools/cache" "k8s.io/client-go/tools/cache"
) )
type Operation int type Operation int
@ -103,7 +103,7 @@ func StartEndpointsWatcher(clientset *kubernetes.Clientset, resyncPeriod time.Du
ew.clientset = clientset ew.clientset = clientset
ew.broadcaster = utils.NewBroadcaster() ew.broadcaster = utils.NewBroadcaster()
lw := cache.NewListWatchFromClient(clientset.Core().RESTClient(), "endpoints", metav1.NamespaceAll, fields.Everything()) lw := cache.NewListWatchFromClient(clientset.CoreV1().RESTClient(), "endpoints", metav1.NamespaceAll, fields.Everything())
ew.endpointsLister, ew.endpointsController = cache.NewIndexerInformer( ew.endpointsLister, ew.endpointsController = cache.NewIndexerInformer(
lw, lw,
&api.Endpoints{}, resyncPeriod, eventHandler, &api.Endpoints{}, resyncPeriod, eventHandler,

View File

@ -11,7 +11,7 @@ import (
"k8s.io/apimachinery/pkg/labels" "k8s.io/apimachinery/pkg/labels"
"k8s.io/client-go/kubernetes" "k8s.io/client-go/kubernetes"
listers "k8s.io/client-go/listers/core/v1" listers "k8s.io/client-go/listers/core/v1"
cache "k8s.io/client-go/tools/cache" "k8s.io/client-go/tools/cache"
) )
type NamespaceUpdate struct { type NamespaceUpdate struct {
@ -97,7 +97,7 @@ func StartNamespaceWatcher(clientset *kubernetes.Clientset, resyncPeriod time.Du
nsw.clientset = clientset nsw.clientset = clientset
nsw.broadcaster = utils.NewBroadcaster() nsw.broadcaster = utils.NewBroadcaster()
lw := cache.NewListWatchFromClient(clientset.Core().RESTClient(), "namespaces", metav1.NamespaceAll, fields.Everything()) lw := cache.NewListWatchFromClient(clientset.CoreV1().RESTClient(), "namespaces", metav1.NamespaceAll, fields.Everything())
nsw.namespaceLister, nsw.namespaceController = cache.NewIndexerInformer( nsw.namespaceLister, nsw.namespaceController = cache.NewIndexerInformer(
lw, lw,
&api.Namespace{}, resyncPeriod, eventHandler, &api.Namespace{}, resyncPeriod, eventHandler,

View File

@ -12,7 +12,7 @@ import (
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/fields" "k8s.io/apimachinery/pkg/fields"
"k8s.io/client-go/kubernetes" "k8s.io/client-go/kubernetes"
cache "k8s.io/client-go/tools/cache" "k8s.io/client-go/tools/cache"
) )
type NetworkPolicyUpdate struct { type NetworkPolicyUpdate struct {
@ -92,13 +92,13 @@ func StartNetworkPolicyWatcher(clientset *kubernetes.Clientset, resyncPeriod tim
npw.broadcaster = utils.NewBroadcaster() npw.broadcaster = utils.NewBroadcaster()
var lw *cache.ListWatch var lw *cache.ListWatch
if v1NetworkPolicy { if v1NetworkPolicy {
lw = cache.NewListWatchFromClient(clientset.Networking().RESTClient(), "networkpolicies", metav1.NamespaceAll, fields.Everything()) lw = cache.NewListWatchFromClient(clientset.NetworkingV1().RESTClient(), "networkpolicies", metav1.NamespaceAll, fields.Everything())
npw.networkPolicyLister, npw.networkPolicyController = cache.NewIndexerInformer( npw.networkPolicyLister, npw.networkPolicyController = cache.NewIndexerInformer(
lw, &networking.NetworkPolicy{}, resyncPeriod, eventHandler, lw, &networking.NetworkPolicy{}, resyncPeriod, eventHandler,
cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc}, cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc},
) )
} else { } else {
lw = cache.NewListWatchFromClient(clientset.Extensions().RESTClient(), "networkpolicies", metav1.NamespaceAll, fields.Everything()) lw = cache.NewListWatchFromClient(clientset.ExtensionsV1beta1().RESTClient(), "networkpolicies", metav1.NamespaceAll, fields.Everything())
npw.networkPolicyLister, npw.networkPolicyController = cache.NewIndexerInformer( npw.networkPolicyLister, npw.networkPolicyController = cache.NewIndexerInformer(
lw, &apiextensions.NetworkPolicy{}, resyncPeriod, eventHandler, lw, &apiextensions.NetworkPolicy{}, resyncPeriod, eventHandler,
cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc}, cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc},

View File

@ -8,7 +8,7 @@ import (
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/fields" "k8s.io/apimachinery/pkg/fields"
"k8s.io/client-go/kubernetes" "k8s.io/client-go/kubernetes"
cache "k8s.io/client-go/tools/cache" "k8s.io/client-go/tools/cache"
) )
type NodeUpdate struct { type NodeUpdate struct {
@ -85,7 +85,7 @@ func StartNodeWatcher(clientset *kubernetes.Clientset, resyncPeriod time.Duratio
nw.clientset = clientset nw.clientset = clientset
nw.broadcaster = utils.NewBroadcaster() nw.broadcaster = utils.NewBroadcaster()
lw := cache.NewListWatchFromClient(clientset.Core().RESTClient(), "nodes", metav1.NamespaceAll, fields.Everything()) lw := cache.NewListWatchFromClient(clientset.CoreV1().RESTClient(), "nodes", metav1.NamespaceAll, fields.Everything())
nw.nodeLister, nw.nodeController = cache.NewIndexerInformer( nw.nodeLister, nw.nodeController = cache.NewIndexerInformer(
lw, lw,
&api.Node{}, resyncPeriod, eventHandler, &api.Node{}, resyncPeriod, eventHandler,

View File

@ -11,7 +11,7 @@ import (
"k8s.io/apimachinery/pkg/labels" "k8s.io/apimachinery/pkg/labels"
"k8s.io/client-go/kubernetes" "k8s.io/client-go/kubernetes"
listers "k8s.io/client-go/listers/core/v1" listers "k8s.io/client-go/listers/core/v1"
cache "k8s.io/client-go/tools/cache" "k8s.io/client-go/tools/cache"
) )
type PodUpdate struct { type PodUpdate struct {
@ -102,7 +102,7 @@ func StartPodWatcher(clientset *kubernetes.Clientset, resyncPeriod time.Duration
pw.clientset = clientset pw.clientset = clientset
pw.broadcaster = utils.NewBroadcaster() pw.broadcaster = utils.NewBroadcaster()
lw := cache.NewListWatchFromClient(clientset.Core().RESTClient(), "pods", metav1.NamespaceAll, fields.Everything()) lw := cache.NewListWatchFromClient(clientset.CoreV1().RESTClient(), "pods", metav1.NamespaceAll, fields.Everything())
pw.podLister, pw.podController = cache.NewIndexerInformer( pw.podLister, pw.podController = cache.NewIndexerInformer(
lw, lw,
&api.Pod{}, resyncPeriod, eventHandler, &api.Pod{}, resyncPeriod, eventHandler,