9306e99368
changed from capnslog to logrus for logging JSON structured message. finished issue #383
373 lines
11 KiB
Go
373 lines
11 KiB
Go
// Copyright 2017 clair authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package clair
|
|
|
|
import (
|
|
"math/rand"
|
|
"strconv"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/pborman/uuid"
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
log "github.com/sirupsen/logrus"
|
|
|
|
"github.com/coreos/clair/database"
|
|
"github.com/coreos/clair/ext/vulnmdsrc"
|
|
"github.com/coreos/clair/ext/vulnsrc"
|
|
"github.com/coreos/clair/pkg/stopper"
|
|
)
|
|
|
|
const (
|
|
updaterLastFlagName = "updater/last"
|
|
updaterLockName = "updater"
|
|
updaterLockDuration = updaterLockRefreshDuration + time.Minute*2
|
|
updaterLockRefreshDuration = time.Minute * 8
|
|
)
|
|
|
|
var (
|
|
promUpdaterErrorsTotal = prometheus.NewCounter(prometheus.CounterOpts{
|
|
Name: "clair_updater_errors_total",
|
|
Help: "Numbers of errors that the updater generated.",
|
|
})
|
|
|
|
promUpdaterDurationSeconds = prometheus.NewGauge(prometheus.GaugeOpts{
|
|
Name: "clair_updater_duration_seconds",
|
|
Help: "Time it takes to update the vulnerability database.",
|
|
})
|
|
|
|
promUpdaterNotesTotal = prometheus.NewGauge(prometheus.GaugeOpts{
|
|
Name: "clair_updater_notes_total",
|
|
Help: "Number of notes that the vulnerability fetchers generated.",
|
|
})
|
|
)
|
|
|
|
func init() {
|
|
prometheus.MustRegister(promUpdaterErrorsTotal)
|
|
prometheus.MustRegister(promUpdaterDurationSeconds)
|
|
prometheus.MustRegister(promUpdaterNotesTotal)
|
|
}
|
|
|
|
// UpdaterConfig is the configuration for the Updater service.
|
|
type UpdaterConfig struct {
|
|
Interval time.Duration
|
|
}
|
|
|
|
// RunUpdater begins a process that updates the vulnerability database at
|
|
// regular intervals.
|
|
func RunUpdater(config *UpdaterConfig, datastore database.Datastore, st *stopper.Stopper) {
|
|
defer st.End()
|
|
|
|
// Do not run the updater if there is no config or if the interval is 0.
|
|
if config == nil || config.Interval == 0 {
|
|
log.Info("updater service is disabled.")
|
|
return
|
|
}
|
|
|
|
whoAmI := uuid.New()
|
|
log.WithField("lock identifier", whoAmI).Info("updater service started")
|
|
|
|
for {
|
|
var stop bool
|
|
|
|
// Determine if this is the first update and define the next update time.
|
|
// The next update time is (last update time + interval) or now if this is the first update.
|
|
nextUpdate := time.Now().UTC()
|
|
lastUpdate, firstUpdate, err := getLastUpdate(datastore)
|
|
if err != nil {
|
|
log.WithError(err).Error("an error occured while getting the last update time")
|
|
nextUpdate = nextUpdate.Add(config.Interval)
|
|
} else if firstUpdate == false {
|
|
nextUpdate = lastUpdate.Add(config.Interval)
|
|
}
|
|
|
|
// If the next update timer is in the past, then try to update.
|
|
if nextUpdate.Before(time.Now().UTC()) {
|
|
// Attempt to get a lock on the the update.
|
|
log.Debug("attempting to obtain update lock")
|
|
hasLock, hasLockUntil := datastore.Lock(updaterLockName, whoAmI, updaterLockDuration, false)
|
|
if hasLock {
|
|
// Launch update in a new go routine.
|
|
doneC := make(chan bool, 1)
|
|
go func() {
|
|
update(datastore, firstUpdate)
|
|
doneC <- true
|
|
}()
|
|
|
|
for done := false; !done && !stop; {
|
|
select {
|
|
case <-doneC:
|
|
done = true
|
|
case <-time.After(updaterLockRefreshDuration):
|
|
// Refresh the lock until the update is done.
|
|
datastore.Lock(updaterLockName, whoAmI, updaterLockDuration, true)
|
|
case <-st.Chan():
|
|
stop = true
|
|
}
|
|
}
|
|
|
|
// Unlock the update.
|
|
datastore.Unlock(updaterLockName, whoAmI)
|
|
|
|
if stop {
|
|
break
|
|
}
|
|
continue
|
|
} else {
|
|
lockOwner, lockExpiration, err := datastore.FindLock(updaterLockName)
|
|
if err != nil {
|
|
log.Debug("update lock is already taken")
|
|
nextUpdate = hasLockUntil
|
|
} else {
|
|
log.WithFields(log.Fields{"lock owner": lockOwner, "lock expiration": lockExpiration}).Debug("update lock is already taken")
|
|
nextUpdate = lockExpiration
|
|
}
|
|
}
|
|
}
|
|
|
|
// Sleep, but remain stoppable until approximately the next update time.
|
|
now := time.Now().UTC()
|
|
waitUntil := nextUpdate.Add(time.Duration(rand.ExpFloat64()/0.5) * time.Second)
|
|
log.WithField("scheduled time", waitUntil).Debug("next update attempt scheduled")
|
|
if !waitUntil.Before(now) {
|
|
if !st.Sleep(waitUntil.Sub(time.Now())) {
|
|
break
|
|
}
|
|
}
|
|
}
|
|
|
|
// Clean resources.
|
|
for _, appenders := range vulnmdsrc.Appenders() {
|
|
appenders.Clean()
|
|
}
|
|
for _, updaters := range vulnsrc.Updaters() {
|
|
updaters.Clean()
|
|
}
|
|
|
|
log.Info("updater service stopped")
|
|
}
|
|
|
|
// update fetches all the vulnerabilities from the registered fetchers, upserts
|
|
// them into the database and then sends notifications.
|
|
func update(datastore database.Datastore, firstUpdate bool) {
|
|
defer setUpdaterDuration(time.Now())
|
|
|
|
log.Info("updating vulnerabilities")
|
|
|
|
// Fetch updates.
|
|
status, vulnerabilities, flags, notes := fetch(datastore)
|
|
|
|
// Insert vulnerabilities.
|
|
log.WithField("count", len(vulnerabilities)).Debug("inserting vulnerabilities for update")
|
|
err := datastore.InsertVulnerabilities(vulnerabilities, !firstUpdate)
|
|
if err != nil {
|
|
promUpdaterErrorsTotal.Inc()
|
|
log.WithError(err).Error("an error occured when inserting vulnerabilities for update")
|
|
return
|
|
}
|
|
vulnerabilities = nil
|
|
|
|
// Update flags.
|
|
for flagName, flagValue := range flags {
|
|
datastore.InsertKeyValue(flagName, flagValue)
|
|
}
|
|
|
|
// Log notes.
|
|
for _, note := range notes {
|
|
log.WithField("note", note).Warning("fetcher note")
|
|
}
|
|
promUpdaterNotesTotal.Set(float64(len(notes)))
|
|
|
|
// Update last successful update if every fetchers worked properly.
|
|
if status {
|
|
datastore.InsertKeyValue(updaterLastFlagName, strconv.FormatInt(time.Now().UTC().Unix(), 10))
|
|
}
|
|
|
|
log.Info("update finished")
|
|
}
|
|
|
|
func setUpdaterDuration(start time.Time) {
|
|
promUpdaterDurationSeconds.Set(time.Since(start).Seconds())
|
|
}
|
|
|
|
// fetch get data from the registered fetchers, in parallel.
|
|
func fetch(datastore database.Datastore) (bool, []database.Vulnerability, map[string]string, []string) {
|
|
var vulnerabilities []database.Vulnerability
|
|
var notes []string
|
|
status := true
|
|
flags := make(map[string]string)
|
|
|
|
// Fetch updates in parallel.
|
|
log.Info("fetching vulnerability updates")
|
|
var responseC = make(chan *vulnsrc.UpdateResponse, 0)
|
|
for n, u := range vulnsrc.Updaters() {
|
|
go func(name string, u vulnsrc.Updater) {
|
|
response, err := u.Update(datastore)
|
|
if err != nil {
|
|
promUpdaterErrorsTotal.Inc()
|
|
log.WithError(err).WithField("updater name", name).Error("an error occured when fetching update")
|
|
status = false
|
|
responseC <- nil
|
|
return
|
|
}
|
|
|
|
responseC <- &response
|
|
log.WithField("updater name", name).Info("finished fetching")
|
|
}(n, u)
|
|
}
|
|
|
|
// Collect results of updates.
|
|
for i := 0; i < len(vulnsrc.Updaters()); i++ {
|
|
resp := <-responseC
|
|
if resp != nil {
|
|
vulnerabilities = append(vulnerabilities, doVulnerabilitiesNamespacing(resp.Vulnerabilities)...)
|
|
notes = append(notes, resp.Notes...)
|
|
if resp.FlagName != "" && resp.FlagValue != "" {
|
|
flags[resp.FlagName] = resp.FlagValue
|
|
}
|
|
}
|
|
}
|
|
|
|
close(responseC)
|
|
return status, addMetadata(datastore, vulnerabilities), flags, notes
|
|
}
|
|
|
|
// Add metadata to the specified vulnerabilities using the registered MetadataFetchers, in parallel.
|
|
func addMetadata(datastore database.Datastore, vulnerabilities []database.Vulnerability) []database.Vulnerability {
|
|
if len(vulnmdsrc.Appenders()) == 0 {
|
|
return vulnerabilities
|
|
}
|
|
|
|
log.Info("adding metadata to vulnerabilities")
|
|
|
|
// Add a mutex to each vulnerability to ensure that only one appender at a
|
|
// time can modify the vulnerability's Metadata map.
|
|
lockableVulnerabilities := make([]*lockableVulnerability, 0, len(vulnerabilities))
|
|
for i := 0; i < len(vulnerabilities); i++ {
|
|
lockableVulnerabilities = append(lockableVulnerabilities, &lockableVulnerability{
|
|
Vulnerability: &vulnerabilities[i],
|
|
})
|
|
}
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(len(vulnmdsrc.Appenders()))
|
|
|
|
for n, a := range vulnmdsrc.Appenders() {
|
|
go func(name string, appender vulnmdsrc.Appender) {
|
|
defer wg.Done()
|
|
|
|
// Build up a metadata cache.
|
|
if err := appender.BuildCache(datastore); err != nil {
|
|
promUpdaterErrorsTotal.Inc()
|
|
log.WithError(err).WithField("appender name", name).Error("an error occured when loading metadata fetcher")
|
|
return
|
|
}
|
|
|
|
// Append vulnerability metadata to each vulnerability.
|
|
for _, vulnerability := range lockableVulnerabilities {
|
|
appender.Append(vulnerability.Name, vulnerability.appendFunc)
|
|
}
|
|
|
|
// Purge the metadata cache.
|
|
appender.PurgeCache()
|
|
}(n, a)
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
return vulnerabilities
|
|
}
|
|
|
|
func getLastUpdate(datastore database.Datastore) (time.Time, bool, error) {
|
|
lastUpdateTSS, err := datastore.GetKeyValue(updaterLastFlagName)
|
|
if err != nil {
|
|
return time.Time{}, false, err
|
|
}
|
|
|
|
if lastUpdateTSS == "" {
|
|
// This is the first update.
|
|
return time.Time{}, true, nil
|
|
}
|
|
|
|
lastUpdateTS, err := strconv.ParseInt(lastUpdateTSS, 10, 64)
|
|
if err != nil {
|
|
return time.Time{}, false, err
|
|
}
|
|
|
|
return time.Unix(lastUpdateTS, 0).UTC(), false, nil
|
|
}
|
|
|
|
type lockableVulnerability struct {
|
|
*database.Vulnerability
|
|
sync.Mutex
|
|
}
|
|
|
|
func (lv *lockableVulnerability) appendFunc(metadataKey string, metadata interface{}, severity database.Severity) {
|
|
lv.Lock()
|
|
defer lv.Unlock()
|
|
|
|
// If necessary, initialize the metadata map for the vulnerability.
|
|
if lv.Metadata == nil {
|
|
lv.Metadata = make(map[string]interface{})
|
|
}
|
|
|
|
// Append the metadata.
|
|
lv.Metadata[metadataKey] = metadata
|
|
|
|
// If necessary, provide a severity for the vulnerability.
|
|
if lv.Severity == database.UnknownSeverity {
|
|
lv.Severity = severity
|
|
}
|
|
}
|
|
|
|
// doVulnerabilitiesNamespacing takes Vulnerabilities that don't have a
|
|
// Namespace and split them into multiple vulnerabilities that have a Namespace
|
|
// and only contains the FixedIn FeatureVersions corresponding to their
|
|
// Namespace.
|
|
//
|
|
// It helps simplifying the fetchers that share the same metadata about a
|
|
// Vulnerability regardless of their actual namespace (ie. same vulnerability
|
|
// information for every version of a distro).
|
|
func doVulnerabilitiesNamespacing(vulnerabilities []database.Vulnerability) []database.Vulnerability {
|
|
vulnerabilitiesMap := make(map[string]*database.Vulnerability)
|
|
|
|
for _, v := range vulnerabilities {
|
|
featureVersions := v.FixedIn
|
|
v.FixedIn = []database.FeatureVersion{}
|
|
|
|
for _, fv := range featureVersions {
|
|
index := fv.Feature.Namespace.Name + ":" + v.Name
|
|
|
|
if vulnerability, ok := vulnerabilitiesMap[index]; !ok {
|
|
newVulnerability := v
|
|
newVulnerability.Namespace = fv.Feature.Namespace
|
|
newVulnerability.FixedIn = []database.FeatureVersion{fv}
|
|
|
|
vulnerabilitiesMap[index] = &newVulnerability
|
|
} else {
|
|
vulnerability.FixedIn = append(vulnerability.FixedIn, fv)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Convert map into a slice.
|
|
var response []database.Vulnerability
|
|
for _, vulnerability := range vulnerabilitiesMap {
|
|
response = append(response, *vulnerability)
|
|
}
|
|
|
|
return response
|
|
}
|