mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-12-27 15:30:23 +00:00
171ba6a65d
Bumps the github-dependencies group with 8 updates in the / directory: | Package | From | To | | --- | --- | --- | | [github.com/IBM/keyprotect-go-client](https://github.com/IBM/keyprotect-go-client) | `0.12.2` | `0.14.1` | | [github.com/aws/aws-sdk-go](https://github.com/aws/aws-sdk-go) | `1.53.14` | `1.54.6` | | [github.com/aws/aws-sdk-go-v2/service/sts](https://github.com/aws/aws-sdk-go-v2) | `1.28.1` | `1.29.1` | | [github.com/hashicorp/vault/api](https://github.com/hashicorp/vault) | `1.12.0` | `1.14.0` | | [github.com/kubernetes-csi/csi-lib-utils](https://github.com/kubernetes-csi/csi-lib-utils) | `0.17.0` | `0.18.1` | | [github.com/onsi/ginkgo/v2](https://github.com/onsi/ginkgo) | `2.17.1` | `2.19.0` | | [github.com/prometheus/client_golang](https://github.com/prometheus/client_golang) | `1.18.0` | `1.19.1` | | [github.com/Azure/azure-sdk-for-go/sdk/azidentity](https://github.com/Azure/azure-sdk-for-go) | `1.6.0` | `1.7.0` | Updates `github.com/IBM/keyprotect-go-client` from 0.12.2 to 0.14.1 - [Release notes](https://github.com/IBM/keyprotect-go-client/releases) - [Changelog](https://github.com/IBM/keyprotect-go-client/blob/master/CHANGELOG.md) - [Commits](https://github.com/IBM/keyprotect-go-client/compare/v0.12.2...v0.14.1) Updates `github.com/aws/aws-sdk-go` from 1.53.14 to 1.54.6 - [Release notes](https://github.com/aws/aws-sdk-go/releases) - [Commits](https://github.com/aws/aws-sdk-go/compare/v1.53.14...v1.54.6) Updates `github.com/aws/aws-sdk-go-v2/service/sts` from 1.28.1 to 1.29.1 - [Release notes](https://github.com/aws/aws-sdk-go-v2/releases) - [Commits](https://github.com/aws/aws-sdk-go-v2/compare/service/ecr/v1.28.1...service/s3/v1.29.1) Updates `github.com/hashicorp/vault/api` from 1.12.0 to 1.14.0 - [Release notes](https://github.com/hashicorp/vault/releases) - [Changelog](https://github.com/hashicorp/vault/blob/main/CHANGELOG.md) - [Commits](https://github.com/hashicorp/vault/compare/v1.12.0...v1.14.0) Updates `github.com/kubernetes-csi/csi-lib-utils` from 0.17.0 to 0.18.1 - [Release notes](https://github.com/kubernetes-csi/csi-lib-utils/releases) - [Commits](https://github.com/kubernetes-csi/csi-lib-utils/compare/v0.17.0...v0.18.1) Updates `github.com/onsi/ginkgo/v2` from 2.17.1 to 2.19.0 - [Release notes](https://github.com/onsi/ginkgo/releases) - [Changelog](https://github.com/onsi/ginkgo/blob/master/CHANGELOG.md) - [Commits](https://github.com/onsi/ginkgo/compare/v2.17.1...v2.19.0) Updates `github.com/onsi/gomega` from 1.32.0 to 1.33.1 - [Release notes](https://github.com/onsi/gomega/releases) - [Changelog](https://github.com/onsi/gomega/blob/master/CHANGELOG.md) - [Commits](https://github.com/onsi/gomega/compare/v1.32.0...v1.33.1) Updates `github.com/prometheus/client_golang` from 1.18.0 to 1.19.1 - [Release notes](https://github.com/prometheus/client_golang/releases) - [Changelog](https://github.com/prometheus/client_golang/blob/main/CHANGELOG.md) - [Commits](https://github.com/prometheus/client_golang/compare/v1.18.0...v1.19.1) Updates `github.com/Azure/azure-sdk-for-go/sdk/azidentity` from 1.6.0 to 1.7.0 - [Release notes](https://github.com/Azure/azure-sdk-for-go/releases) - [Changelog](https://github.com/Azure/azure-sdk-for-go/blob/main/documentation/release.md) - [Commits](https://github.com/Azure/azure-sdk-for-go/compare/sdk/azcore/v1.6.0...sdk/azcore/v1.7.0) --- updated-dependencies: - dependency-name: github.com/IBM/keyprotect-go-client dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies - dependency-name: github.com/aws/aws-sdk-go dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies - dependency-name: github.com/aws/aws-sdk-go-v2/service/sts dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies - dependency-name: github.com/hashicorp/vault/api dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies - dependency-name: github.com/kubernetes-csi/csi-lib-utils dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies - dependency-name: github.com/onsi/ginkgo/v2 dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies - dependency-name: github.com/onsi/gomega dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies - dependency-name: github.com/prometheus/client_golang dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies - dependency-name: github.com/Azure/azure-sdk-for-go/sdk/azidentity dependency-type: direct:production update-type: version-update:semver-minor dependency-group: github-dependencies ... Signed-off-by: dependabot[bot] <support@github.com>
670 lines
17 KiB
Go
670 lines
17 KiB
Go
// Copyright 2014 Google Inc. All Rights Reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package profile
|
|
|
|
import (
|
|
"encoding/binary"
|
|
"fmt"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
)
|
|
|
|
// Compact performs garbage collection on a profile to remove any
|
|
// unreferenced fields. This is useful to reduce the size of a profile
|
|
// after samples or locations have been removed.
|
|
func (p *Profile) Compact() *Profile {
|
|
p, _ = Merge([]*Profile{p})
|
|
return p
|
|
}
|
|
|
|
// Merge merges all the profiles in profs into a single Profile.
|
|
// Returns a new profile independent of the input profiles. The merged
|
|
// profile is compacted to eliminate unused samples, locations,
|
|
// functions and mappings. Profiles must have identical profile sample
|
|
// and period types or the merge will fail. profile.Period of the
|
|
// resulting profile will be the maximum of all profiles, and
|
|
// profile.TimeNanos will be the earliest nonzero one. Merges are
|
|
// associative with the caveat of the first profile having some
|
|
// specialization in how headers are combined. There may be other
|
|
// subtleties now or in the future regarding associativity.
|
|
func Merge(srcs []*Profile) (*Profile, error) {
|
|
if len(srcs) == 0 {
|
|
return nil, fmt.Errorf("no profiles to merge")
|
|
}
|
|
p, err := combineHeaders(srcs)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
pm := &profileMerger{
|
|
p: p,
|
|
samples: make(map[sampleKey]*Sample, len(srcs[0].Sample)),
|
|
locations: make(map[locationKey]*Location, len(srcs[0].Location)),
|
|
functions: make(map[functionKey]*Function, len(srcs[0].Function)),
|
|
mappings: make(map[mappingKey]*Mapping, len(srcs[0].Mapping)),
|
|
}
|
|
|
|
for _, src := range srcs {
|
|
// Clear the profile-specific hash tables
|
|
pm.locationsByID = makeLocationIDMap(len(src.Location))
|
|
pm.functionsByID = make(map[uint64]*Function, len(src.Function))
|
|
pm.mappingsByID = make(map[uint64]mapInfo, len(src.Mapping))
|
|
|
|
if len(pm.mappings) == 0 && len(src.Mapping) > 0 {
|
|
// The Mapping list has the property that the first mapping
|
|
// represents the main binary. Take the first Mapping we see,
|
|
// otherwise the operations below will add mappings in an
|
|
// arbitrary order.
|
|
pm.mapMapping(src.Mapping[0])
|
|
}
|
|
|
|
for _, s := range src.Sample {
|
|
if !isZeroSample(s) {
|
|
pm.mapSample(s)
|
|
}
|
|
}
|
|
}
|
|
|
|
for _, s := range p.Sample {
|
|
if isZeroSample(s) {
|
|
// If there are any zero samples, re-merge the profile to GC
|
|
// them.
|
|
return Merge([]*Profile{p})
|
|
}
|
|
}
|
|
|
|
return p, nil
|
|
}
|
|
|
|
// Normalize normalizes the source profile by multiplying each value in profile by the
|
|
// ratio of the sum of the base profile's values of that sample type to the sum of the
|
|
// source profile's value of that sample type.
|
|
func (p *Profile) Normalize(pb *Profile) error {
|
|
|
|
if err := p.compatible(pb); err != nil {
|
|
return err
|
|
}
|
|
|
|
baseVals := make([]int64, len(p.SampleType))
|
|
for _, s := range pb.Sample {
|
|
for i, v := range s.Value {
|
|
baseVals[i] += v
|
|
}
|
|
}
|
|
|
|
srcVals := make([]int64, len(p.SampleType))
|
|
for _, s := range p.Sample {
|
|
for i, v := range s.Value {
|
|
srcVals[i] += v
|
|
}
|
|
}
|
|
|
|
normScale := make([]float64, len(baseVals))
|
|
for i := range baseVals {
|
|
if srcVals[i] == 0 {
|
|
normScale[i] = 0.0
|
|
} else {
|
|
normScale[i] = float64(baseVals[i]) / float64(srcVals[i])
|
|
}
|
|
}
|
|
p.ScaleN(normScale)
|
|
return nil
|
|
}
|
|
|
|
func isZeroSample(s *Sample) bool {
|
|
for _, v := range s.Value {
|
|
if v != 0 {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
type profileMerger struct {
|
|
p *Profile
|
|
|
|
// Memoization tables within a profile.
|
|
locationsByID locationIDMap
|
|
functionsByID map[uint64]*Function
|
|
mappingsByID map[uint64]mapInfo
|
|
|
|
// Memoization tables for profile entities.
|
|
samples map[sampleKey]*Sample
|
|
locations map[locationKey]*Location
|
|
functions map[functionKey]*Function
|
|
mappings map[mappingKey]*Mapping
|
|
}
|
|
|
|
type mapInfo struct {
|
|
m *Mapping
|
|
offset int64
|
|
}
|
|
|
|
func (pm *profileMerger) mapSample(src *Sample) *Sample {
|
|
// Check memoization table
|
|
k := pm.sampleKey(src)
|
|
if ss, ok := pm.samples[k]; ok {
|
|
for i, v := range src.Value {
|
|
ss.Value[i] += v
|
|
}
|
|
return ss
|
|
}
|
|
|
|
// Make new sample.
|
|
s := &Sample{
|
|
Location: make([]*Location, len(src.Location)),
|
|
Value: make([]int64, len(src.Value)),
|
|
Label: make(map[string][]string, len(src.Label)),
|
|
NumLabel: make(map[string][]int64, len(src.NumLabel)),
|
|
NumUnit: make(map[string][]string, len(src.NumLabel)),
|
|
}
|
|
for i, l := range src.Location {
|
|
s.Location[i] = pm.mapLocation(l)
|
|
}
|
|
for k, v := range src.Label {
|
|
vv := make([]string, len(v))
|
|
copy(vv, v)
|
|
s.Label[k] = vv
|
|
}
|
|
for k, v := range src.NumLabel {
|
|
u := src.NumUnit[k]
|
|
vv := make([]int64, len(v))
|
|
uu := make([]string, len(u))
|
|
copy(vv, v)
|
|
copy(uu, u)
|
|
s.NumLabel[k] = vv
|
|
s.NumUnit[k] = uu
|
|
}
|
|
copy(s.Value, src.Value)
|
|
pm.samples[k] = s
|
|
pm.p.Sample = append(pm.p.Sample, s)
|
|
return s
|
|
}
|
|
|
|
func (pm *profileMerger) sampleKey(sample *Sample) sampleKey {
|
|
// Accumulate contents into a string.
|
|
var buf strings.Builder
|
|
buf.Grow(64) // Heuristic to avoid extra allocs
|
|
|
|
// encode a number
|
|
putNumber := func(v uint64) {
|
|
var num [binary.MaxVarintLen64]byte
|
|
n := binary.PutUvarint(num[:], v)
|
|
buf.Write(num[:n])
|
|
}
|
|
|
|
// encode a string prefixed with its length.
|
|
putDelimitedString := func(s string) {
|
|
putNumber(uint64(len(s)))
|
|
buf.WriteString(s)
|
|
}
|
|
|
|
for _, l := range sample.Location {
|
|
// Get the location in the merged profile, which may have a different ID.
|
|
if loc := pm.mapLocation(l); loc != nil {
|
|
putNumber(loc.ID)
|
|
}
|
|
}
|
|
putNumber(0) // Delimiter
|
|
|
|
for _, l := range sortedKeys1(sample.Label) {
|
|
putDelimitedString(l)
|
|
values := sample.Label[l]
|
|
putNumber(uint64(len(values)))
|
|
for _, v := range values {
|
|
putDelimitedString(v)
|
|
}
|
|
}
|
|
|
|
for _, l := range sortedKeys2(sample.NumLabel) {
|
|
putDelimitedString(l)
|
|
values := sample.NumLabel[l]
|
|
putNumber(uint64(len(values)))
|
|
for _, v := range values {
|
|
putNumber(uint64(v))
|
|
}
|
|
units := sample.NumUnit[l]
|
|
putNumber(uint64(len(units)))
|
|
for _, v := range units {
|
|
putDelimitedString(v)
|
|
}
|
|
}
|
|
|
|
return sampleKey(buf.String())
|
|
}
|
|
|
|
type sampleKey string
|
|
|
|
// sortedKeys1 returns the sorted keys found in a string->[]string map.
|
|
//
|
|
// Note: this is currently non-generic since github pprof runs golint,
|
|
// which does not support generics. When that issue is fixed, it can
|
|
// be merged with sortedKeys2 and made into a generic function.
|
|
func sortedKeys1(m map[string][]string) []string {
|
|
if len(m) == 0 {
|
|
return nil
|
|
}
|
|
keys := make([]string, 0, len(m))
|
|
for k := range m {
|
|
keys = append(keys, k)
|
|
}
|
|
sort.Strings(keys)
|
|
return keys
|
|
}
|
|
|
|
// sortedKeys2 returns the sorted keys found in a string->[]int64 map.
|
|
//
|
|
// Note: this is currently non-generic since github pprof runs golint,
|
|
// which does not support generics. When that issue is fixed, it can
|
|
// be merged with sortedKeys1 and made into a generic function.
|
|
func sortedKeys2(m map[string][]int64) []string {
|
|
if len(m) == 0 {
|
|
return nil
|
|
}
|
|
keys := make([]string, 0, len(m))
|
|
for k := range m {
|
|
keys = append(keys, k)
|
|
}
|
|
sort.Strings(keys)
|
|
return keys
|
|
}
|
|
|
|
func (pm *profileMerger) mapLocation(src *Location) *Location {
|
|
if src == nil {
|
|
return nil
|
|
}
|
|
|
|
if l := pm.locationsByID.get(src.ID); l != nil {
|
|
return l
|
|
}
|
|
|
|
mi := pm.mapMapping(src.Mapping)
|
|
l := &Location{
|
|
ID: uint64(len(pm.p.Location) + 1),
|
|
Mapping: mi.m,
|
|
Address: uint64(int64(src.Address) + mi.offset),
|
|
Line: make([]Line, len(src.Line)),
|
|
IsFolded: src.IsFolded,
|
|
}
|
|
for i, ln := range src.Line {
|
|
l.Line[i] = pm.mapLine(ln)
|
|
}
|
|
// Check memoization table. Must be done on the remapped location to
|
|
// account for the remapped mapping ID.
|
|
k := l.key()
|
|
if ll, ok := pm.locations[k]; ok {
|
|
pm.locationsByID.set(src.ID, ll)
|
|
return ll
|
|
}
|
|
pm.locationsByID.set(src.ID, l)
|
|
pm.locations[k] = l
|
|
pm.p.Location = append(pm.p.Location, l)
|
|
return l
|
|
}
|
|
|
|
// key generates locationKey to be used as a key for maps.
|
|
func (l *Location) key() locationKey {
|
|
key := locationKey{
|
|
addr: l.Address,
|
|
isFolded: l.IsFolded,
|
|
}
|
|
if l.Mapping != nil {
|
|
// Normalizes address to handle address space randomization.
|
|
key.addr -= l.Mapping.Start
|
|
key.mappingID = l.Mapping.ID
|
|
}
|
|
lines := make([]string, len(l.Line)*3)
|
|
for i, line := range l.Line {
|
|
if line.Function != nil {
|
|
lines[i*2] = strconv.FormatUint(line.Function.ID, 16)
|
|
}
|
|
lines[i*2+1] = strconv.FormatInt(line.Line, 16)
|
|
lines[i*2+2] = strconv.FormatInt(line.Column, 16)
|
|
}
|
|
key.lines = strings.Join(lines, "|")
|
|
return key
|
|
}
|
|
|
|
type locationKey struct {
|
|
addr, mappingID uint64
|
|
lines string
|
|
isFolded bool
|
|
}
|
|
|
|
func (pm *profileMerger) mapMapping(src *Mapping) mapInfo {
|
|
if src == nil {
|
|
return mapInfo{}
|
|
}
|
|
|
|
if mi, ok := pm.mappingsByID[src.ID]; ok {
|
|
return mi
|
|
}
|
|
|
|
// Check memoization tables.
|
|
mk := src.key()
|
|
if m, ok := pm.mappings[mk]; ok {
|
|
mi := mapInfo{m, int64(m.Start) - int64(src.Start)}
|
|
pm.mappingsByID[src.ID] = mi
|
|
return mi
|
|
}
|
|
m := &Mapping{
|
|
ID: uint64(len(pm.p.Mapping) + 1),
|
|
Start: src.Start,
|
|
Limit: src.Limit,
|
|
Offset: src.Offset,
|
|
File: src.File,
|
|
KernelRelocationSymbol: src.KernelRelocationSymbol,
|
|
BuildID: src.BuildID,
|
|
HasFunctions: src.HasFunctions,
|
|
HasFilenames: src.HasFilenames,
|
|
HasLineNumbers: src.HasLineNumbers,
|
|
HasInlineFrames: src.HasInlineFrames,
|
|
}
|
|
pm.p.Mapping = append(pm.p.Mapping, m)
|
|
|
|
// Update memoization tables.
|
|
pm.mappings[mk] = m
|
|
mi := mapInfo{m, 0}
|
|
pm.mappingsByID[src.ID] = mi
|
|
return mi
|
|
}
|
|
|
|
// key generates encoded strings of Mapping to be used as a key for
|
|
// maps.
|
|
func (m *Mapping) key() mappingKey {
|
|
// Normalize addresses to handle address space randomization.
|
|
// Round up to next 4K boundary to avoid minor discrepancies.
|
|
const mapsizeRounding = 0x1000
|
|
|
|
size := m.Limit - m.Start
|
|
size = size + mapsizeRounding - 1
|
|
size = size - (size % mapsizeRounding)
|
|
key := mappingKey{
|
|
size: size,
|
|
offset: m.Offset,
|
|
}
|
|
|
|
switch {
|
|
case m.BuildID != "":
|
|
key.buildIDOrFile = m.BuildID
|
|
case m.File != "":
|
|
key.buildIDOrFile = m.File
|
|
default:
|
|
// A mapping containing neither build ID nor file name is a fake mapping. A
|
|
// key with empty buildIDOrFile is used for fake mappings so that they are
|
|
// treated as the same mapping during merging.
|
|
}
|
|
return key
|
|
}
|
|
|
|
type mappingKey struct {
|
|
size, offset uint64
|
|
buildIDOrFile string
|
|
}
|
|
|
|
func (pm *profileMerger) mapLine(src Line) Line {
|
|
ln := Line{
|
|
Function: pm.mapFunction(src.Function),
|
|
Line: src.Line,
|
|
Column: src.Column,
|
|
}
|
|
return ln
|
|
}
|
|
|
|
func (pm *profileMerger) mapFunction(src *Function) *Function {
|
|
if src == nil {
|
|
return nil
|
|
}
|
|
if f, ok := pm.functionsByID[src.ID]; ok {
|
|
return f
|
|
}
|
|
k := src.key()
|
|
if f, ok := pm.functions[k]; ok {
|
|
pm.functionsByID[src.ID] = f
|
|
return f
|
|
}
|
|
f := &Function{
|
|
ID: uint64(len(pm.p.Function) + 1),
|
|
Name: src.Name,
|
|
SystemName: src.SystemName,
|
|
Filename: src.Filename,
|
|
StartLine: src.StartLine,
|
|
}
|
|
pm.functions[k] = f
|
|
pm.functionsByID[src.ID] = f
|
|
pm.p.Function = append(pm.p.Function, f)
|
|
return f
|
|
}
|
|
|
|
// key generates a struct to be used as a key for maps.
|
|
func (f *Function) key() functionKey {
|
|
return functionKey{
|
|
f.StartLine,
|
|
f.Name,
|
|
f.SystemName,
|
|
f.Filename,
|
|
}
|
|
}
|
|
|
|
type functionKey struct {
|
|
startLine int64
|
|
name, systemName, fileName string
|
|
}
|
|
|
|
// combineHeaders checks that all profiles can be merged and returns
|
|
// their combined profile.
|
|
func combineHeaders(srcs []*Profile) (*Profile, error) {
|
|
for _, s := range srcs[1:] {
|
|
if err := srcs[0].compatible(s); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
var timeNanos, durationNanos, period int64
|
|
var comments []string
|
|
seenComments := map[string]bool{}
|
|
var defaultSampleType string
|
|
for _, s := range srcs {
|
|
if timeNanos == 0 || s.TimeNanos < timeNanos {
|
|
timeNanos = s.TimeNanos
|
|
}
|
|
durationNanos += s.DurationNanos
|
|
if period == 0 || period < s.Period {
|
|
period = s.Period
|
|
}
|
|
for _, c := range s.Comments {
|
|
if seen := seenComments[c]; !seen {
|
|
comments = append(comments, c)
|
|
seenComments[c] = true
|
|
}
|
|
}
|
|
if defaultSampleType == "" {
|
|
defaultSampleType = s.DefaultSampleType
|
|
}
|
|
}
|
|
|
|
p := &Profile{
|
|
SampleType: make([]*ValueType, len(srcs[0].SampleType)),
|
|
|
|
DropFrames: srcs[0].DropFrames,
|
|
KeepFrames: srcs[0].KeepFrames,
|
|
|
|
TimeNanos: timeNanos,
|
|
DurationNanos: durationNanos,
|
|
PeriodType: srcs[0].PeriodType,
|
|
Period: period,
|
|
|
|
Comments: comments,
|
|
DefaultSampleType: defaultSampleType,
|
|
}
|
|
copy(p.SampleType, srcs[0].SampleType)
|
|
return p, nil
|
|
}
|
|
|
|
// compatible determines if two profiles can be compared/merged.
|
|
// returns nil if the profiles are compatible; otherwise an error with
|
|
// details on the incompatibility.
|
|
func (p *Profile) compatible(pb *Profile) error {
|
|
if !equalValueType(p.PeriodType, pb.PeriodType) {
|
|
return fmt.Errorf("incompatible period types %v and %v", p.PeriodType, pb.PeriodType)
|
|
}
|
|
|
|
if len(p.SampleType) != len(pb.SampleType) {
|
|
return fmt.Errorf("incompatible sample types %v and %v", p.SampleType, pb.SampleType)
|
|
}
|
|
|
|
for i := range p.SampleType {
|
|
if !equalValueType(p.SampleType[i], pb.SampleType[i]) {
|
|
return fmt.Errorf("incompatible sample types %v and %v", p.SampleType, pb.SampleType)
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// equalValueType returns true if the two value types are semantically
|
|
// equal. It ignores the internal fields used during encode/decode.
|
|
func equalValueType(st1, st2 *ValueType) bool {
|
|
return st1.Type == st2.Type && st1.Unit == st2.Unit
|
|
}
|
|
|
|
// locationIDMap is like a map[uint64]*Location, but provides efficiency for
|
|
// ids that are densely numbered, which is often the case.
|
|
type locationIDMap struct {
|
|
dense []*Location // indexed by id for id < len(dense)
|
|
sparse map[uint64]*Location // indexed by id for id >= len(dense)
|
|
}
|
|
|
|
func makeLocationIDMap(n int) locationIDMap {
|
|
return locationIDMap{
|
|
dense: make([]*Location, n),
|
|
sparse: map[uint64]*Location{},
|
|
}
|
|
}
|
|
|
|
func (lm locationIDMap) get(id uint64) *Location {
|
|
if id < uint64(len(lm.dense)) {
|
|
return lm.dense[int(id)]
|
|
}
|
|
return lm.sparse[id]
|
|
}
|
|
|
|
func (lm locationIDMap) set(id uint64, loc *Location) {
|
|
if id < uint64(len(lm.dense)) {
|
|
lm.dense[id] = loc
|
|
return
|
|
}
|
|
lm.sparse[id] = loc
|
|
}
|
|
|
|
// CompatibilizeSampleTypes makes profiles compatible to be compared/merged. It
|
|
// keeps sample types that appear in all profiles only and drops/reorders the
|
|
// sample types as necessary.
|
|
//
|
|
// In the case of sample types order is not the same for given profiles the
|
|
// order is derived from the first profile.
|
|
//
|
|
// Profiles are modified in-place.
|
|
//
|
|
// It returns an error if the sample type's intersection is empty.
|
|
func CompatibilizeSampleTypes(ps []*Profile) error {
|
|
sTypes := commonSampleTypes(ps)
|
|
if len(sTypes) == 0 {
|
|
return fmt.Errorf("profiles have empty common sample type list")
|
|
}
|
|
for _, p := range ps {
|
|
if err := compatibilizeSampleTypes(p, sTypes); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// commonSampleTypes returns sample types that appear in all profiles in the
|
|
// order how they ordered in the first profile.
|
|
func commonSampleTypes(ps []*Profile) []string {
|
|
if len(ps) == 0 {
|
|
return nil
|
|
}
|
|
sTypes := map[string]int{}
|
|
for _, p := range ps {
|
|
for _, st := range p.SampleType {
|
|
sTypes[st.Type]++
|
|
}
|
|
}
|
|
var res []string
|
|
for _, st := range ps[0].SampleType {
|
|
if sTypes[st.Type] == len(ps) {
|
|
res = append(res, st.Type)
|
|
}
|
|
}
|
|
return res
|
|
}
|
|
|
|
// compatibilizeSampleTypes drops sample types that are not present in sTypes
|
|
// list and reorder them if needed.
|
|
//
|
|
// It sets DefaultSampleType to sType[0] if it is not in sType list.
|
|
//
|
|
// It assumes that all sample types from the sTypes list are present in the
|
|
// given profile otherwise it returns an error.
|
|
func compatibilizeSampleTypes(p *Profile, sTypes []string) error {
|
|
if len(sTypes) == 0 {
|
|
return fmt.Errorf("sample type list is empty")
|
|
}
|
|
defaultSampleType := sTypes[0]
|
|
reMap, needToModify := make([]int, len(sTypes)), false
|
|
for i, st := range sTypes {
|
|
if st == p.DefaultSampleType {
|
|
defaultSampleType = p.DefaultSampleType
|
|
}
|
|
idx := searchValueType(p.SampleType, st)
|
|
if idx < 0 {
|
|
return fmt.Errorf("%q sample type is not found in profile", st)
|
|
}
|
|
reMap[i] = idx
|
|
if idx != i {
|
|
needToModify = true
|
|
}
|
|
}
|
|
if !needToModify && len(sTypes) == len(p.SampleType) {
|
|
return nil
|
|
}
|
|
p.DefaultSampleType = defaultSampleType
|
|
oldSampleTypes := p.SampleType
|
|
p.SampleType = make([]*ValueType, len(sTypes))
|
|
for i, idx := range reMap {
|
|
p.SampleType[i] = oldSampleTypes[idx]
|
|
}
|
|
values := make([]int64, len(sTypes))
|
|
for _, s := range p.Sample {
|
|
for i, idx := range reMap {
|
|
values[i] = s.Value[idx]
|
|
}
|
|
s.Value = s.Value[:len(values)]
|
|
copy(s.Value, values)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func searchValueType(vts []*ValueType, s string) int {
|
|
for i, vt := range vts {
|
|
if vt.Type == s {
|
|
return i
|
|
}
|
|
}
|
|
return -1
|
|
}
|