mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-11-09 16:00:22 +00:00
rebase: bump github.com/prometheus/client_golang from 1.12.2 to 1.14.0
Bumps [github.com/prometheus/client_golang](https://github.com/prometheus/client_golang) from 1.12.2 to 1.14.0. - [Release notes](https://github.com/prometheus/client_golang/releases) - [Changelog](https://github.com/prometheus/client_golang/blob/main/CHANGELOG.md) - [Commits](https://github.com/prometheus/client_golang/compare/v1.12.2...v1.14.0) --- updated-dependencies: - dependency-name: github.com/prometheus/client_golang dependency-type: direct:production update-type: version-update:semver-minor ... Signed-off-by: dependabot[bot] <support@github.com>
This commit is contained in:
parent
bfbd17581b
commit
ec242d4cc8
8
go.mod
8
go.mod
@ -25,7 +25,7 @@ require (
|
|||||||
github.com/onsi/ginkgo/v2 v2.4.0
|
github.com/onsi/ginkgo/v2 v2.4.0
|
||||||
github.com/onsi/gomega v1.23.0
|
github.com/onsi/gomega v1.23.0
|
||||||
github.com/pkg/xattr v0.4.9
|
github.com/pkg/xattr v0.4.9
|
||||||
github.com/prometheus/client_golang v1.12.2
|
github.com/prometheus/client_golang v1.14.0
|
||||||
github.com/stretchr/testify v1.8.1
|
github.com/stretchr/testify v1.8.1
|
||||||
golang.org/x/crypto v0.3.0
|
golang.org/x/crypto v0.3.0
|
||||||
golang.org/x/net v0.2.0
|
golang.org/x/net v0.2.0
|
||||||
@ -129,9 +129,9 @@ require (
|
|||||||
github.com/pierrec/lz4 v2.5.2+incompatible // indirect
|
github.com/pierrec/lz4 v2.5.2+incompatible // indirect
|
||||||
github.com/pkg/errors v0.9.1 // indirect
|
github.com/pkg/errors v0.9.1 // indirect
|
||||||
github.com/pmezard/go-difflib v1.0.0 // indirect
|
github.com/pmezard/go-difflib v1.0.0 // indirect
|
||||||
github.com/prometheus/client_model v0.2.0 // indirect
|
github.com/prometheus/client_model v0.3.0 // indirect
|
||||||
github.com/prometheus/common v0.32.1 // indirect
|
github.com/prometheus/common v0.37.0 // indirect
|
||||||
github.com/prometheus/procfs v0.7.3 // indirect
|
github.com/prometheus/procfs v0.8.0 // indirect
|
||||||
github.com/ryanuber/go-glob v1.0.0 // indirect
|
github.com/ryanuber/go-glob v1.0.0 // indirect
|
||||||
github.com/spf13/cobra v1.4.0 // indirect
|
github.com/spf13/cobra v1.4.0 // indirect
|
||||||
github.com/spf13/pflag v1.0.5 // indirect
|
github.com/spf13/pflag v1.0.5 // indirect
|
||||||
|
17
go.sum
17
go.sum
@ -328,11 +328,13 @@ github.com/go-gl/glfw/v3.3/glfw v0.0.0-20200222043503-6f7a984d4dc4/go.mod h1:tQ2
|
|||||||
github.com/go-kit/kit v0.8.0/go.mod h1:xBxKIO96dXMWWy0MnWVtmwkA9/13aqxPnvrjFYMA2as=
|
github.com/go-kit/kit v0.8.0/go.mod h1:xBxKIO96dXMWWy0MnWVtmwkA9/13aqxPnvrjFYMA2as=
|
||||||
github.com/go-kit/kit v0.9.0/go.mod h1:xBxKIO96dXMWWy0MnWVtmwkA9/13aqxPnvrjFYMA2as=
|
github.com/go-kit/kit v0.9.0/go.mod h1:xBxKIO96dXMWWy0MnWVtmwkA9/13aqxPnvrjFYMA2as=
|
||||||
github.com/go-kit/log v0.1.0/go.mod h1:zbhenjAZHb184qTLMA9ZjW7ThYL0H2mk7Q6pNt4vbaY=
|
github.com/go-kit/log v0.1.0/go.mod h1:zbhenjAZHb184qTLMA9ZjW7ThYL0H2mk7Q6pNt4vbaY=
|
||||||
|
github.com/go-kit/log v0.2.0/go.mod h1:NwTd00d/i8cPZ3xOwwiv2PO5MOcx78fFErGNcVmBjv0=
|
||||||
github.com/go-ldap/ldap v3.0.2+incompatible/go.mod h1:qfd9rJvER9Q0/D/Sqn1DfHRoBp40uXYvFoEVrNEPqRc=
|
github.com/go-ldap/ldap v3.0.2+incompatible/go.mod h1:qfd9rJvER9Q0/D/Sqn1DfHRoBp40uXYvFoEVrNEPqRc=
|
||||||
github.com/go-ldap/ldap/v3 v3.1.3/go.mod h1:3rbOH3jRS2u6jg2rJnKAMLE/xQyCKIveG2Sa/Cohzb8=
|
github.com/go-ldap/ldap/v3 v3.1.3/go.mod h1:3rbOH3jRS2u6jg2rJnKAMLE/xQyCKIveG2Sa/Cohzb8=
|
||||||
github.com/go-logfmt/logfmt v0.3.0/go.mod h1:Qt1PoO58o5twSAckw1HlFXLmHsOX5/0LbT9GBnD5lWE=
|
github.com/go-logfmt/logfmt v0.3.0/go.mod h1:Qt1PoO58o5twSAckw1HlFXLmHsOX5/0LbT9GBnD5lWE=
|
||||||
github.com/go-logfmt/logfmt v0.4.0/go.mod h1:3RMwSq7FuexP4Kalkev3ejPJsZTpXXBr9+V4qmtdjCk=
|
github.com/go-logfmt/logfmt v0.4.0/go.mod h1:3RMwSq7FuexP4Kalkev3ejPJsZTpXXBr9+V4qmtdjCk=
|
||||||
github.com/go-logfmt/logfmt v0.5.0/go.mod h1:wCYkCAKZfumFQihp8CzCvQ3paCTfi41vtzG1KdI/P7A=
|
github.com/go-logfmt/logfmt v0.5.0/go.mod h1:wCYkCAKZfumFQihp8CzCvQ3paCTfi41vtzG1KdI/P7A=
|
||||||
|
github.com/go-logfmt/logfmt v0.5.1/go.mod h1:WYhtIu8zTZfxdn5+rREduYbwxfcBr/Vr6KEVveWlfTs=
|
||||||
github.com/go-logr/logr v0.1.0/go.mod h1:ixOQHD9gLJUVQQ2ZOR7zLEifBX6tGkNJF4QyIY7sIas=
|
github.com/go-logr/logr v0.1.0/go.mod h1:ixOQHD9gLJUVQQ2ZOR7zLEifBX6tGkNJF4QyIY7sIas=
|
||||||
github.com/go-logr/logr v0.2.0/go.mod h1:z6/tIYblkpsD+a4lm/fGIIU9mZ+XfAiaFtq7xTgseGU=
|
github.com/go-logr/logr v0.2.0/go.mod h1:z6/tIYblkpsD+a4lm/fGIIU9mZ+XfAiaFtq7xTgseGU=
|
||||||
github.com/go-logr/logr v0.4.0/go.mod h1:z6/tIYblkpsD+a4lm/fGIIU9mZ+XfAiaFtq7xTgseGU=
|
github.com/go-logr/logr v0.4.0/go.mod h1:z6/tIYblkpsD+a4lm/fGIIU9mZ+XfAiaFtq7xTgseGU=
|
||||||
@ -926,14 +928,15 @@ github.com/prometheus/client_golang v1.7.1/go.mod h1:PY5Wy2awLA44sXw4AOSfFBetzPP
|
|||||||
github.com/prometheus/client_golang v1.11.0/go.mod h1:Z6t4BnS23TR94PD6BsDNk8yVqroYurpAkEiz0P2BEV0=
|
github.com/prometheus/client_golang v1.11.0/go.mod h1:Z6t4BnS23TR94PD6BsDNk8yVqroYurpAkEiz0P2BEV0=
|
||||||
github.com/prometheus/client_golang v1.11.1/go.mod h1:Z6t4BnS23TR94PD6BsDNk8yVqroYurpAkEiz0P2BEV0=
|
github.com/prometheus/client_golang v1.11.1/go.mod h1:Z6t4BnS23TR94PD6BsDNk8yVqroYurpAkEiz0P2BEV0=
|
||||||
github.com/prometheus/client_golang v1.12.1/go.mod h1:3Z9XVyYiZYEO+YQWt3RD2R3jrbd179Rt297l4aS6nDY=
|
github.com/prometheus/client_golang v1.12.1/go.mod h1:3Z9XVyYiZYEO+YQWt3RD2R3jrbd179Rt297l4aS6nDY=
|
||||||
github.com/prometheus/client_golang v1.12.2 h1:51L9cDoUHVrXx4zWYlcLQIZ+d+VXHgqnYKkIuq4g/34=
|
github.com/prometheus/client_golang v1.14.0 h1:nJdhIvne2eSX/XRAFV9PcvFFRbrjbcTUj0VP62TMhnw=
|
||||||
github.com/prometheus/client_golang v1.12.2/go.mod h1:3Z9XVyYiZYEO+YQWt3RD2R3jrbd179Rt297l4aS6nDY=
|
github.com/prometheus/client_golang v1.14.0/go.mod h1:8vpkKitgIVNcqrRBWh1C4TIUQgYNtG/XQE4E/Zae36Y=
|
||||||
github.com/prometheus/client_model v0.0.0-20180712105110-5c3871d89910/go.mod h1:MbSGuTsp3dbXC40dX6PRTWyKYBIrTGTE9sqQNg2J8bo=
|
github.com/prometheus/client_model v0.0.0-20180712105110-5c3871d89910/go.mod h1:MbSGuTsp3dbXC40dX6PRTWyKYBIrTGTE9sqQNg2J8bo=
|
||||||
github.com/prometheus/client_model v0.0.0-20190115171406-56726106282f/go.mod h1:MbSGuTsp3dbXC40dX6PRTWyKYBIrTGTE9sqQNg2J8bo=
|
github.com/prometheus/client_model v0.0.0-20190115171406-56726106282f/go.mod h1:MbSGuTsp3dbXC40dX6PRTWyKYBIrTGTE9sqQNg2J8bo=
|
||||||
github.com/prometheus/client_model v0.0.0-20190129233127-fd36f4220a90/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
github.com/prometheus/client_model v0.0.0-20190129233127-fd36f4220a90/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
||||||
github.com/prometheus/client_model v0.0.0-20190812154241-14fe0d1b01d4/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
github.com/prometheus/client_model v0.0.0-20190812154241-14fe0d1b01d4/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
||||||
github.com/prometheus/client_model v0.2.0 h1:uq5h0d+GuxiXLJLNABMgp2qUWDPiLvgCzz2dUR+/W/M=
|
|
||||||
github.com/prometheus/client_model v0.2.0/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
github.com/prometheus/client_model v0.2.0/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
||||||
|
github.com/prometheus/client_model v0.3.0 h1:UBgGFHqYdG/TPFD1B1ogZywDqEkwp3fBMvqdiQ7Xew4=
|
||||||
|
github.com/prometheus/client_model v0.3.0/go.mod h1:LDGWKZIo7rky3hgvBe+caln+Dr3dPggB5dvjtD7w9+w=
|
||||||
github.com/prometheus/common v0.0.0-20180801064454-c7de2306084e/go.mod h1:daVV7qP5qjZbuso7PdcryaAu0sAZbrN9i7WWcTMWvro=
|
github.com/prometheus/common v0.0.0-20180801064454-c7de2306084e/go.mod h1:daVV7qP5qjZbuso7PdcryaAu0sAZbrN9i7WWcTMWvro=
|
||||||
github.com/prometheus/common v0.0.0-20181113130724-41aa239b4cce/go.mod h1:daVV7qP5qjZbuso7PdcryaAu0sAZbrN9i7WWcTMWvro=
|
github.com/prometheus/common v0.0.0-20181113130724-41aa239b4cce/go.mod h1:daVV7qP5qjZbuso7PdcryaAu0sAZbrN9i7WWcTMWvro=
|
||||||
github.com/prometheus/common v0.0.0-20181126121408-4724e9255275/go.mod h1:daVV7qP5qjZbuso7PdcryaAu0sAZbrN9i7WWcTMWvro=
|
github.com/prometheus/common v0.0.0-20181126121408-4724e9255275/go.mod h1:daVV7qP5qjZbuso7PdcryaAu0sAZbrN9i7WWcTMWvro=
|
||||||
@ -943,8 +946,9 @@ github.com/prometheus/common v0.4.1/go.mod h1:TNfzLD0ON7rHzMJeJkieUDPYmFC7Snx/y8
|
|||||||
github.com/prometheus/common v0.9.1/go.mod h1:yhUN8i9wzaXS3w1O07YhxHEBxD+W35wd8bs7vj7HSQ4=
|
github.com/prometheus/common v0.9.1/go.mod h1:yhUN8i9wzaXS3w1O07YhxHEBxD+W35wd8bs7vj7HSQ4=
|
||||||
github.com/prometheus/common v0.10.0/go.mod h1:Tlit/dnDKsSWFlCLTWaA1cyBgKHSMdTB80sz/V91rCo=
|
github.com/prometheus/common v0.10.0/go.mod h1:Tlit/dnDKsSWFlCLTWaA1cyBgKHSMdTB80sz/V91rCo=
|
||||||
github.com/prometheus/common v0.26.0/go.mod h1:M7rCNAaPfAosfx8veZJCuw84e35h3Cfd9VFqTh1DIvc=
|
github.com/prometheus/common v0.26.0/go.mod h1:M7rCNAaPfAosfx8veZJCuw84e35h3Cfd9VFqTh1DIvc=
|
||||||
github.com/prometheus/common v0.32.1 h1:hWIdL3N2HoUx3B8j3YN9mWor0qhY/NlEKZEaXxuIRh4=
|
|
||||||
github.com/prometheus/common v0.32.1/go.mod h1:vu+V0TpY+O6vW9J44gczi3Ap/oXXR10b+M/gUGO4Hls=
|
github.com/prometheus/common v0.32.1/go.mod h1:vu+V0TpY+O6vW9J44gczi3Ap/oXXR10b+M/gUGO4Hls=
|
||||||
|
github.com/prometheus/common v0.37.0 h1:ccBbHCgIiT9uSoFY0vX8H3zsNR5eLt17/RQLUvn8pXE=
|
||||||
|
github.com/prometheus/common v0.37.0/go.mod h1:phzohg0JFMnBEFGxTDbfu3QyL5GI8gTQJFhYO5B3mfA=
|
||||||
github.com/prometheus/procfs v0.0.0-20180725123919-05ee40e3a273/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk=
|
github.com/prometheus/procfs v0.0.0-20180725123919-05ee40e3a273/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk=
|
||||||
github.com/prometheus/procfs v0.0.0-20181005140218-185b4288413d/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk=
|
github.com/prometheus/procfs v0.0.0-20181005140218-185b4288413d/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk=
|
||||||
github.com/prometheus/procfs v0.0.0-20181204211112-1dc9a6cbc91a/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk=
|
github.com/prometheus/procfs v0.0.0-20181204211112-1dc9a6cbc91a/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk=
|
||||||
@ -954,8 +958,9 @@ github.com/prometheus/procfs v0.0.2/go.mod h1:TjEm7ze935MbeOT/UhFTIMYKhuLP4wbCsT
|
|||||||
github.com/prometheus/procfs v0.0.8/go.mod h1:7Qr8sr6344vo1JqZ6HhLceV9o3AJ1Ff+GxbHq6oeK9A=
|
github.com/prometheus/procfs v0.0.8/go.mod h1:7Qr8sr6344vo1JqZ6HhLceV9o3AJ1Ff+GxbHq6oeK9A=
|
||||||
github.com/prometheus/procfs v0.1.3/go.mod h1:lV6e/gmhEcM9IjHGsFOCxxuZ+z1YqCvr4OA4YeYWdaU=
|
github.com/prometheus/procfs v0.1.3/go.mod h1:lV6e/gmhEcM9IjHGsFOCxxuZ+z1YqCvr4OA4YeYWdaU=
|
||||||
github.com/prometheus/procfs v0.6.0/go.mod h1:cz+aTbrPOrUb4q7XlbU9ygM+/jj0fzG6c1xBZuNvfVA=
|
github.com/prometheus/procfs v0.6.0/go.mod h1:cz+aTbrPOrUb4q7XlbU9ygM+/jj0fzG6c1xBZuNvfVA=
|
||||||
github.com/prometheus/procfs v0.7.3 h1:4jVXhlkAyzOScmCkXBTOLRLTz8EeU+eyjrwB/EPq0VU=
|
|
||||||
github.com/prometheus/procfs v0.7.3/go.mod h1:cz+aTbrPOrUb4q7XlbU9ygM+/jj0fzG6c1xBZuNvfVA=
|
github.com/prometheus/procfs v0.7.3/go.mod h1:cz+aTbrPOrUb4q7XlbU9ygM+/jj0fzG6c1xBZuNvfVA=
|
||||||
|
github.com/prometheus/procfs v0.8.0 h1:ODq8ZFEaYeCaZOJlZZdJA2AbQR98dSHSM1KW/You5mo=
|
||||||
|
github.com/prometheus/procfs v0.8.0/go.mod h1:z7EfXMXOkbkqb9IINtpCn86r/to3BnA0uaxHdg830/4=
|
||||||
github.com/prometheus/tsdb v0.7.1/go.mod h1:qhTCs0VvXwvX/y3TZrWD7rabWM+ijKTux40TwIPHuXU=
|
github.com/prometheus/tsdb v0.7.1/go.mod h1:qhTCs0VvXwvX/y3TZrWD7rabWM+ijKTux40TwIPHuXU=
|
||||||
github.com/rcrowley/go-metrics v0.0.0-20181016184325-3113b8401b8a/go.mod h1:bCqnVzQkZxMG4s8nGwiZ5l3QUCyqpo9Y+/ZMZ9VjZe4=
|
github.com/rcrowley/go-metrics v0.0.0-20181016184325-3113b8401b8a/go.mod h1:bCqnVzQkZxMG4s8nGwiZ5l3QUCyqpo9Y+/ZMZ9VjZe4=
|
||||||
github.com/rogpeppe/fastuuid v0.0.0-20150106093220-6724a57986af/go.mod h1:XWv6SoW27p1b0cqNHllgS5HIMJraePCO15w5zCzIWYg=
|
github.com/rogpeppe/fastuuid v0.0.0-20150106093220-6724a57986af/go.mod h1:XWv6SoW27p1b0cqNHllgS5HIMJraePCO15w5zCzIWYg=
|
||||||
@ -1231,6 +1236,7 @@ golang.org/x/net v0.0.0-20210525063256-abc453219eb5/go.mod h1:9nx3DQGgdP8bBQD5qx
|
|||||||
golang.org/x/net v0.0.0-20210805182204-aaa1db679c0d/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
golang.org/x/net v0.0.0-20210805182204-aaa1db679c0d/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
||||||
golang.org/x/net v0.0.0-20211015210444-4f30a5c0130f/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
golang.org/x/net v0.0.0-20211015210444-4f30a5c0130f/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
||||||
golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
||||||
|
golang.org/x/net v0.0.0-20220127200216-cd36cc0744dd/go.mod h1:CfG3xpIq0wQ8r1q4Su4UZFWDARRcnwPjda9FqA0JpMk=
|
||||||
golang.org/x/net v0.0.0-20220225172249-27dd8689420f/go.mod h1:CfG3xpIq0wQ8r1q4Su4UZFWDARRcnwPjda9FqA0JpMk=
|
golang.org/x/net v0.0.0-20220225172249-27dd8689420f/go.mod h1:CfG3xpIq0wQ8r1q4Su4UZFWDARRcnwPjda9FqA0JpMk=
|
||||||
golang.org/x/net v0.0.0-20220425223048-2871e0cb64e4/go.mod h1:CfG3xpIq0wQ8r1q4Su4UZFWDARRcnwPjda9FqA0JpMk=
|
golang.org/x/net v0.0.0-20220425223048-2871e0cb64e4/go.mod h1:CfG3xpIq0wQ8r1q4Su4UZFWDARRcnwPjda9FqA0JpMk=
|
||||||
golang.org/x/net v0.0.0-20220722155237-a158d28d115b/go.mod h1:XRhObCWvk6IyKnWLug+ECip1KBveYUHfp+8e9klMJ9c=
|
golang.org/x/net v0.0.0-20220722155237-a158d28d115b/go.mod h1:XRhObCWvk6IyKnWLug+ECip1KBveYUHfp+8e9klMJ9c=
|
||||||
@ -1257,6 +1263,7 @@ golang.org/x/oauth2 v0.0.0-20210628180205-a41e5a781914/go.mod h1:KelEdhl1UZF7XfJ
|
|||||||
golang.org/x/oauth2 v0.0.0-20210805134026-6f1e6394065a/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A=
|
golang.org/x/oauth2 v0.0.0-20210805134026-6f1e6394065a/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A=
|
||||||
golang.org/x/oauth2 v0.0.0-20210819190943-2bc19b11175f/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A=
|
golang.org/x/oauth2 v0.0.0-20210819190943-2bc19b11175f/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A=
|
||||||
golang.org/x/oauth2 v0.0.0-20211104180415-d3ed0bb246c8/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A=
|
golang.org/x/oauth2 v0.0.0-20211104180415-d3ed0bb246c8/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A=
|
||||||
|
golang.org/x/oauth2 v0.0.0-20220223155221-ee480838109b/go.mod h1:DAh4E804XQdzx2j+YRIaUnCqCV2RuMz24cGBJ5QYIrc=
|
||||||
golang.org/x/oauth2 v0.2.0 h1:GtQkldQ9m7yvzCL1V+LrYow3Khe0eJH0w7RbX/VbaIU=
|
golang.org/x/oauth2 v0.2.0 h1:GtQkldQ9m7yvzCL1V+LrYow3Khe0eJH0w7RbX/VbaIU=
|
||||||
golang.org/x/oauth2 v0.2.0/go.mod h1:Cwn6afJ8jrQwYMxQDTpISoXmXW9I6qF6vDeuuoX3Ibs=
|
golang.org/x/oauth2 v0.2.0/go.mod h1:Cwn6afJ8jrQwYMxQDTpISoXmXW9I6qF6vDeuuoX3Ibs=
|
||||||
golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
|
6
vendor/github.com/prometheus/client_golang/prometheus/collector.go
generated
vendored
6
vendor/github.com/prometheus/client_golang/prometheus/collector.go
generated
vendored
@ -69,9 +69,9 @@ type Collector interface {
|
|||||||
// If a Collector collects the same metrics throughout its lifetime, its
|
// If a Collector collects the same metrics throughout its lifetime, its
|
||||||
// Describe method can simply be implemented as:
|
// Describe method can simply be implemented as:
|
||||||
//
|
//
|
||||||
// func (c customCollector) Describe(ch chan<- *Desc) {
|
// func (c customCollector) Describe(ch chan<- *Desc) {
|
||||||
// DescribeByCollect(c, ch)
|
// DescribeByCollect(c, ch)
|
||||||
// }
|
// }
|
||||||
//
|
//
|
||||||
// However, this will not work if the metrics collected change dynamically over
|
// However, this will not work if the metrics collected change dynamically over
|
||||||
// the lifetime of the Collector in a way that their combined set of descriptors
|
// the lifetime of the Collector in a way that their combined set of descriptors
|
||||||
|
@ -101,7 +101,7 @@ func (c *dbStatsCollector) Describe(ch chan<- *prometheus.Desc) {
|
|||||||
ch <- c.waitDuration
|
ch <- c.waitDuration
|
||||||
ch <- c.maxIdleClosed
|
ch <- c.maxIdleClosed
|
||||||
ch <- c.maxLifetimeClosed
|
ch <- c.maxLifetimeClosed
|
||||||
c.describeNewInGo115(ch)
|
ch <- c.maxIdleTimeClosed
|
||||||
}
|
}
|
||||||
|
|
||||||
// Collect implements Collector.
|
// Collect implements Collector.
|
||||||
@ -115,5 +115,5 @@ func (c *dbStatsCollector) Collect(ch chan<- prometheus.Metric) {
|
|||||||
ch <- prometheus.MustNewConstMetric(c.waitDuration, prometheus.CounterValue, stats.WaitDuration.Seconds())
|
ch <- prometheus.MustNewConstMetric(c.waitDuration, prometheus.CounterValue, stats.WaitDuration.Seconds())
|
||||||
ch <- prometheus.MustNewConstMetric(c.maxIdleClosed, prometheus.CounterValue, float64(stats.MaxIdleClosed))
|
ch <- prometheus.MustNewConstMetric(c.maxIdleClosed, prometheus.CounterValue, float64(stats.MaxIdleClosed))
|
||||||
ch <- prometheus.MustNewConstMetric(c.maxLifetimeClosed, prometheus.CounterValue, float64(stats.MaxLifetimeClosed))
|
ch <- prometheus.MustNewConstMetric(c.maxLifetimeClosed, prometheus.CounterValue, float64(stats.MaxLifetimeClosed))
|
||||||
c.collectNewInGo115(ch, stats)
|
ch <- prometheus.MustNewConstMetric(c.maxIdleTimeClosed, prometheus.CounterValue, float64(stats.MaxIdleTimeClosed))
|
||||||
}
|
}
|
||||||
|
185
vendor/github.com/prometheus/client_golang/prometheus/collectors/go_collector_latest.go
generated
vendored
185
vendor/github.com/prometheus/client_golang/prometheus/collectors/go_collector_latest.go
generated
vendored
@ -16,76 +16,145 @@
|
|||||||
|
|
||||||
package collectors
|
package collectors
|
||||||
|
|
||||||
import "github.com/prometheus/client_golang/prometheus"
|
import (
|
||||||
|
"regexp"
|
||||||
|
|
||||||
//nolint:staticcheck // Ignore SA1019 until v2.
|
"github.com/prometheus/client_golang/prometheus"
|
||||||
type goOptions = prometheus.GoCollectorOptions
|
"github.com/prometheus/client_golang/prometheus/internal"
|
||||||
type goOption func(o *goOptions)
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// MetricsAll allows all the metrics to be collected from Go runtime.
|
||||||
|
MetricsAll = GoRuntimeMetricsRule{regexp.MustCompile("/.*")}
|
||||||
|
// MetricsGC allows only GC metrics to be collected from Go runtime.
|
||||||
|
// e.g. go_gc_cycles_automatic_gc_cycles_total
|
||||||
|
MetricsGC = GoRuntimeMetricsRule{regexp.MustCompile(`^/gc/.*`)}
|
||||||
|
// MetricsMemory allows only memory metrics to be collected from Go runtime.
|
||||||
|
// e.g. go_memory_classes_heap_free_bytes
|
||||||
|
MetricsMemory = GoRuntimeMetricsRule{regexp.MustCompile(`^/memory/.*`)}
|
||||||
|
// MetricsScheduler allows only scheduler metrics to be collected from Go runtime.
|
||||||
|
// e.g. go_sched_goroutines_goroutines
|
||||||
|
MetricsScheduler = GoRuntimeMetricsRule{regexp.MustCompile(`^/sched/.*`)}
|
||||||
|
)
|
||||||
|
|
||||||
|
// WithGoCollectorMemStatsMetricsDisabled disables metrics that is gathered in runtime.MemStats structure such as:
|
||||||
|
//
|
||||||
|
// go_memstats_alloc_bytes
|
||||||
|
// go_memstats_alloc_bytes_total
|
||||||
|
// go_memstats_sys_bytes
|
||||||
|
// go_memstats_lookups_total
|
||||||
|
// go_memstats_mallocs_total
|
||||||
|
// go_memstats_frees_total
|
||||||
|
// go_memstats_heap_alloc_bytes
|
||||||
|
// go_memstats_heap_sys_bytes
|
||||||
|
// go_memstats_heap_idle_bytes
|
||||||
|
// go_memstats_heap_inuse_bytes
|
||||||
|
// go_memstats_heap_released_bytes
|
||||||
|
// go_memstats_heap_objects
|
||||||
|
// go_memstats_stack_inuse_bytes
|
||||||
|
// go_memstats_stack_sys_bytes
|
||||||
|
// go_memstats_mspan_inuse_bytes
|
||||||
|
// go_memstats_mspan_sys_bytes
|
||||||
|
// go_memstats_mcache_inuse_bytes
|
||||||
|
// go_memstats_mcache_sys_bytes
|
||||||
|
// go_memstats_buck_hash_sys_bytes
|
||||||
|
// go_memstats_gc_sys_bytes
|
||||||
|
// go_memstats_other_sys_bytes
|
||||||
|
// go_memstats_next_gc_bytes
|
||||||
|
//
|
||||||
|
// so the metrics known from pre client_golang v1.12.0,
|
||||||
|
//
|
||||||
|
// NOTE(bwplotka): The above represents runtime.MemStats statistics, but they are
|
||||||
|
// actually implemented using new runtime/metrics package. (except skipped go_memstats_gc_cpu_fraction
|
||||||
|
// -- see https://github.com/prometheus/client_golang/issues/842#issuecomment-861812034 for explanation).
|
||||||
|
//
|
||||||
|
// Some users might want to disable this on collector level (although you can use scrape relabelling on Prometheus),
|
||||||
|
// because similar metrics can be now obtained using WithGoCollectorRuntimeMetrics. Note that the semantics of new
|
||||||
|
// metrics might be different, plus the names can be change over time with different Go version.
|
||||||
|
//
|
||||||
|
// NOTE(bwplotka): Changing metric names can be tedious at times as the alerts, recording rules and dashboards have to be adjusted.
|
||||||
|
// The old metrics are also very useful, with many guides and books written about how to interpret them.
|
||||||
|
//
|
||||||
|
// As a result our recommendation would be to stick with MemStats like metrics and enable other runtime/metrics if you are interested
|
||||||
|
// in advanced insights Go provides. See ExampleGoCollector_WithAdvancedGoMetrics.
|
||||||
|
func WithGoCollectorMemStatsMetricsDisabled() func(options *internal.GoCollectorOptions) {
|
||||||
|
return func(o *internal.GoCollectorOptions) {
|
||||||
|
o.DisableMemStatsLikeMetrics = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// GoRuntimeMetricsRule allow enabling and configuring particular group of runtime/metrics.
|
||||||
|
// TODO(bwplotka): Consider adding ability to adjust buckets.
|
||||||
|
type GoRuntimeMetricsRule struct {
|
||||||
|
// Matcher represents RE2 expression will match the runtime/metrics from https://golang.bg/src/runtime/metrics/description.go
|
||||||
|
// Use `regexp.MustCompile` or `regexp.Compile` to create this field.
|
||||||
|
Matcher *regexp.Regexp
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithGoCollectorRuntimeMetrics allows enabling and configuring particular group of runtime/metrics.
|
||||||
|
// See the list of metrics https://golang.bg/src/runtime/metrics/description.go (pick the Go version you use there!).
|
||||||
|
// You can use this option in repeated manner, which will add new rules. The order of rules is important, the last rule
|
||||||
|
// that matches particular metrics is applied.
|
||||||
|
func WithGoCollectorRuntimeMetrics(rules ...GoRuntimeMetricsRule) func(options *internal.GoCollectorOptions) {
|
||||||
|
rs := make([]internal.GoCollectorRule, len(rules))
|
||||||
|
for i, r := range rules {
|
||||||
|
rs[i] = internal.GoCollectorRule{
|
||||||
|
Matcher: r.Matcher,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return func(o *internal.GoCollectorOptions) {
|
||||||
|
o.RuntimeMetricRules = append(o.RuntimeMetricRules, rs...)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithoutGoCollectorRuntimeMetrics allows disabling group of runtime/metrics that you might have added in WithGoCollectorRuntimeMetrics.
|
||||||
|
// It behaves similarly to WithGoCollectorRuntimeMetrics just with deny-list semantics.
|
||||||
|
func WithoutGoCollectorRuntimeMetrics(matchers ...*regexp.Regexp) func(options *internal.GoCollectorOptions) {
|
||||||
|
rs := make([]internal.GoCollectorRule, len(matchers))
|
||||||
|
for i, m := range matchers {
|
||||||
|
rs[i] = internal.GoCollectorRule{
|
||||||
|
Matcher: m,
|
||||||
|
Deny: true,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return func(o *internal.GoCollectorOptions) {
|
||||||
|
o.RuntimeMetricRules = append(o.RuntimeMetricRules, rs...)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// GoCollectionOption represents Go collection option flag.
|
||||||
|
// Deprecated.
|
||||||
type GoCollectionOption uint32
|
type GoCollectionOption uint32
|
||||||
|
|
||||||
const (
|
const (
|
||||||
// GoRuntimeMemStatsCollection represents the metrics represented by runtime.MemStats structure such as
|
// GoRuntimeMemStatsCollection represents the metrics represented by runtime.MemStats structure.
|
||||||
// go_memstats_alloc_bytes
|
// Deprecated. Use WithGoCollectorMemStatsMetricsDisabled() function to disable those metrics in the collector.
|
||||||
// go_memstats_alloc_bytes_total
|
|
||||||
// go_memstats_sys_bytes
|
|
||||||
// go_memstats_lookups_total
|
|
||||||
// go_memstats_mallocs_total
|
|
||||||
// go_memstats_frees_total
|
|
||||||
// go_memstats_heap_alloc_bytes
|
|
||||||
// go_memstats_heap_sys_bytes
|
|
||||||
// go_memstats_heap_idle_bytes
|
|
||||||
// go_memstats_heap_inuse_bytes
|
|
||||||
// go_memstats_heap_released_bytes
|
|
||||||
// go_memstats_heap_objects
|
|
||||||
// go_memstats_stack_inuse_bytes
|
|
||||||
// go_memstats_stack_sys_bytes
|
|
||||||
// go_memstats_mspan_inuse_bytes
|
|
||||||
// go_memstats_mspan_sys_bytes
|
|
||||||
// go_memstats_mcache_inuse_bytes
|
|
||||||
// go_memstats_mcache_sys_bytes
|
|
||||||
// go_memstats_buck_hash_sys_bytes
|
|
||||||
// go_memstats_gc_sys_bytes
|
|
||||||
// go_memstats_other_sys_bytes
|
|
||||||
// go_memstats_next_gc_bytes
|
|
||||||
// so the metrics known from pre client_golang v1.12.0, except skipped go_memstats_gc_cpu_fraction (see
|
|
||||||
// https://github.com/prometheus/client_golang/issues/842#issuecomment-861812034 for explanation.
|
|
||||||
//
|
|
||||||
// NOTE that this mode represents runtime.MemStats statistics, but they are
|
|
||||||
// actually implemented using new runtime/metrics package.
|
|
||||||
// Deprecated: Use GoRuntimeMetricsCollection instead going forward.
|
|
||||||
GoRuntimeMemStatsCollection GoCollectionOption = 1 << iota
|
GoRuntimeMemStatsCollection GoCollectionOption = 1 << iota
|
||||||
// GoRuntimeMetricsCollection is the new set of metrics represented by runtime/metrics package and follows
|
// GoRuntimeMetricsCollection is the new set of metrics represented by runtime/metrics package.
|
||||||
// consistent naming. The exposed metric set depends on Go version, but it is controlled against
|
// Deprecated. Use WithGoCollectorRuntimeMetrics(GoRuntimeMetricsRule{Matcher: regexp.MustCompile("/.*")})
|
||||||
// unexpected cardinality. This set has overlapping information with GoRuntimeMemStatsCollection, just with
|
// function to enable those metrics in the collector.
|
||||||
// new names. GoRuntimeMetricsCollection is what is recommended for using going forward.
|
|
||||||
GoRuntimeMetricsCollection
|
GoRuntimeMetricsCollection
|
||||||
)
|
)
|
||||||
|
|
||||||
// WithGoCollections allows enabling different collections for Go collector on top of base metrics
|
// WithGoCollections allows enabling different collections for Go collector on top of base metrics.
|
||||||
// like go_goroutines, go_threads, go_gc_duration_seconds, go_memstats_last_gc_time_seconds, go_info.
|
// Deprecated. Use WithGoCollectorRuntimeMetrics() and WithGoCollectorMemStatsMetricsDisabled() instead to control metrics.
|
||||||
//
|
func WithGoCollections(flags GoCollectionOption) func(options *internal.GoCollectorOptions) {
|
||||||
// Check GoRuntimeMemStatsCollection and GoRuntimeMetricsCollection for more details. You can use none,
|
return func(options *internal.GoCollectorOptions) {
|
||||||
// one or more collections at once. For example:
|
if flags&GoRuntimeMemStatsCollection == 0 {
|
||||||
// WithGoCollections(GoRuntimeMemStatsCollection | GoRuntimeMetricsCollection) means both GoRuntimeMemStatsCollection
|
WithGoCollectorMemStatsMetricsDisabled()(options)
|
||||||
// metrics and GoRuntimeMetricsCollection will be exposed.
|
}
|
||||||
//
|
|
||||||
// The current default is GoRuntimeMemStatsCollection, so the compatibility mode with
|
if flags&GoRuntimeMetricsCollection != 0 {
|
||||||
// client_golang pre v1.12 (move to runtime/metrics).
|
WithGoCollectorRuntimeMetrics(GoRuntimeMetricsRule{Matcher: regexp.MustCompile("/.*")})(options)
|
||||||
func WithGoCollections(flags GoCollectionOption) goOption {
|
}
|
||||||
return func(o *goOptions) {
|
|
||||||
o.EnabledCollections = uint32(flags)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewGoCollector returns a collector that exports metrics about the current Go
|
// NewGoCollector returns a collector that exports metrics about the current Go
|
||||||
// process using debug.GCStats using runtime/metrics.
|
// process using debug.GCStats (base metrics) and runtime/metrics (both in MemStats style and new ones).
|
||||||
func NewGoCollector(opts ...goOption) prometheus.Collector {
|
func NewGoCollector(opts ...func(o *internal.GoCollectorOptions)) prometheus.Collector {
|
||||||
//nolint:staticcheck // Ignore SA1019 until v2.
|
//nolint:staticcheck // Ignore SA1019 until v2.
|
||||||
promPkgOpts := make([]func(o *prometheus.GoCollectorOptions), len(opts))
|
return prometheus.NewGoCollector(opts...)
|
||||||
for i, opt := range opts {
|
|
||||||
promPkgOpts[i] = opt
|
|
||||||
}
|
|
||||||
//nolint:staticcheck // Ignore SA1019 until v2.
|
|
||||||
return prometheus.NewGoCollector(promPkgOpts...)
|
|
||||||
}
|
}
|
||||||
|
13
vendor/github.com/prometheus/client_golang/prometheus/counter.go
generated
vendored
13
vendor/github.com/prometheus/client_golang/prometheus/counter.go
generated
vendored
@ -51,7 +51,7 @@ type Counter interface {
|
|||||||
// will lead to a valid (label-less) exemplar. But if Labels is nil, the current
|
// will lead to a valid (label-less) exemplar. But if Labels is nil, the current
|
||||||
// exemplar is left in place. AddWithExemplar panics if the value is < 0, if any
|
// exemplar is left in place. AddWithExemplar panics if the value is < 0, if any
|
||||||
// of the provided labels are invalid, or if the provided labels contain more
|
// of the provided labels are invalid, or if the provided labels contain more
|
||||||
// than 64 runes in total.
|
// than 128 runes in total.
|
||||||
type ExemplarAdder interface {
|
type ExemplarAdder interface {
|
||||||
AddWithExemplar(value float64, exemplar Labels)
|
AddWithExemplar(value float64, exemplar Labels)
|
||||||
}
|
}
|
||||||
@ -140,12 +140,13 @@ func (c *counter) get() float64 {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (c *counter) Write(out *dto.Metric) error {
|
func (c *counter) Write(out *dto.Metric) error {
|
||||||
val := c.get()
|
// Read the Exemplar first and the value second. This is to avoid a race condition
|
||||||
|
// where users see an exemplar for a not-yet-existing observation.
|
||||||
var exemplar *dto.Exemplar
|
var exemplar *dto.Exemplar
|
||||||
if e := c.exemplar.Load(); e != nil {
|
if e := c.exemplar.Load(); e != nil {
|
||||||
exemplar = e.(*dto.Exemplar)
|
exemplar = e.(*dto.Exemplar)
|
||||||
}
|
}
|
||||||
|
val := c.get()
|
||||||
|
|
||||||
return populateMetric(CounterValue, val, c.labelPairs, exemplar, out)
|
return populateMetric(CounterValue, val, c.labelPairs, exemplar, out)
|
||||||
}
|
}
|
||||||
@ -245,7 +246,8 @@ func (v *CounterVec) GetMetricWith(labels Labels) (Counter, error) {
|
|||||||
// WithLabelValues works as GetMetricWithLabelValues, but panics where
|
// WithLabelValues works as GetMetricWithLabelValues, but panics where
|
||||||
// GetMetricWithLabelValues would have returned an error. Not returning an
|
// GetMetricWithLabelValues would have returned an error. Not returning an
|
||||||
// error allows shortcuts like
|
// error allows shortcuts like
|
||||||
// myVec.WithLabelValues("404", "GET").Add(42)
|
//
|
||||||
|
// myVec.WithLabelValues("404", "GET").Add(42)
|
||||||
func (v *CounterVec) WithLabelValues(lvs ...string) Counter {
|
func (v *CounterVec) WithLabelValues(lvs ...string) Counter {
|
||||||
c, err := v.GetMetricWithLabelValues(lvs...)
|
c, err := v.GetMetricWithLabelValues(lvs...)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -256,7 +258,8 @@ func (v *CounterVec) WithLabelValues(lvs ...string) Counter {
|
|||||||
|
|
||||||
// With works as GetMetricWith, but panics where GetMetricWithLabels would have
|
// With works as GetMetricWith, but panics where GetMetricWithLabels would have
|
||||||
// returned an error. Not returning an error allows shortcuts like
|
// returned an error. Not returning an error allows shortcuts like
|
||||||
// myVec.With(prometheus.Labels{"code": "404", "method": "GET"}).Add(42)
|
//
|
||||||
|
// myVec.With(prometheus.Labels{"code": "404", "method": "GET"}).Add(42)
|
||||||
func (v *CounterVec) With(labels Labels) Counter {
|
func (v *CounterVec) With(labels Labels) Counter {
|
||||||
c, err := v.GetMetricWith(labels)
|
c, err := v.GetMetricWith(labels)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
5
vendor/github.com/prometheus/client_golang/prometheus/desc.go
generated
vendored
5
vendor/github.com/prometheus/client_golang/prometheus/desc.go
generated
vendored
@ -20,6 +20,9 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/cespare/xxhash/v2"
|
"github.com/cespare/xxhash/v2"
|
||||||
|
|
||||||
|
"github.com/prometheus/client_golang/prometheus/internal"
|
||||||
|
|
||||||
//nolint:staticcheck // Ignore SA1019. Need to keep deprecated package for compatibility.
|
//nolint:staticcheck // Ignore SA1019. Need to keep deprecated package for compatibility.
|
||||||
"github.com/golang/protobuf/proto"
|
"github.com/golang/protobuf/proto"
|
||||||
"github.com/prometheus/common/model"
|
"github.com/prometheus/common/model"
|
||||||
@ -154,7 +157,7 @@ func NewDesc(fqName, help string, variableLabels []string, constLabels Labels) *
|
|||||||
Value: proto.String(v),
|
Value: proto.String(v),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
sort.Sort(labelPairSorter(d.constLabelPairs))
|
sort.Sort(internal.LabelPairSorter(d.constLabelPairs))
|
||||||
return d
|
return d
|
||||||
}
|
}
|
||||||
|
|
||||||
|
93
vendor/github.com/prometheus/client_golang/prometheus/doc.go
generated
vendored
93
vendor/github.com/prometheus/client_golang/prometheus/doc.go
generated
vendored
@ -21,55 +21,66 @@
|
|||||||
// All exported functions and methods are safe to be used concurrently unless
|
// All exported functions and methods are safe to be used concurrently unless
|
||||||
// specified otherwise.
|
// specified otherwise.
|
||||||
//
|
//
|
||||||
// A Basic Example
|
// # A Basic Example
|
||||||
//
|
//
|
||||||
// As a starting point, a very basic usage example:
|
// As a starting point, a very basic usage example:
|
||||||
//
|
//
|
||||||
// package main
|
// package main
|
||||||
//
|
//
|
||||||
// import (
|
// import (
|
||||||
// "log"
|
// "log"
|
||||||
// "net/http"
|
// "net/http"
|
||||||
//
|
//
|
||||||
// "github.com/prometheus/client_golang/prometheus"
|
// "github.com/prometheus/client_golang/prometheus"
|
||||||
// "github.com/prometheus/client_golang/prometheus/promhttp"
|
// "github.com/prometheus/client_golang/prometheus/promhttp"
|
||||||
// )
|
// )
|
||||||
//
|
//
|
||||||
// var (
|
// type metrics struct {
|
||||||
// cpuTemp = prometheus.NewGauge(prometheus.GaugeOpts{
|
// cpuTemp prometheus.Gauge
|
||||||
// Name: "cpu_temperature_celsius",
|
// hdFailures *prometheus.CounterVec
|
||||||
// Help: "Current temperature of the CPU.",
|
// }
|
||||||
// })
|
|
||||||
// hdFailures = prometheus.NewCounterVec(
|
|
||||||
// prometheus.CounterOpts{
|
|
||||||
// Name: "hd_errors_total",
|
|
||||||
// Help: "Number of hard-disk errors.",
|
|
||||||
// },
|
|
||||||
// []string{"device"},
|
|
||||||
// )
|
|
||||||
// )
|
|
||||||
//
|
//
|
||||||
// func init() {
|
// func NewMetrics(reg prometheus.Registerer) *metrics {
|
||||||
// // Metrics have to be registered to be exposed:
|
// m := &metrics{
|
||||||
// prometheus.MustRegister(cpuTemp)
|
// cpuTemp: prometheus.NewGauge(prometheus.GaugeOpts{
|
||||||
// prometheus.MustRegister(hdFailures)
|
// Name: "cpu_temperature_celsius",
|
||||||
// }
|
// Help: "Current temperature of the CPU.",
|
||||||
|
// }),
|
||||||
|
// hdFailures: prometheus.NewCounterVec(
|
||||||
|
// prometheus.CounterOpts{
|
||||||
|
// Name: "hd_errors_total",
|
||||||
|
// Help: "Number of hard-disk errors.",
|
||||||
|
// },
|
||||||
|
// []string{"device"},
|
||||||
|
// ),
|
||||||
|
// }
|
||||||
|
// reg.MustRegister(m.cpuTemp)
|
||||||
|
// reg.MustRegister(m.hdFailures)
|
||||||
|
// return m
|
||||||
|
// }
|
||||||
//
|
//
|
||||||
// func main() {
|
// func main() {
|
||||||
// cpuTemp.Set(65.3)
|
// // Create a non-global registry.
|
||||||
// hdFailures.With(prometheus.Labels{"device":"/dev/sda"}).Inc()
|
// reg := prometheus.NewRegistry()
|
||||||
//
|
//
|
||||||
// // The Handler function provides a default handler to expose metrics
|
// // Create new metrics and register them using the custom registry.
|
||||||
// // via an HTTP server. "/metrics" is the usual endpoint for that.
|
// m := NewMetrics(reg)
|
||||||
// http.Handle("/metrics", promhttp.Handler())
|
// // Set values for the new created metrics.
|
||||||
// log.Fatal(http.ListenAndServe(":8080", nil))
|
// m.cpuTemp.Set(65.3)
|
||||||
// }
|
// m.hdFailures.With(prometheus.Labels{"device":"/dev/sda"}).Inc()
|
||||||
//
|
//
|
||||||
|
// // Expose metrics and custom registry via an HTTP server
|
||||||
|
// // using the HandleFor function. "/metrics" is the usual endpoint for that.
|
||||||
|
// http.Handle("/metrics", promhttp.HandlerFor(reg, promhttp.HandlerOpts{Registry: reg}))
|
||||||
|
// log.Fatal(http.ListenAndServe(":8080", nil))
|
||||||
|
// }
|
||||||
//
|
//
|
||||||
// This is a complete program that exports two metrics, a Gauge and a Counter,
|
// This is a complete program that exports two metrics, a Gauge and a Counter,
|
||||||
// the latter with a label attached to turn it into a (one-dimensional) vector.
|
// the latter with a label attached to turn it into a (one-dimensional) vector.
|
||||||
|
// It register the metrics using a custom registry and exposes them via an HTTP server
|
||||||
|
// on the /metrics endpoint.
|
||||||
//
|
//
|
||||||
// Metrics
|
// # Metrics
|
||||||
//
|
//
|
||||||
// The number of exported identifiers in this package might appear a bit
|
// The number of exported identifiers in this package might appear a bit
|
||||||
// overwhelming. However, in addition to the basic plumbing shown in the example
|
// overwhelming. However, in addition to the basic plumbing shown in the example
|
||||||
@ -100,7 +111,7 @@
|
|||||||
// To create instances of Metrics and their vector versions, you need a suitable
|
// To create instances of Metrics and their vector versions, you need a suitable
|
||||||
// …Opts struct, i.e. GaugeOpts, CounterOpts, SummaryOpts, or HistogramOpts.
|
// …Opts struct, i.e. GaugeOpts, CounterOpts, SummaryOpts, or HistogramOpts.
|
||||||
//
|
//
|
||||||
// Custom Collectors and constant Metrics
|
// # Custom Collectors and constant Metrics
|
||||||
//
|
//
|
||||||
// While you could create your own implementations of Metric, most likely you
|
// While you could create your own implementations of Metric, most likely you
|
||||||
// will only ever implement the Collector interface on your own. At a first
|
// will only ever implement the Collector interface on your own. At a first
|
||||||
@ -141,7 +152,7 @@
|
|||||||
// a metric, GaugeFunc, CounterFunc, or UntypedFunc might be interesting
|
// a metric, GaugeFunc, CounterFunc, or UntypedFunc might be interesting
|
||||||
// shortcuts.
|
// shortcuts.
|
||||||
//
|
//
|
||||||
// Advanced Uses of the Registry
|
// # Advanced Uses of the Registry
|
||||||
//
|
//
|
||||||
// While MustRegister is the by far most common way of registering a Collector,
|
// While MustRegister is the by far most common way of registering a Collector,
|
||||||
// sometimes you might want to handle the errors the registration might cause.
|
// sometimes you might want to handle the errors the registration might cause.
|
||||||
@ -176,23 +187,23 @@
|
|||||||
// NewProcessCollector). With a custom registry, you are in control and decide
|
// NewProcessCollector). With a custom registry, you are in control and decide
|
||||||
// yourself about the Collectors to register.
|
// yourself about the Collectors to register.
|
||||||
//
|
//
|
||||||
// HTTP Exposition
|
// # HTTP Exposition
|
||||||
//
|
//
|
||||||
// The Registry implements the Gatherer interface. The caller of the Gather
|
// The Registry implements the Gatherer interface. The caller of the Gather
|
||||||
// method can then expose the gathered metrics in some way. Usually, the metrics
|
// method can then expose the gathered metrics in some way. Usually, the metrics
|
||||||
// are served via HTTP on the /metrics endpoint. That's happening in the example
|
// are served via HTTP on the /metrics endpoint. That's happening in the example
|
||||||
// above. The tools to expose metrics via HTTP are in the promhttp sub-package.
|
// above. The tools to expose metrics via HTTP are in the promhttp sub-package.
|
||||||
//
|
//
|
||||||
// Pushing to the Pushgateway
|
// # Pushing to the Pushgateway
|
||||||
//
|
//
|
||||||
// Function for pushing to the Pushgateway can be found in the push sub-package.
|
// Function for pushing to the Pushgateway can be found in the push sub-package.
|
||||||
//
|
//
|
||||||
// Graphite Bridge
|
// # Graphite Bridge
|
||||||
//
|
//
|
||||||
// Functions and examples to push metrics from a Gatherer to Graphite can be
|
// Functions and examples to push metrics from a Gatherer to Graphite can be
|
||||||
// found in the graphite sub-package.
|
// found in the graphite sub-package.
|
||||||
//
|
//
|
||||||
// Other Means of Exposition
|
// # Other Means of Exposition
|
||||||
//
|
//
|
||||||
// More ways of exposing metrics can easily be added by following the approaches
|
// More ways of exposing metrics can easily be added by following the approaches
|
||||||
// of the existing implementations.
|
// of the existing implementations.
|
||||||
|
6
vendor/github.com/prometheus/client_golang/prometheus/gauge.go
generated
vendored
6
vendor/github.com/prometheus/client_golang/prometheus/gauge.go
generated
vendored
@ -210,7 +210,8 @@ func (v *GaugeVec) GetMetricWith(labels Labels) (Gauge, error) {
|
|||||||
// WithLabelValues works as GetMetricWithLabelValues, but panics where
|
// WithLabelValues works as GetMetricWithLabelValues, but panics where
|
||||||
// GetMetricWithLabelValues would have returned an error. Not returning an
|
// GetMetricWithLabelValues would have returned an error. Not returning an
|
||||||
// error allows shortcuts like
|
// error allows shortcuts like
|
||||||
// myVec.WithLabelValues("404", "GET").Add(42)
|
//
|
||||||
|
// myVec.WithLabelValues("404", "GET").Add(42)
|
||||||
func (v *GaugeVec) WithLabelValues(lvs ...string) Gauge {
|
func (v *GaugeVec) WithLabelValues(lvs ...string) Gauge {
|
||||||
g, err := v.GetMetricWithLabelValues(lvs...)
|
g, err := v.GetMetricWithLabelValues(lvs...)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -221,7 +222,8 @@ func (v *GaugeVec) WithLabelValues(lvs ...string) Gauge {
|
|||||||
|
|
||||||
// With works as GetMetricWith, but panics where GetMetricWithLabels would have
|
// With works as GetMetricWith, but panics where GetMetricWithLabels would have
|
||||||
// returned an error. Not returning an error allows shortcuts like
|
// returned an error. Not returning an error allows shortcuts like
|
||||||
// myVec.With(prometheus.Labels{"code": "404", "method": "GET"}).Add(42)
|
//
|
||||||
|
// myVec.With(prometheus.Labels{"code": "404", "method": "GET"}).Add(42)
|
||||||
func (v *GaugeVec) With(labels Labels) Gauge {
|
func (v *GaugeVec) With(labels Labels) Gauge {
|
||||||
g, err := v.GetMetricWith(labels)
|
g, err := v.GetMetricWith(labels)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
// Copyright 2021 The Prometheus Authors
|
// Copyright 2015 The Prometheus Authors
|
||||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
// you may not use this file except in compliance with the License.
|
// you may not use this file except in compliance with the License.
|
||||||
// You may obtain a copy of the License at
|
// You may obtain a copy of the License at
|
||||||
@ -11,17 +11,16 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
//go:build !go1.15
|
//go:build !js || wasm
|
||||||
// +build !go1.15
|
// +build !js wasm
|
||||||
|
|
||||||
package collectors
|
package prometheus
|
||||||
|
|
||||||
import (
|
import "os"
|
||||||
"database/sql"
|
|
||||||
|
|
||||||
"github.com/prometheus/client_golang/prometheus"
|
func getPIDFn() func() (int, error) {
|
||||||
)
|
pid := os.Getpid()
|
||||||
|
return func() (int, error) {
|
||||||
func (c *dbStatsCollector) describeNewInGo115(ch chan<- *prometheus.Desc) {}
|
return pid, nil
|
||||||
|
}
|
||||||
func (c *dbStatsCollector) collectNewInGo115(ch chan<- prometheus.Metric, stats sql.DBStats) {}
|
}
|
23
vendor/github.com/prometheus/client_golang/prometheus/get_pid_gopherjs.go
generated
vendored
Normal file
23
vendor/github.com/prometheus/client_golang/prometheus/get_pid_gopherjs.go
generated
vendored
Normal file
@ -0,0 +1,23 @@
|
|||||||
|
// Copyright 2015 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build js && !wasm
|
||||||
|
// +build js,!wasm
|
||||||
|
|
||||||
|
package prometheus
|
||||||
|
|
||||||
|
func getPIDFn() func() (int, error) {
|
||||||
|
return func() (int, error) {
|
||||||
|
return 1, nil
|
||||||
|
}
|
||||||
|
}
|
12
vendor/github.com/prometheus/client_golang/prometheus/go_collector.go
generated
vendored
12
vendor/github.com/prometheus/client_golang/prometheus/go_collector.go
generated
vendored
@ -19,6 +19,10 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// goRuntimeMemStats provides the metrics initially provided by runtime.ReadMemStats.
|
||||||
|
// From Go 1.17 those similar (and better) statistics are provided by runtime/metrics, so
|
||||||
|
// while eval closure works on runtime.MemStats, the struct from Go 1.17+ is
|
||||||
|
// populated using runtime/metrics.
|
||||||
func goRuntimeMemStats() memStatsMetrics {
|
func goRuntimeMemStats() memStatsMetrics {
|
||||||
return memStatsMetrics{
|
return memStatsMetrics{
|
||||||
{
|
{
|
||||||
@ -224,7 +228,7 @@ func newBaseGoCollector() baseGoCollector {
|
|||||||
"A summary of the pause duration of garbage collection cycles.",
|
"A summary of the pause duration of garbage collection cycles.",
|
||||||
nil, nil),
|
nil, nil),
|
||||||
gcLastTimeDesc: NewDesc(
|
gcLastTimeDesc: NewDesc(
|
||||||
memstatNamespace("last_gc_time_seconds"),
|
"go_memstats_last_gc_time_seconds",
|
||||||
"Number of seconds since 1970 of last garbage collection.",
|
"Number of seconds since 1970 of last garbage collection.",
|
||||||
nil, nil),
|
nil, nil),
|
||||||
goInfoDesc: NewDesc(
|
goInfoDesc: NewDesc(
|
||||||
@ -246,8 +250,9 @@ func (c *baseGoCollector) Describe(ch chan<- *Desc) {
|
|||||||
// Collect returns the current state of all metrics of the collector.
|
// Collect returns the current state of all metrics of the collector.
|
||||||
func (c *baseGoCollector) Collect(ch chan<- Metric) {
|
func (c *baseGoCollector) Collect(ch chan<- Metric) {
|
||||||
ch <- MustNewConstMetric(c.goroutinesDesc, GaugeValue, float64(runtime.NumGoroutine()))
|
ch <- MustNewConstMetric(c.goroutinesDesc, GaugeValue, float64(runtime.NumGoroutine()))
|
||||||
n, _ := runtime.ThreadCreateProfile(nil)
|
|
||||||
ch <- MustNewConstMetric(c.threadsDesc, GaugeValue, float64(n))
|
n := getRuntimeNumThreads()
|
||||||
|
ch <- MustNewConstMetric(c.threadsDesc, GaugeValue, n)
|
||||||
|
|
||||||
var stats debug.GCStats
|
var stats debug.GCStats
|
||||||
stats.PauseQuantiles = make([]time.Duration, 5)
|
stats.PauseQuantiles = make([]time.Duration, 5)
|
||||||
@ -269,7 +274,6 @@ func memstatNamespace(s string) string {
|
|||||||
|
|
||||||
// memStatsMetrics provide description, evaluator, runtime/metrics name, and
|
// memStatsMetrics provide description, evaluator, runtime/metrics name, and
|
||||||
// value type for memstat metrics.
|
// value type for memstat metrics.
|
||||||
// TODO(bwplotka): Remove with end Go 1.16 EOL and replace with runtime/metrics.Description
|
|
||||||
type memStatsMetrics []struct {
|
type memStatsMetrics []struct {
|
||||||
desc *Desc
|
desc *Desc
|
||||||
eval func(*runtime.MemStats) float64
|
eval func(*runtime.MemStats) float64
|
||||||
|
249
vendor/github.com/prometheus/client_golang/prometheus/go_collector_latest.go
generated
vendored
249
vendor/github.com/prometheus/client_golang/prometheus/go_collector_latest.go
generated
vendored
@ -31,9 +31,11 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
// constants for strings referenced more than once.
|
||||||
goGCHeapTinyAllocsObjects = "/gc/heap/tiny/allocs:objects"
|
goGCHeapTinyAllocsObjects = "/gc/heap/tiny/allocs:objects"
|
||||||
goGCHeapAllocsObjects = "/gc/heap/allocs:objects"
|
goGCHeapAllocsObjects = "/gc/heap/allocs:objects"
|
||||||
goGCHeapFreesObjects = "/gc/heap/frees:objects"
|
goGCHeapFreesObjects = "/gc/heap/frees:objects"
|
||||||
|
goGCHeapFreesBytes = "/gc/heap/frees:bytes"
|
||||||
goGCHeapAllocsBytes = "/gc/heap/allocs:bytes"
|
goGCHeapAllocsBytes = "/gc/heap/allocs:bytes"
|
||||||
goGCHeapObjects = "/gc/heap/objects:objects"
|
goGCHeapObjects = "/gc/heap/objects:objects"
|
||||||
goGCHeapGoalBytes = "/gc/heap/goal:bytes"
|
goGCHeapGoalBytes = "/gc/heap/goal:bytes"
|
||||||
@ -53,8 +55,9 @@ const (
|
|||||||
goMemoryClassesOtherBytes = "/memory/classes/other:bytes"
|
goMemoryClassesOtherBytes = "/memory/classes/other:bytes"
|
||||||
)
|
)
|
||||||
|
|
||||||
// runtime/metrics names required for runtimeMemStats like logic.
|
// rmNamesForMemStatsMetrics represents runtime/metrics names required to populate goRuntimeMemStats from like logic.
|
||||||
var rmForMemStats = []string{goGCHeapTinyAllocsObjects,
|
var rmNamesForMemStatsMetrics = []string{
|
||||||
|
goGCHeapTinyAllocsObjects,
|
||||||
goGCHeapAllocsObjects,
|
goGCHeapAllocsObjects,
|
||||||
goGCHeapFreesObjects,
|
goGCHeapFreesObjects,
|
||||||
goGCHeapAllocsBytes,
|
goGCHeapAllocsBytes,
|
||||||
@ -89,74 +92,90 @@ func bestEffortLookupRM(lookup []string) []metrics.Description {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type goCollector struct {
|
type goCollector struct {
|
||||||
opt GoCollectorOptions
|
|
||||||
base baseGoCollector
|
base baseGoCollector
|
||||||
|
|
||||||
// mu protects updates to all fields ensuring a consistent
|
// mu protects updates to all fields ensuring a consistent
|
||||||
// snapshot is always produced by Collect.
|
// snapshot is always produced by Collect.
|
||||||
mu sync.Mutex
|
mu sync.Mutex
|
||||||
|
|
||||||
// rm... fields all pertain to the runtime/metrics package.
|
// Contains all samples that has to retrieved from runtime/metrics (not all of them will be exposed).
|
||||||
rmSampleBuf []metrics.Sample
|
sampleBuf []metrics.Sample
|
||||||
rmSampleMap map[string]*metrics.Sample
|
// sampleMap allows lookup for MemStats metrics and runtime/metrics histograms for exact sums.
|
||||||
rmMetrics []collectorMetric
|
sampleMap map[string]*metrics.Sample
|
||||||
|
|
||||||
|
// rmExposedMetrics represents all runtime/metrics package metrics
|
||||||
|
// that were configured to be exposed.
|
||||||
|
rmExposedMetrics []collectorMetric
|
||||||
|
rmExactSumMapForHist map[string]string
|
||||||
|
|
||||||
// With Go 1.17, the runtime/metrics package was introduced.
|
// With Go 1.17, the runtime/metrics package was introduced.
|
||||||
// From that point on, metric names produced by the runtime/metrics
|
// From that point on, metric names produced by the runtime/metrics
|
||||||
// package could be generated from runtime/metrics names. However,
|
// package could be generated from runtime/metrics names. However,
|
||||||
// these differ from the old names for the same values.
|
// these differ from the old names for the same values.
|
||||||
//
|
//
|
||||||
// This field exist to export the same values under the old names
|
// This field exists to export the same values under the old names
|
||||||
// as well.
|
// as well.
|
||||||
msMetrics memStatsMetrics
|
msMetrics memStatsMetrics
|
||||||
|
msMetricsEnabled bool
|
||||||
}
|
}
|
||||||
|
|
||||||
const (
|
type rmMetricDesc struct {
|
||||||
// Those are not exposed due to need to move Go collector to another package in v2.
|
metrics.Description
|
||||||
// See issue https://github.com/prometheus/client_golang/issues/1030.
|
|
||||||
goRuntimeMemStatsCollection uint32 = 1 << iota
|
|
||||||
goRuntimeMetricsCollection
|
|
||||||
)
|
|
||||||
|
|
||||||
// GoCollectorOptions should not be used be directly by anything, except `collectors` package.
|
|
||||||
// Use it via collectors package instead. See issue
|
|
||||||
// https://github.com/prometheus/client_golang/issues/1030.
|
|
||||||
//
|
|
||||||
// Deprecated: Use collectors.WithGoCollections
|
|
||||||
type GoCollectorOptions struct {
|
|
||||||
// EnabledCollection sets what type of collections collector should expose on top of base collection.
|
|
||||||
// By default it's goMemStatsCollection | goRuntimeMetricsCollection.
|
|
||||||
EnabledCollections uint32
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c GoCollectorOptions) isEnabled(flag uint32) bool {
|
func matchRuntimeMetricsRules(rules []internal.GoCollectorRule) []rmMetricDesc {
|
||||||
return c.EnabledCollections&flag != 0
|
var descs []rmMetricDesc
|
||||||
|
for _, d := range metrics.All() {
|
||||||
|
var (
|
||||||
|
deny = true
|
||||||
|
desc rmMetricDesc
|
||||||
|
)
|
||||||
|
|
||||||
|
for _, r := range rules {
|
||||||
|
if !r.Matcher.MatchString(d.Name) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
deny = r.Deny
|
||||||
|
}
|
||||||
|
if deny {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
desc.Description = d
|
||||||
|
descs = append(descs, desc)
|
||||||
|
}
|
||||||
|
return descs
|
||||||
}
|
}
|
||||||
|
|
||||||
const defaultGoCollections = goRuntimeMemStatsCollection
|
func defaultGoCollectorOptions() internal.GoCollectorOptions {
|
||||||
|
return internal.GoCollectorOptions{
|
||||||
|
RuntimeMetricSumForHist: map[string]string{
|
||||||
|
"/gc/heap/allocs-by-size:bytes": goGCHeapAllocsBytes,
|
||||||
|
"/gc/heap/frees-by-size:bytes": goGCHeapFreesBytes,
|
||||||
|
},
|
||||||
|
RuntimeMetricRules: []internal.GoCollectorRule{
|
||||||
|
//{Matcher: regexp.MustCompile("")},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// NewGoCollector is the obsolete version of collectors.NewGoCollector.
|
// NewGoCollector is the obsolete version of collectors.NewGoCollector.
|
||||||
// See there for documentation.
|
// See there for documentation.
|
||||||
//
|
//
|
||||||
// Deprecated: Use collectors.NewGoCollector instead.
|
// Deprecated: Use collectors.NewGoCollector instead.
|
||||||
func NewGoCollector(opts ...func(o *GoCollectorOptions)) Collector {
|
func NewGoCollector(opts ...func(o *internal.GoCollectorOptions)) Collector {
|
||||||
opt := GoCollectorOptions{EnabledCollections: defaultGoCollections}
|
opt := defaultGoCollectorOptions()
|
||||||
for _, o := range opts {
|
for _, o := range opts {
|
||||||
o(&opt)
|
o(&opt)
|
||||||
}
|
}
|
||||||
|
|
||||||
var descriptions []metrics.Description
|
exposedDescriptions := matchRuntimeMetricsRules(opt.RuntimeMetricRules)
|
||||||
if opt.isEnabled(goRuntimeMetricsCollection) {
|
|
||||||
descriptions = metrics.All()
|
|
||||||
} else if opt.isEnabled(goRuntimeMemStatsCollection) {
|
|
||||||
descriptions = bestEffortLookupRM(rmForMemStats)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Collect all histogram samples so that we can get their buckets.
|
// Collect all histogram samples so that we can get their buckets.
|
||||||
// The API guarantees that the buckets are always fixed for the lifetime
|
// The API guarantees that the buckets are always fixed for the lifetime
|
||||||
// of the process.
|
// of the process.
|
||||||
var histograms []metrics.Sample
|
var histograms []metrics.Sample
|
||||||
for _, d := range descriptions {
|
for _, d := range exposedDescriptions {
|
||||||
if d.Kind == metrics.KindFloat64Histogram {
|
if d.Kind == metrics.KindFloat64Histogram {
|
||||||
histograms = append(histograms, metrics.Sample{Name: d.Name})
|
histograms = append(histograms, metrics.Sample{Name: d.Name})
|
||||||
}
|
}
|
||||||
@ -171,13 +190,14 @@ func NewGoCollector(opts ...func(o *GoCollectorOptions)) Collector {
|
|||||||
bucketsMap[histograms[i].Name] = histograms[i].Value.Float64Histogram().Buckets
|
bucketsMap[histograms[i].Name] = histograms[i].Value.Float64Histogram().Buckets
|
||||||
}
|
}
|
||||||
|
|
||||||
// Generate a Desc and ValueType for each runtime/metrics metric.
|
// Generate a collector for each exposed runtime/metrics metric.
|
||||||
metricSet := make([]collectorMetric, 0, len(descriptions))
|
metricSet := make([]collectorMetric, 0, len(exposedDescriptions))
|
||||||
sampleBuf := make([]metrics.Sample, 0, len(descriptions))
|
// SampleBuf is used for reading from runtime/metrics.
|
||||||
sampleMap := make(map[string]*metrics.Sample, len(descriptions))
|
// We are assuming the largest case to have stable pointers for sampleMap purposes.
|
||||||
for i := range descriptions {
|
sampleBuf := make([]metrics.Sample, 0, len(exposedDescriptions)+len(opt.RuntimeMetricSumForHist)+len(rmNamesForMemStatsMetrics))
|
||||||
d := &descriptions[i]
|
sampleMap := make(map[string]*metrics.Sample, len(exposedDescriptions))
|
||||||
namespace, subsystem, name, ok := internal.RuntimeMetricsToProm(d)
|
for _, d := range exposedDescriptions {
|
||||||
|
namespace, subsystem, name, ok := internal.RuntimeMetricsToProm(&d.Description)
|
||||||
if !ok {
|
if !ok {
|
||||||
// Just ignore this metric; we can't do anything with it here.
|
// Just ignore this metric; we can't do anything with it here.
|
||||||
// If a user decides to use the latest version of Go, we don't want
|
// If a user decides to use the latest version of Go, we don't want
|
||||||
@ -185,19 +205,17 @@ func NewGoCollector(opts ...func(o *GoCollectorOptions)) Collector {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
// Set up sample buffer for reading, and a map
|
|
||||||
// for quick lookup of sample values.
|
|
||||||
sampleBuf = append(sampleBuf, metrics.Sample{Name: d.Name})
|
sampleBuf = append(sampleBuf, metrics.Sample{Name: d.Name})
|
||||||
sampleMap[d.Name] = &sampleBuf[len(sampleBuf)-1]
|
sampleMap[d.Name] = &sampleBuf[len(sampleBuf)-1]
|
||||||
|
|
||||||
var m collectorMetric
|
var m collectorMetric
|
||||||
if d.Kind == metrics.KindFloat64Histogram {
|
if d.Kind == metrics.KindFloat64Histogram {
|
||||||
_, hasSum := rmExactSumMap[d.Name]
|
_, hasSum := opt.RuntimeMetricSumForHist[d.Name]
|
||||||
unit := d.Name[strings.IndexRune(d.Name, ':')+1:]
|
unit := d.Name[strings.IndexRune(d.Name, ':')+1:]
|
||||||
m = newBatchHistogram(
|
m = newBatchHistogram(
|
||||||
NewDesc(
|
NewDesc(
|
||||||
BuildFQName(namespace, subsystem, name),
|
BuildFQName(namespace, subsystem, name),
|
||||||
d.Description,
|
d.Description.Description,
|
||||||
nil,
|
nil,
|
||||||
nil,
|
nil,
|
||||||
),
|
),
|
||||||
@ -209,30 +227,61 @@ func NewGoCollector(opts ...func(o *GoCollectorOptions)) Collector {
|
|||||||
Namespace: namespace,
|
Namespace: namespace,
|
||||||
Subsystem: subsystem,
|
Subsystem: subsystem,
|
||||||
Name: name,
|
Name: name,
|
||||||
Help: d.Description,
|
Help: d.Description.Description,
|
||||||
})
|
},
|
||||||
|
)
|
||||||
} else {
|
} else {
|
||||||
m = NewGauge(GaugeOpts{
|
m = NewGauge(GaugeOpts{
|
||||||
Namespace: namespace,
|
Namespace: namespace,
|
||||||
Subsystem: subsystem,
|
Subsystem: subsystem,
|
||||||
Name: name,
|
Name: name,
|
||||||
Help: d.Description,
|
Help: d.Description.Description,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
metricSet = append(metricSet, m)
|
metricSet = append(metricSet, m)
|
||||||
}
|
}
|
||||||
|
|
||||||
var msMetrics memStatsMetrics
|
// Add exact sum metrics to sampleBuf if not added before.
|
||||||
if opt.isEnabled(goRuntimeMemStatsCollection) {
|
for _, h := range histograms {
|
||||||
msMetrics = goRuntimeMemStats()
|
sumMetric, ok := opt.RuntimeMetricSumForHist[h.Name]
|
||||||
|
if !ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if _, ok := sampleMap[sumMetric]; ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
sampleBuf = append(sampleBuf, metrics.Sample{Name: sumMetric})
|
||||||
|
sampleMap[sumMetric] = &sampleBuf[len(sampleBuf)-1]
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
msMetrics memStatsMetrics
|
||||||
|
msDescriptions []metrics.Description
|
||||||
|
)
|
||||||
|
|
||||||
|
if !opt.DisableMemStatsLikeMetrics {
|
||||||
|
msMetrics = goRuntimeMemStats()
|
||||||
|
msDescriptions = bestEffortLookupRM(rmNamesForMemStatsMetrics)
|
||||||
|
|
||||||
|
// Check if metric was not exposed before and if not, add to sampleBuf.
|
||||||
|
for _, mdDesc := range msDescriptions {
|
||||||
|
if _, ok := sampleMap[mdDesc.Name]; ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
sampleBuf = append(sampleBuf, metrics.Sample{Name: mdDesc.Name})
|
||||||
|
sampleMap[mdDesc.Name] = &sampleBuf[len(sampleBuf)-1]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
return &goCollector{
|
return &goCollector{
|
||||||
opt: opt,
|
base: newBaseGoCollector(),
|
||||||
base: newBaseGoCollector(),
|
sampleBuf: sampleBuf,
|
||||||
rmSampleBuf: sampleBuf,
|
sampleMap: sampleMap,
|
||||||
rmSampleMap: sampleMap,
|
rmExposedMetrics: metricSet,
|
||||||
rmMetrics: metricSet,
|
rmExactSumMapForHist: opt.RuntimeMetricSumForHist,
|
||||||
msMetrics: msMetrics,
|
msMetrics: msMetrics,
|
||||||
|
msMetricsEnabled: !opt.DisableMemStatsLikeMetrics,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -242,7 +291,7 @@ func (c *goCollector) Describe(ch chan<- *Desc) {
|
|||||||
for _, i := range c.msMetrics {
|
for _, i := range c.msMetrics {
|
||||||
ch <- i.desc
|
ch <- i.desc
|
||||||
}
|
}
|
||||||
for _, m := range c.rmMetrics {
|
for _, m := range c.rmExposedMetrics {
|
||||||
ch <- m.Desc()
|
ch <- m.Desc()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -252,8 +301,12 @@ func (c *goCollector) Collect(ch chan<- Metric) {
|
|||||||
// Collect base non-memory metrics.
|
// Collect base non-memory metrics.
|
||||||
c.base.Collect(ch)
|
c.base.Collect(ch)
|
||||||
|
|
||||||
|
if len(c.sampleBuf) == 0 {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
// Collect must be thread-safe, so prevent concurrent use of
|
// Collect must be thread-safe, so prevent concurrent use of
|
||||||
// rmSampleBuf. Just read into rmSampleBuf but write all the data
|
// sampleBuf elements. Just read into sampleBuf but write all the data
|
||||||
// we get into our Metrics or MemStats.
|
// we get into our Metrics or MemStats.
|
||||||
//
|
//
|
||||||
// This lock also ensures that the Metrics we send out are all from
|
// This lock also ensures that the Metrics we send out are all from
|
||||||
@ -267,44 +320,43 @@ func (c *goCollector) Collect(ch chan<- Metric) {
|
|||||||
c.mu.Lock()
|
c.mu.Lock()
|
||||||
defer c.mu.Unlock()
|
defer c.mu.Unlock()
|
||||||
|
|
||||||
if len(c.rmSampleBuf) > 0 {
|
// Populate runtime/metrics sample buffer.
|
||||||
// Populate runtime/metrics sample buffer.
|
metrics.Read(c.sampleBuf)
|
||||||
metrics.Read(c.rmSampleBuf)
|
|
||||||
}
|
|
||||||
|
|
||||||
if c.opt.isEnabled(goRuntimeMetricsCollection) {
|
// Collect all our runtime/metrics user chose to expose from sampleBuf (if any).
|
||||||
// Collect all our metrics from rmSampleBuf.
|
for i, metric := range c.rmExposedMetrics {
|
||||||
for i, sample := range c.rmSampleBuf {
|
// We created samples for exposed metrics first in order, so indexes match.
|
||||||
// N.B. switch on concrete type because it's significantly more efficient
|
sample := c.sampleBuf[i]
|
||||||
// than checking for the Counter and Gauge interface implementations. In
|
|
||||||
// this case, we control all the types here.
|
// N.B. switch on concrete type because it's significantly more efficient
|
||||||
switch m := c.rmMetrics[i].(type) {
|
// than checking for the Counter and Gauge interface implementations. In
|
||||||
case *counter:
|
// this case, we control all the types here.
|
||||||
// Guard against decreases. This should never happen, but a failure
|
switch m := metric.(type) {
|
||||||
// to do so will result in a panic, which is a harsh consequence for
|
case *counter:
|
||||||
// a metrics collection bug.
|
// Guard against decreases. This should never happen, but a failure
|
||||||
v0, v1 := m.get(), unwrapScalarRMValue(sample.Value)
|
// to do so will result in a panic, which is a harsh consequence for
|
||||||
if v1 > v0 {
|
// a metrics collection bug.
|
||||||
m.Add(unwrapScalarRMValue(sample.Value) - m.get())
|
v0, v1 := m.get(), unwrapScalarRMValue(sample.Value)
|
||||||
}
|
if v1 > v0 {
|
||||||
m.Collect(ch)
|
m.Add(unwrapScalarRMValue(sample.Value) - m.get())
|
||||||
case *gauge:
|
|
||||||
m.Set(unwrapScalarRMValue(sample.Value))
|
|
||||||
m.Collect(ch)
|
|
||||||
case *batchHistogram:
|
|
||||||
m.update(sample.Value.Float64Histogram(), c.exactSumFor(sample.Name))
|
|
||||||
m.Collect(ch)
|
|
||||||
default:
|
|
||||||
panic("unexpected metric type")
|
|
||||||
}
|
}
|
||||||
|
m.Collect(ch)
|
||||||
|
case *gauge:
|
||||||
|
m.Set(unwrapScalarRMValue(sample.Value))
|
||||||
|
m.Collect(ch)
|
||||||
|
case *batchHistogram:
|
||||||
|
m.update(sample.Value.Float64Histogram(), c.exactSumFor(sample.Name))
|
||||||
|
m.Collect(ch)
|
||||||
|
default:
|
||||||
|
panic("unexpected metric type")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// ms is a dummy MemStats that we populate ourselves so that we can
|
if c.msMetricsEnabled {
|
||||||
// populate the old metrics from it if goMemStatsCollection is enabled.
|
// ms is a dummy MemStats that we populate ourselves so that we can
|
||||||
if c.opt.isEnabled(goRuntimeMemStatsCollection) {
|
// populate the old metrics from it if goMemStatsCollection is enabled.
|
||||||
var ms runtime.MemStats
|
var ms runtime.MemStats
|
||||||
memStatsFromRM(&ms, c.rmSampleMap)
|
memStatsFromRM(&ms, c.sampleMap)
|
||||||
for _, i := range c.msMetrics {
|
for _, i := range c.msMetrics {
|
||||||
ch <- MustNewConstMetric(i.desc, i.valType, i.eval(&ms))
|
ch <- MustNewConstMetric(i.desc, i.valType, i.eval(&ms))
|
||||||
}
|
}
|
||||||
@ -335,11 +387,6 @@ func unwrapScalarRMValue(v metrics.Value) float64 {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
var rmExactSumMap = map[string]string{
|
|
||||||
"/gc/heap/allocs-by-size:bytes": "/gc/heap/allocs:bytes",
|
|
||||||
"/gc/heap/frees-by-size:bytes": "/gc/heap/frees:bytes",
|
|
||||||
}
|
|
||||||
|
|
||||||
// exactSumFor takes a runtime/metrics metric name (that is assumed to
|
// exactSumFor takes a runtime/metrics metric name (that is assumed to
|
||||||
// be of kind KindFloat64Histogram) and returns its exact sum and whether
|
// be of kind KindFloat64Histogram) and returns its exact sum and whether
|
||||||
// its exact sum exists.
|
// its exact sum exists.
|
||||||
@ -347,11 +394,11 @@ var rmExactSumMap = map[string]string{
|
|||||||
// The runtime/metrics API for histograms doesn't currently expose exact
|
// The runtime/metrics API for histograms doesn't currently expose exact
|
||||||
// sums, but some of the other metrics are in fact exact sums of histograms.
|
// sums, but some of the other metrics are in fact exact sums of histograms.
|
||||||
func (c *goCollector) exactSumFor(rmName string) float64 {
|
func (c *goCollector) exactSumFor(rmName string) float64 {
|
||||||
sumName, ok := rmExactSumMap[rmName]
|
sumName, ok := c.rmExactSumMapForHist[rmName]
|
||||||
if !ok {
|
if !ok {
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
s, ok := c.rmSampleMap[sumName]
|
s, ok := c.sampleMap[sumName]
|
||||||
if !ok {
|
if !ok {
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
978
vendor/github.com/prometheus/client_golang/prometheus/histogram.go
generated
vendored
978
vendor/github.com/prometheus/client_golang/prometheus/histogram.go
generated
vendored
File diff suppressed because it is too large
Load Diff
60
vendor/github.com/prometheus/client_golang/prometheus/internal/almost_equal.go
generated
vendored
Normal file
60
vendor/github.com/prometheus/client_golang/prometheus/internal/almost_equal.go
generated
vendored
Normal file
@ -0,0 +1,60 @@
|
|||||||
|
// Copyright (c) 2015 Björn Rabenstein
|
||||||
|
//
|
||||||
|
// Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||||
|
// of this software and associated documentation files (the "Software"), to deal
|
||||||
|
// in the Software without restriction, including without limitation the rights
|
||||||
|
// to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||||
|
// copies of the Software, and to permit persons to whom the Software is
|
||||||
|
// furnished to do so, subject to the following conditions:
|
||||||
|
//
|
||||||
|
// The above copyright notice and this permission notice shall be included in all
|
||||||
|
// copies or substantial portions of the Software.
|
||||||
|
//
|
||||||
|
// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||||
|
// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||||
|
// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||||
|
// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||||
|
// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||||
|
// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||||
|
// SOFTWARE.
|
||||||
|
//
|
||||||
|
// The code in this package is copy/paste to avoid a dependency. Hence this file
|
||||||
|
// carries the copyright of the original repo.
|
||||||
|
// https://github.com/beorn7/floats
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import (
|
||||||
|
"math"
|
||||||
|
)
|
||||||
|
|
||||||
|
// minNormalFloat64 is the smallest positive normal value of type float64.
|
||||||
|
var minNormalFloat64 = math.Float64frombits(0x0010000000000000)
|
||||||
|
|
||||||
|
// AlmostEqualFloat64 returns true if a and b are equal within a relative error
|
||||||
|
// of epsilon. See http://floating-point-gui.de/errors/comparison/ for the
|
||||||
|
// details of the applied method.
|
||||||
|
func AlmostEqualFloat64(a, b, epsilon float64) bool {
|
||||||
|
if a == b {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
absA := math.Abs(a)
|
||||||
|
absB := math.Abs(b)
|
||||||
|
diff := math.Abs(a - b)
|
||||||
|
if a == 0 || b == 0 || absA+absB < minNormalFloat64 {
|
||||||
|
return diff < epsilon*minNormalFloat64
|
||||||
|
}
|
||||||
|
return diff/math.Min(absA+absB, math.MaxFloat64) < epsilon
|
||||||
|
}
|
||||||
|
|
||||||
|
// AlmostEqualFloat64s is the slice form of AlmostEqualFloat64.
|
||||||
|
func AlmostEqualFloat64s(a, b []float64, epsilon float64) bool {
|
||||||
|
if len(a) != len(b) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
for i := range a {
|
||||||
|
if !AlmostEqualFloat64(a[i], b[i], epsilon) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
654
vendor/github.com/prometheus/client_golang/prometheus/internal/difflib.go
generated
vendored
Normal file
654
vendor/github.com/prometheus/client_golang/prometheus/internal/difflib.go
generated
vendored
Normal file
@ -0,0 +1,654 @@
|
|||||||
|
// Copyright 2022 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
//
|
||||||
|
// It provides tools to compare sequences of strings and generate textual diffs.
|
||||||
|
//
|
||||||
|
// Maintaining `GetUnifiedDiffString` here because original repository
|
||||||
|
// (https://github.com/pmezard/go-difflib) is no loger maintained.
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
func min(a, b int) int {
|
||||||
|
if a < b {
|
||||||
|
return a
|
||||||
|
}
|
||||||
|
return b
|
||||||
|
}
|
||||||
|
|
||||||
|
func max(a, b int) int {
|
||||||
|
if a > b {
|
||||||
|
return a
|
||||||
|
}
|
||||||
|
return b
|
||||||
|
}
|
||||||
|
|
||||||
|
func calculateRatio(matches, length int) float64 {
|
||||||
|
if length > 0 {
|
||||||
|
return 2.0 * float64(matches) / float64(length)
|
||||||
|
}
|
||||||
|
return 1.0
|
||||||
|
}
|
||||||
|
|
||||||
|
type Match struct {
|
||||||
|
A int
|
||||||
|
B int
|
||||||
|
Size int
|
||||||
|
}
|
||||||
|
|
||||||
|
type OpCode struct {
|
||||||
|
Tag byte
|
||||||
|
I1 int
|
||||||
|
I2 int
|
||||||
|
J1 int
|
||||||
|
J2 int
|
||||||
|
}
|
||||||
|
|
||||||
|
// SequenceMatcher compares sequence of strings. The basic
|
||||||
|
// algorithm predates, and is a little fancier than, an algorithm
|
||||||
|
// published in the late 1980's by Ratcliff and Obershelp under the
|
||||||
|
// hyperbolic name "gestalt pattern matching". The basic idea is to find
|
||||||
|
// the longest contiguous matching subsequence that contains no "junk"
|
||||||
|
// elements (R-O doesn't address junk). The same idea is then applied
|
||||||
|
// recursively to the pieces of the sequences to the left and to the right
|
||||||
|
// of the matching subsequence. This does not yield minimal edit
|
||||||
|
// sequences, but does tend to yield matches that "look right" to people.
|
||||||
|
//
|
||||||
|
// SequenceMatcher tries to compute a "human-friendly diff" between two
|
||||||
|
// sequences. Unlike e.g. UNIX(tm) diff, the fundamental notion is the
|
||||||
|
// longest *contiguous* & junk-free matching subsequence. That's what
|
||||||
|
// catches peoples' eyes. The Windows(tm) windiff has another interesting
|
||||||
|
// notion, pairing up elements that appear uniquely in each sequence.
|
||||||
|
// That, and the method here, appear to yield more intuitive difference
|
||||||
|
// reports than does diff. This method appears to be the least vulnerable
|
||||||
|
// to synching up on blocks of "junk lines", though (like blank lines in
|
||||||
|
// ordinary text files, or maybe "<P>" lines in HTML files). That may be
|
||||||
|
// because this is the only method of the 3 that has a *concept* of
|
||||||
|
// "junk" <wink>.
|
||||||
|
//
|
||||||
|
// Timing: Basic R-O is cubic time worst case and quadratic time expected
|
||||||
|
// case. SequenceMatcher is quadratic time for the worst case and has
|
||||||
|
// expected-case behavior dependent in a complicated way on how many
|
||||||
|
// elements the sequences have in common; best case time is linear.
|
||||||
|
type SequenceMatcher struct {
|
||||||
|
a []string
|
||||||
|
b []string
|
||||||
|
b2j map[string][]int
|
||||||
|
IsJunk func(string) bool
|
||||||
|
autoJunk bool
|
||||||
|
bJunk map[string]struct{}
|
||||||
|
matchingBlocks []Match
|
||||||
|
fullBCount map[string]int
|
||||||
|
bPopular map[string]struct{}
|
||||||
|
opCodes []OpCode
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewMatcher(a, b []string) *SequenceMatcher {
|
||||||
|
m := SequenceMatcher{autoJunk: true}
|
||||||
|
m.SetSeqs(a, b)
|
||||||
|
return &m
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewMatcherWithJunk(a, b []string, autoJunk bool,
|
||||||
|
isJunk func(string) bool,
|
||||||
|
) *SequenceMatcher {
|
||||||
|
m := SequenceMatcher{IsJunk: isJunk, autoJunk: autoJunk}
|
||||||
|
m.SetSeqs(a, b)
|
||||||
|
return &m
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set two sequences to be compared.
|
||||||
|
func (m *SequenceMatcher) SetSeqs(a, b []string) {
|
||||||
|
m.SetSeq1(a)
|
||||||
|
m.SetSeq2(b)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set the first sequence to be compared. The second sequence to be compared is
|
||||||
|
// not changed.
|
||||||
|
//
|
||||||
|
// SequenceMatcher computes and caches detailed information about the second
|
||||||
|
// sequence, so if you want to compare one sequence S against many sequences,
|
||||||
|
// use .SetSeq2(s) once and call .SetSeq1(x) repeatedly for each of the other
|
||||||
|
// sequences.
|
||||||
|
//
|
||||||
|
// See also SetSeqs() and SetSeq2().
|
||||||
|
func (m *SequenceMatcher) SetSeq1(a []string) {
|
||||||
|
if &a == &m.a {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
m.a = a
|
||||||
|
m.matchingBlocks = nil
|
||||||
|
m.opCodes = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set the second sequence to be compared. The first sequence to be compared is
|
||||||
|
// not changed.
|
||||||
|
func (m *SequenceMatcher) SetSeq2(b []string) {
|
||||||
|
if &b == &m.b {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
m.b = b
|
||||||
|
m.matchingBlocks = nil
|
||||||
|
m.opCodes = nil
|
||||||
|
m.fullBCount = nil
|
||||||
|
m.chainB()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *SequenceMatcher) chainB() {
|
||||||
|
// Populate line -> index mapping
|
||||||
|
b2j := map[string][]int{}
|
||||||
|
for i, s := range m.b {
|
||||||
|
indices := b2j[s]
|
||||||
|
indices = append(indices, i)
|
||||||
|
b2j[s] = indices
|
||||||
|
}
|
||||||
|
|
||||||
|
// Purge junk elements
|
||||||
|
m.bJunk = map[string]struct{}{}
|
||||||
|
if m.IsJunk != nil {
|
||||||
|
junk := m.bJunk
|
||||||
|
for s := range b2j {
|
||||||
|
if m.IsJunk(s) {
|
||||||
|
junk[s] = struct{}{}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for s := range junk {
|
||||||
|
delete(b2j, s)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Purge remaining popular elements
|
||||||
|
popular := map[string]struct{}{}
|
||||||
|
n := len(m.b)
|
||||||
|
if m.autoJunk && n >= 200 {
|
||||||
|
ntest := n/100 + 1
|
||||||
|
for s, indices := range b2j {
|
||||||
|
if len(indices) > ntest {
|
||||||
|
popular[s] = struct{}{}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for s := range popular {
|
||||||
|
delete(b2j, s)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
m.bPopular = popular
|
||||||
|
m.b2j = b2j
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *SequenceMatcher) isBJunk(s string) bool {
|
||||||
|
_, ok := m.bJunk[s]
|
||||||
|
return ok
|
||||||
|
}
|
||||||
|
|
||||||
|
// Find longest matching block in a[alo:ahi] and b[blo:bhi].
|
||||||
|
//
|
||||||
|
// If IsJunk is not defined:
|
||||||
|
//
|
||||||
|
// Return (i,j,k) such that a[i:i+k] is equal to b[j:j+k], where
|
||||||
|
//
|
||||||
|
// alo <= i <= i+k <= ahi
|
||||||
|
// blo <= j <= j+k <= bhi
|
||||||
|
//
|
||||||
|
// and for all (i',j',k') meeting those conditions,
|
||||||
|
//
|
||||||
|
// k >= k'
|
||||||
|
// i <= i'
|
||||||
|
// and if i == i', j <= j'
|
||||||
|
//
|
||||||
|
// In other words, of all maximal matching blocks, return one that
|
||||||
|
// starts earliest in a, and of all those maximal matching blocks that
|
||||||
|
// start earliest in a, return the one that starts earliest in b.
|
||||||
|
//
|
||||||
|
// If IsJunk is defined, first the longest matching block is
|
||||||
|
// determined as above, but with the additional restriction that no
|
||||||
|
// junk element appears in the block. Then that block is extended as
|
||||||
|
// far as possible by matching (only) junk elements on both sides. So
|
||||||
|
// the resulting block never matches on junk except as identical junk
|
||||||
|
// happens to be adjacent to an "interesting" match.
|
||||||
|
//
|
||||||
|
// If no blocks match, return (alo, blo, 0).
|
||||||
|
func (m *SequenceMatcher) findLongestMatch(alo, ahi, blo, bhi int) Match {
|
||||||
|
// CAUTION: stripping common prefix or suffix would be incorrect.
|
||||||
|
// E.g.,
|
||||||
|
// ab
|
||||||
|
// acab
|
||||||
|
// Longest matching block is "ab", but if common prefix is
|
||||||
|
// stripped, it's "a" (tied with "b"). UNIX(tm) diff does so
|
||||||
|
// strip, so ends up claiming that ab is changed to acab by
|
||||||
|
// inserting "ca" in the middle. That's minimal but unintuitive:
|
||||||
|
// "it's obvious" that someone inserted "ac" at the front.
|
||||||
|
// Windiff ends up at the same place as diff, but by pairing up
|
||||||
|
// the unique 'b's and then matching the first two 'a's.
|
||||||
|
besti, bestj, bestsize := alo, blo, 0
|
||||||
|
|
||||||
|
// find longest junk-free match
|
||||||
|
// during an iteration of the loop, j2len[j] = length of longest
|
||||||
|
// junk-free match ending with a[i-1] and b[j]
|
||||||
|
j2len := map[int]int{}
|
||||||
|
for i := alo; i != ahi; i++ {
|
||||||
|
// look at all instances of a[i] in b; note that because
|
||||||
|
// b2j has no junk keys, the loop is skipped if a[i] is junk
|
||||||
|
newj2len := map[int]int{}
|
||||||
|
for _, j := range m.b2j[m.a[i]] {
|
||||||
|
// a[i] matches b[j]
|
||||||
|
if j < blo {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if j >= bhi {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
k := j2len[j-1] + 1
|
||||||
|
newj2len[j] = k
|
||||||
|
if k > bestsize {
|
||||||
|
besti, bestj, bestsize = i-k+1, j-k+1, k
|
||||||
|
}
|
||||||
|
}
|
||||||
|
j2len = newj2len
|
||||||
|
}
|
||||||
|
|
||||||
|
// Extend the best by non-junk elements on each end. In particular,
|
||||||
|
// "popular" non-junk elements aren't in b2j, which greatly speeds
|
||||||
|
// the inner loop above, but also means "the best" match so far
|
||||||
|
// doesn't contain any junk *or* popular non-junk elements.
|
||||||
|
for besti > alo && bestj > blo && !m.isBJunk(m.b[bestj-1]) &&
|
||||||
|
m.a[besti-1] == m.b[bestj-1] {
|
||||||
|
besti, bestj, bestsize = besti-1, bestj-1, bestsize+1
|
||||||
|
}
|
||||||
|
for besti+bestsize < ahi && bestj+bestsize < bhi &&
|
||||||
|
!m.isBJunk(m.b[bestj+bestsize]) &&
|
||||||
|
m.a[besti+bestsize] == m.b[bestj+bestsize] {
|
||||||
|
bestsize++
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now that we have a wholly interesting match (albeit possibly
|
||||||
|
// empty!), we may as well suck up the matching junk on each
|
||||||
|
// side of it too. Can't think of a good reason not to, and it
|
||||||
|
// saves post-processing the (possibly considerable) expense of
|
||||||
|
// figuring out what to do with it. In the case of an empty
|
||||||
|
// interesting match, this is clearly the right thing to do,
|
||||||
|
// because no other kind of match is possible in the regions.
|
||||||
|
for besti > alo && bestj > blo && m.isBJunk(m.b[bestj-1]) &&
|
||||||
|
m.a[besti-1] == m.b[bestj-1] {
|
||||||
|
besti, bestj, bestsize = besti-1, bestj-1, bestsize+1
|
||||||
|
}
|
||||||
|
for besti+bestsize < ahi && bestj+bestsize < bhi &&
|
||||||
|
m.isBJunk(m.b[bestj+bestsize]) &&
|
||||||
|
m.a[besti+bestsize] == m.b[bestj+bestsize] {
|
||||||
|
bestsize++
|
||||||
|
}
|
||||||
|
|
||||||
|
return Match{A: besti, B: bestj, Size: bestsize}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return list of triples describing matching subsequences.
|
||||||
|
//
|
||||||
|
// Each triple is of the form (i, j, n), and means that
|
||||||
|
// a[i:i+n] == b[j:j+n]. The triples are monotonically increasing in
|
||||||
|
// i and in j. It's also guaranteed that if (i, j, n) and (i', j', n') are
|
||||||
|
// adjacent triples in the list, and the second is not the last triple in the
|
||||||
|
// list, then i+n != i' or j+n != j'. IOW, adjacent triples never describe
|
||||||
|
// adjacent equal blocks.
|
||||||
|
//
|
||||||
|
// The last triple is a dummy, (len(a), len(b), 0), and is the only
|
||||||
|
// triple with n==0.
|
||||||
|
func (m *SequenceMatcher) GetMatchingBlocks() []Match {
|
||||||
|
if m.matchingBlocks != nil {
|
||||||
|
return m.matchingBlocks
|
||||||
|
}
|
||||||
|
|
||||||
|
var matchBlocks func(alo, ahi, blo, bhi int, matched []Match) []Match
|
||||||
|
matchBlocks = func(alo, ahi, blo, bhi int, matched []Match) []Match {
|
||||||
|
match := m.findLongestMatch(alo, ahi, blo, bhi)
|
||||||
|
i, j, k := match.A, match.B, match.Size
|
||||||
|
if match.Size > 0 {
|
||||||
|
if alo < i && blo < j {
|
||||||
|
matched = matchBlocks(alo, i, blo, j, matched)
|
||||||
|
}
|
||||||
|
matched = append(matched, match)
|
||||||
|
if i+k < ahi && j+k < bhi {
|
||||||
|
matched = matchBlocks(i+k, ahi, j+k, bhi, matched)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return matched
|
||||||
|
}
|
||||||
|
matched := matchBlocks(0, len(m.a), 0, len(m.b), nil)
|
||||||
|
|
||||||
|
// It's possible that we have adjacent equal blocks in the
|
||||||
|
// matching_blocks list now.
|
||||||
|
nonAdjacent := []Match{}
|
||||||
|
i1, j1, k1 := 0, 0, 0
|
||||||
|
for _, b := range matched {
|
||||||
|
// Is this block adjacent to i1, j1, k1?
|
||||||
|
i2, j2, k2 := b.A, b.B, b.Size
|
||||||
|
if i1+k1 == i2 && j1+k1 == j2 {
|
||||||
|
// Yes, so collapse them -- this just increases the length of
|
||||||
|
// the first block by the length of the second, and the first
|
||||||
|
// block so lengthened remains the block to compare against.
|
||||||
|
k1 += k2
|
||||||
|
} else {
|
||||||
|
// Not adjacent. Remember the first block (k1==0 means it's
|
||||||
|
// the dummy we started with), and make the second block the
|
||||||
|
// new block to compare against.
|
||||||
|
if k1 > 0 {
|
||||||
|
nonAdjacent = append(nonAdjacent, Match{i1, j1, k1})
|
||||||
|
}
|
||||||
|
i1, j1, k1 = i2, j2, k2
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if k1 > 0 {
|
||||||
|
nonAdjacent = append(nonAdjacent, Match{i1, j1, k1})
|
||||||
|
}
|
||||||
|
|
||||||
|
nonAdjacent = append(nonAdjacent, Match{len(m.a), len(m.b), 0})
|
||||||
|
m.matchingBlocks = nonAdjacent
|
||||||
|
return m.matchingBlocks
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return list of 5-tuples describing how to turn a into b.
|
||||||
|
//
|
||||||
|
// Each tuple is of the form (tag, i1, i2, j1, j2). The first tuple
|
||||||
|
// has i1 == j1 == 0, and remaining tuples have i1 == the i2 from the
|
||||||
|
// tuple preceding it, and likewise for j1 == the previous j2.
|
||||||
|
//
|
||||||
|
// The tags are characters, with these meanings:
|
||||||
|
//
|
||||||
|
// 'r' (replace): a[i1:i2] should be replaced by b[j1:j2]
|
||||||
|
//
|
||||||
|
// 'd' (delete): a[i1:i2] should be deleted, j1==j2 in this case.
|
||||||
|
//
|
||||||
|
// 'i' (insert): b[j1:j2] should be inserted at a[i1:i1], i1==i2 in this case.
|
||||||
|
//
|
||||||
|
// 'e' (equal): a[i1:i2] == b[j1:j2]
|
||||||
|
func (m *SequenceMatcher) GetOpCodes() []OpCode {
|
||||||
|
if m.opCodes != nil {
|
||||||
|
return m.opCodes
|
||||||
|
}
|
||||||
|
i, j := 0, 0
|
||||||
|
matching := m.GetMatchingBlocks()
|
||||||
|
opCodes := make([]OpCode, 0, len(matching))
|
||||||
|
for _, m := range matching {
|
||||||
|
// invariant: we've pumped out correct diffs to change
|
||||||
|
// a[:i] into b[:j], and the next matching block is
|
||||||
|
// a[ai:ai+size] == b[bj:bj+size]. So we need to pump
|
||||||
|
// out a diff to change a[i:ai] into b[j:bj], pump out
|
||||||
|
// the matching block, and move (i,j) beyond the match
|
||||||
|
ai, bj, size := m.A, m.B, m.Size
|
||||||
|
tag := byte(0)
|
||||||
|
if i < ai && j < bj {
|
||||||
|
tag = 'r'
|
||||||
|
} else if i < ai {
|
||||||
|
tag = 'd'
|
||||||
|
} else if j < bj {
|
||||||
|
tag = 'i'
|
||||||
|
}
|
||||||
|
if tag > 0 {
|
||||||
|
opCodes = append(opCodes, OpCode{tag, i, ai, j, bj})
|
||||||
|
}
|
||||||
|
i, j = ai+size, bj+size
|
||||||
|
// the list of matching blocks is terminated by a
|
||||||
|
// sentinel with size 0
|
||||||
|
if size > 0 {
|
||||||
|
opCodes = append(opCodes, OpCode{'e', ai, i, bj, j})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
m.opCodes = opCodes
|
||||||
|
return m.opCodes
|
||||||
|
}
|
||||||
|
|
||||||
|
// Isolate change clusters by eliminating ranges with no changes.
|
||||||
|
//
|
||||||
|
// Return a generator of groups with up to n lines of context.
|
||||||
|
// Each group is in the same format as returned by GetOpCodes().
|
||||||
|
func (m *SequenceMatcher) GetGroupedOpCodes(n int) [][]OpCode {
|
||||||
|
if n < 0 {
|
||||||
|
n = 3
|
||||||
|
}
|
||||||
|
codes := m.GetOpCodes()
|
||||||
|
if len(codes) == 0 {
|
||||||
|
codes = []OpCode{{'e', 0, 1, 0, 1}}
|
||||||
|
}
|
||||||
|
// Fixup leading and trailing groups if they show no changes.
|
||||||
|
if codes[0].Tag == 'e' {
|
||||||
|
c := codes[0]
|
||||||
|
i1, i2, j1, j2 := c.I1, c.I2, c.J1, c.J2
|
||||||
|
codes[0] = OpCode{c.Tag, max(i1, i2-n), i2, max(j1, j2-n), j2}
|
||||||
|
}
|
||||||
|
if codes[len(codes)-1].Tag == 'e' {
|
||||||
|
c := codes[len(codes)-1]
|
||||||
|
i1, i2, j1, j2 := c.I1, c.I2, c.J1, c.J2
|
||||||
|
codes[len(codes)-1] = OpCode{c.Tag, i1, min(i2, i1+n), j1, min(j2, j1+n)}
|
||||||
|
}
|
||||||
|
nn := n + n
|
||||||
|
groups := [][]OpCode{}
|
||||||
|
group := []OpCode{}
|
||||||
|
for _, c := range codes {
|
||||||
|
i1, i2, j1, j2 := c.I1, c.I2, c.J1, c.J2
|
||||||
|
// End the current group and start a new one whenever
|
||||||
|
// there is a large range with no changes.
|
||||||
|
if c.Tag == 'e' && i2-i1 > nn {
|
||||||
|
group = append(group, OpCode{
|
||||||
|
c.Tag, i1, min(i2, i1+n),
|
||||||
|
j1, min(j2, j1+n),
|
||||||
|
})
|
||||||
|
groups = append(groups, group)
|
||||||
|
group = []OpCode{}
|
||||||
|
i1, j1 = max(i1, i2-n), max(j1, j2-n)
|
||||||
|
}
|
||||||
|
group = append(group, OpCode{c.Tag, i1, i2, j1, j2})
|
||||||
|
}
|
||||||
|
if len(group) > 0 && !(len(group) == 1 && group[0].Tag == 'e') {
|
||||||
|
groups = append(groups, group)
|
||||||
|
}
|
||||||
|
return groups
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return a measure of the sequences' similarity (float in [0,1]).
|
||||||
|
//
|
||||||
|
// Where T is the total number of elements in both sequences, and
|
||||||
|
// M is the number of matches, this is 2.0*M / T.
|
||||||
|
// Note that this is 1 if the sequences are identical, and 0 if
|
||||||
|
// they have nothing in common.
|
||||||
|
//
|
||||||
|
// .Ratio() is expensive to compute if you haven't already computed
|
||||||
|
// .GetMatchingBlocks() or .GetOpCodes(), in which case you may
|
||||||
|
// want to try .QuickRatio() or .RealQuickRation() first to get an
|
||||||
|
// upper bound.
|
||||||
|
func (m *SequenceMatcher) Ratio() float64 {
|
||||||
|
matches := 0
|
||||||
|
for _, m := range m.GetMatchingBlocks() {
|
||||||
|
matches += m.Size
|
||||||
|
}
|
||||||
|
return calculateRatio(matches, len(m.a)+len(m.b))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return an upper bound on ratio() relatively quickly.
|
||||||
|
//
|
||||||
|
// This isn't defined beyond that it is an upper bound on .Ratio(), and
|
||||||
|
// is faster to compute.
|
||||||
|
func (m *SequenceMatcher) QuickRatio() float64 {
|
||||||
|
// viewing a and b as multisets, set matches to the cardinality
|
||||||
|
// of their intersection; this counts the number of matches
|
||||||
|
// without regard to order, so is clearly an upper bound
|
||||||
|
if m.fullBCount == nil {
|
||||||
|
m.fullBCount = map[string]int{}
|
||||||
|
for _, s := range m.b {
|
||||||
|
m.fullBCount[s]++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// avail[x] is the number of times x appears in 'b' less the
|
||||||
|
// number of times we've seen it in 'a' so far ... kinda
|
||||||
|
avail := map[string]int{}
|
||||||
|
matches := 0
|
||||||
|
for _, s := range m.a {
|
||||||
|
n, ok := avail[s]
|
||||||
|
if !ok {
|
||||||
|
n = m.fullBCount[s]
|
||||||
|
}
|
||||||
|
avail[s] = n - 1
|
||||||
|
if n > 0 {
|
||||||
|
matches++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return calculateRatio(matches, len(m.a)+len(m.b))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return an upper bound on ratio() very quickly.
|
||||||
|
//
|
||||||
|
// This isn't defined beyond that it is an upper bound on .Ratio(), and
|
||||||
|
// is faster to compute than either .Ratio() or .QuickRatio().
|
||||||
|
func (m *SequenceMatcher) RealQuickRatio() float64 {
|
||||||
|
la, lb := len(m.a), len(m.b)
|
||||||
|
return calculateRatio(min(la, lb), la+lb)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Convert range to the "ed" format
|
||||||
|
func formatRangeUnified(start, stop int) string {
|
||||||
|
// Per the diff spec at http://www.unix.org/single_unix_specification/
|
||||||
|
beginning := start + 1 // lines start numbering with one
|
||||||
|
length := stop - start
|
||||||
|
if length == 1 {
|
||||||
|
return fmt.Sprintf("%d", beginning)
|
||||||
|
}
|
||||||
|
if length == 0 {
|
||||||
|
beginning-- // empty ranges begin at line just before the range
|
||||||
|
}
|
||||||
|
return fmt.Sprintf("%d,%d", beginning, length)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Unified diff parameters
|
||||||
|
type UnifiedDiff struct {
|
||||||
|
A []string // First sequence lines
|
||||||
|
FromFile string // First file name
|
||||||
|
FromDate string // First file time
|
||||||
|
B []string // Second sequence lines
|
||||||
|
ToFile string // Second file name
|
||||||
|
ToDate string // Second file time
|
||||||
|
Eol string // Headers end of line, defaults to LF
|
||||||
|
Context int // Number of context lines
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compare two sequences of lines; generate the delta as a unified diff.
|
||||||
|
//
|
||||||
|
// Unified diffs are a compact way of showing line changes and a few
|
||||||
|
// lines of context. The number of context lines is set by 'n' which
|
||||||
|
// defaults to three.
|
||||||
|
//
|
||||||
|
// By default, the diff control lines (those with ---, +++, or @@) are
|
||||||
|
// created with a trailing newline. This is helpful so that inputs
|
||||||
|
// created from file.readlines() result in diffs that are suitable for
|
||||||
|
// file.writelines() since both the inputs and outputs have trailing
|
||||||
|
// newlines.
|
||||||
|
//
|
||||||
|
// For inputs that do not have trailing newlines, set the lineterm
|
||||||
|
// argument to "" so that the output will be uniformly newline free.
|
||||||
|
//
|
||||||
|
// The unidiff format normally has a header for filenames and modification
|
||||||
|
// times. Any or all of these may be specified using strings for
|
||||||
|
// 'fromfile', 'tofile', 'fromfiledate', and 'tofiledate'.
|
||||||
|
// The modification times are normally expressed in the ISO 8601 format.
|
||||||
|
func WriteUnifiedDiff(writer io.Writer, diff UnifiedDiff) error {
|
||||||
|
buf := bufio.NewWriter(writer)
|
||||||
|
defer buf.Flush()
|
||||||
|
wf := func(format string, args ...interface{}) error {
|
||||||
|
_, err := buf.WriteString(fmt.Sprintf(format, args...))
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
ws := func(s string) error {
|
||||||
|
_, err := buf.WriteString(s)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(diff.Eol) == 0 {
|
||||||
|
diff.Eol = "\n"
|
||||||
|
}
|
||||||
|
|
||||||
|
started := false
|
||||||
|
m := NewMatcher(diff.A, diff.B)
|
||||||
|
for _, g := range m.GetGroupedOpCodes(diff.Context) {
|
||||||
|
if !started {
|
||||||
|
started = true
|
||||||
|
fromDate := ""
|
||||||
|
if len(diff.FromDate) > 0 {
|
||||||
|
fromDate = "\t" + diff.FromDate
|
||||||
|
}
|
||||||
|
toDate := ""
|
||||||
|
if len(diff.ToDate) > 0 {
|
||||||
|
toDate = "\t" + diff.ToDate
|
||||||
|
}
|
||||||
|
if diff.FromFile != "" || diff.ToFile != "" {
|
||||||
|
err := wf("--- %s%s%s", diff.FromFile, fromDate, diff.Eol)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
err = wf("+++ %s%s%s", diff.ToFile, toDate, diff.Eol)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
first, last := g[0], g[len(g)-1]
|
||||||
|
range1 := formatRangeUnified(first.I1, last.I2)
|
||||||
|
range2 := formatRangeUnified(first.J1, last.J2)
|
||||||
|
if err := wf("@@ -%s +%s @@%s", range1, range2, diff.Eol); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
for _, c := range g {
|
||||||
|
i1, i2, j1, j2 := c.I1, c.I2, c.J1, c.J2
|
||||||
|
if c.Tag == 'e' {
|
||||||
|
for _, line := range diff.A[i1:i2] {
|
||||||
|
if err := ws(" " + line); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if c.Tag == 'r' || c.Tag == 'd' {
|
||||||
|
for _, line := range diff.A[i1:i2] {
|
||||||
|
if err := ws("-" + line); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if c.Tag == 'r' || c.Tag == 'i' {
|
||||||
|
for _, line := range diff.B[j1:j2] {
|
||||||
|
if err := ws("+" + line); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Like WriteUnifiedDiff but returns the diff a string.
|
||||||
|
func GetUnifiedDiffString(diff UnifiedDiff) (string, error) {
|
||||||
|
w := &bytes.Buffer{}
|
||||||
|
err := WriteUnifiedDiff(w, diff)
|
||||||
|
return w.String(), err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Split a string on "\n" while preserving them. The output can be used
|
||||||
|
// as input for UnifiedDiff and ContextDiff structures.
|
||||||
|
func SplitLines(s string) []string {
|
||||||
|
lines := strings.SplitAfter(s, "\n")
|
||||||
|
lines[len(lines)-1] += "\n"
|
||||||
|
return lines
|
||||||
|
}
|
32
vendor/github.com/prometheus/client_golang/prometheus/internal/go_collector_options.go
generated
vendored
Normal file
32
vendor/github.com/prometheus/client_golang/prometheus/internal/go_collector_options.go
generated
vendored
Normal file
@ -0,0 +1,32 @@
|
|||||||
|
// Copyright 2021 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import "regexp"
|
||||||
|
|
||||||
|
type GoCollectorRule struct {
|
||||||
|
Matcher *regexp.Regexp
|
||||||
|
Deny bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// GoCollectorOptions should not be used be directly by anything, except `collectors` package.
|
||||||
|
// Use it via collectors package instead. See issue
|
||||||
|
// https://github.com/prometheus/client_golang/issues/1030.
|
||||||
|
//
|
||||||
|
// This is internal, so external users only can use it via `collector.WithGoCollector*` methods
|
||||||
|
type GoCollectorOptions struct {
|
||||||
|
DisableMemStatsLikeMetrics bool
|
||||||
|
RuntimeMetricSumForHist map[string]string
|
||||||
|
RuntimeMetricRules []GoCollectorRule
|
||||||
|
}
|
4
vendor/github.com/prometheus/client_golang/prometheus/internal/go_runtime_metrics.go
generated
vendored
4
vendor/github.com/prometheus/client_golang/prometheus/internal/go_runtime_metrics.go
generated
vendored
@ -61,9 +61,9 @@ func RuntimeMetricsToProm(d *metrics.Description) (string, string, string, bool)
|
|||||||
// name has - replaced with _ and is concatenated with the unit and
|
// name has - replaced with _ and is concatenated with the unit and
|
||||||
// other data.
|
// other data.
|
||||||
name = strings.ReplaceAll(name, "-", "_")
|
name = strings.ReplaceAll(name, "-", "_")
|
||||||
name = name + "_" + unit
|
name += "_" + unit
|
||||||
if d.Cumulative && d.Kind != metrics.KindFloat64Histogram {
|
if d.Cumulative && d.Kind != metrics.KindFloat64Histogram {
|
||||||
name = name + "_total"
|
name += "_total"
|
||||||
}
|
}
|
||||||
|
|
||||||
valid := model.IsValidMetricName(model.LabelValue(namespace + "_" + subsystem + "_" + name))
|
valid := model.IsValidMetricName(model.LabelValue(namespace + "_" + subsystem + "_" + name))
|
||||||
|
28
vendor/github.com/prometheus/client_golang/prometheus/internal/metric.go
generated
vendored
28
vendor/github.com/prometheus/client_golang/prometheus/internal/metric.go
generated
vendored
@ -19,18 +19,34 @@ import (
|
|||||||
dto "github.com/prometheus/client_model/go"
|
dto "github.com/prometheus/client_model/go"
|
||||||
)
|
)
|
||||||
|
|
||||||
// metricSorter is a sortable slice of *dto.Metric.
|
// LabelPairSorter implements sort.Interface. It is used to sort a slice of
|
||||||
type metricSorter []*dto.Metric
|
// dto.LabelPair pointers.
|
||||||
|
type LabelPairSorter []*dto.LabelPair
|
||||||
|
|
||||||
func (s metricSorter) Len() int {
|
func (s LabelPairSorter) Len() int {
|
||||||
return len(s)
|
return len(s)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s metricSorter) Swap(i, j int) {
|
func (s LabelPairSorter) Swap(i, j int) {
|
||||||
s[i], s[j] = s[j], s[i]
|
s[i], s[j] = s[j], s[i]
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s metricSorter) Less(i, j int) bool {
|
func (s LabelPairSorter) Less(i, j int) bool {
|
||||||
|
return s[i].GetName() < s[j].GetName()
|
||||||
|
}
|
||||||
|
|
||||||
|
// MetricSorter is a sortable slice of *dto.Metric.
|
||||||
|
type MetricSorter []*dto.Metric
|
||||||
|
|
||||||
|
func (s MetricSorter) Len() int {
|
||||||
|
return len(s)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s MetricSorter) Swap(i, j int) {
|
||||||
|
s[i], s[j] = s[j], s[i]
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s MetricSorter) Less(i, j int) bool {
|
||||||
if len(s[i].Label) != len(s[j].Label) {
|
if len(s[i].Label) != len(s[j].Label) {
|
||||||
// This should not happen. The metrics are
|
// This should not happen. The metrics are
|
||||||
// inconsistent. However, we have to deal with the fact, as
|
// inconsistent. However, we have to deal with the fact, as
|
||||||
@ -68,7 +84,7 @@ func (s metricSorter) Less(i, j int) bool {
|
|||||||
// the slice, with the contained Metrics sorted within each MetricFamily.
|
// the slice, with the contained Metrics sorted within each MetricFamily.
|
||||||
func NormalizeMetricFamilies(metricFamiliesByName map[string]*dto.MetricFamily) []*dto.MetricFamily {
|
func NormalizeMetricFamilies(metricFamiliesByName map[string]*dto.MetricFamily) []*dto.MetricFamily {
|
||||||
for _, mf := range metricFamiliesByName {
|
for _, mf := range metricFamiliesByName {
|
||||||
sort.Sort(metricSorter(mf.Metric))
|
sort.Sort(MetricSorter(mf.Metric))
|
||||||
}
|
}
|
||||||
names := make([]string, 0, len(metricFamiliesByName))
|
names := make([]string, 0, len(metricFamiliesByName))
|
||||||
for name, mf := range metricFamiliesByName {
|
for name, mf := range metricFamiliesByName {
|
||||||
|
9
vendor/github.com/prometheus/client_golang/prometheus/labels.go
generated
vendored
9
vendor/github.com/prometheus/client_golang/prometheus/labels.go
generated
vendored
@ -25,7 +25,8 @@ import (
|
|||||||
// Labels represents a collection of label name -> value mappings. This type is
|
// Labels represents a collection of label name -> value mappings. This type is
|
||||||
// commonly used with the With(Labels) and GetMetricWith(Labels) methods of
|
// commonly used with the With(Labels) and GetMetricWith(Labels) methods of
|
||||||
// metric vector Collectors, e.g.:
|
// metric vector Collectors, e.g.:
|
||||||
// myVec.With(Labels{"code": "404", "method": "GET"}).Add(42)
|
//
|
||||||
|
// myVec.With(Labels{"code": "404", "method": "GET"}).Add(42)
|
||||||
//
|
//
|
||||||
// The other use-case is the specification of constant label pairs in Opts or to
|
// The other use-case is the specification of constant label pairs in Opts or to
|
||||||
// create a Desc.
|
// create a Desc.
|
||||||
@ -39,7 +40,7 @@ var errInconsistentCardinality = errors.New("inconsistent label cardinality")
|
|||||||
|
|
||||||
func makeInconsistentCardinalityError(fqName string, labels, labelValues []string) error {
|
func makeInconsistentCardinalityError(fqName string, labels, labelValues []string) error {
|
||||||
return fmt.Errorf(
|
return fmt.Errorf(
|
||||||
"%s: %q has %d variable labels named %q but %d values %q were provided",
|
"%w: %q has %d variable labels named %q but %d values %q were provided",
|
||||||
errInconsistentCardinality, fqName,
|
errInconsistentCardinality, fqName,
|
||||||
len(labels), labels,
|
len(labels), labels,
|
||||||
len(labelValues), labelValues,
|
len(labelValues), labelValues,
|
||||||
@ -49,7 +50,7 @@ func makeInconsistentCardinalityError(fqName string, labels, labelValues []strin
|
|||||||
func validateValuesInLabels(labels Labels, expectedNumberOfValues int) error {
|
func validateValuesInLabels(labels Labels, expectedNumberOfValues int) error {
|
||||||
if len(labels) != expectedNumberOfValues {
|
if len(labels) != expectedNumberOfValues {
|
||||||
return fmt.Errorf(
|
return fmt.Errorf(
|
||||||
"%s: expected %d label values but got %d in %#v",
|
"%w: expected %d label values but got %d in %#v",
|
||||||
errInconsistentCardinality, expectedNumberOfValues,
|
errInconsistentCardinality, expectedNumberOfValues,
|
||||||
len(labels), labels,
|
len(labels), labels,
|
||||||
)
|
)
|
||||||
@ -67,7 +68,7 @@ func validateValuesInLabels(labels Labels, expectedNumberOfValues int) error {
|
|||||||
func validateLabelValues(vals []string, expectedNumberOfValues int) error {
|
func validateLabelValues(vals []string, expectedNumberOfValues int) error {
|
||||||
if len(vals) != expectedNumberOfValues {
|
if len(vals) != expectedNumberOfValues {
|
||||||
return fmt.Errorf(
|
return fmt.Errorf(
|
||||||
"%s: expected %d label values but got %d in %#v",
|
"%w: expected %d label values but got %d in %#v",
|
||||||
errInconsistentCardinality, expectedNumberOfValues,
|
errInconsistentCardinality, expectedNumberOfValues,
|
||||||
len(vals), vals,
|
len(vals), vals,
|
||||||
)
|
)
|
||||||
|
112
vendor/github.com/prometheus/client_golang/prometheus/metric.go
generated
vendored
112
vendor/github.com/prometheus/client_golang/prometheus/metric.go
generated
vendored
@ -14,6 +14,9 @@
|
|||||||
package prometheus
|
package prometheus
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"errors"
|
||||||
|
"math"
|
||||||
|
"sort"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
@ -115,22 +118,6 @@ func BuildFQName(namespace, subsystem, name string) string {
|
|||||||
return name
|
return name
|
||||||
}
|
}
|
||||||
|
|
||||||
// labelPairSorter implements sort.Interface. It is used to sort a slice of
|
|
||||||
// dto.LabelPair pointers.
|
|
||||||
type labelPairSorter []*dto.LabelPair
|
|
||||||
|
|
||||||
func (s labelPairSorter) Len() int {
|
|
||||||
return len(s)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s labelPairSorter) Swap(i, j int) {
|
|
||||||
s[i], s[j] = s[j], s[i]
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s labelPairSorter) Less(i, j int) bool {
|
|
||||||
return s[i].GetName() < s[j].GetName()
|
|
||||||
}
|
|
||||||
|
|
||||||
type invalidMetric struct {
|
type invalidMetric struct {
|
||||||
desc *Desc
|
desc *Desc
|
||||||
err error
|
err error
|
||||||
@ -174,3 +161,96 @@ func (m timestampedMetric) Write(pb *dto.Metric) error {
|
|||||||
func NewMetricWithTimestamp(t time.Time, m Metric) Metric {
|
func NewMetricWithTimestamp(t time.Time, m Metric) Metric {
|
||||||
return timestampedMetric{Metric: m, t: t}
|
return timestampedMetric{Metric: m, t: t}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type withExemplarsMetric struct {
|
||||||
|
Metric
|
||||||
|
|
||||||
|
exemplars []*dto.Exemplar
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *withExemplarsMetric) Write(pb *dto.Metric) error {
|
||||||
|
if err := m.Metric.Write(pb); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case pb.Counter != nil:
|
||||||
|
pb.Counter.Exemplar = m.exemplars[len(m.exemplars)-1]
|
||||||
|
case pb.Histogram != nil:
|
||||||
|
for _, e := range m.exemplars {
|
||||||
|
// pb.Histogram.Bucket are sorted by UpperBound.
|
||||||
|
i := sort.Search(len(pb.Histogram.Bucket), func(i int) bool {
|
||||||
|
return pb.Histogram.Bucket[i].GetUpperBound() >= e.GetValue()
|
||||||
|
})
|
||||||
|
if i < len(pb.Histogram.Bucket) {
|
||||||
|
pb.Histogram.Bucket[i].Exemplar = e
|
||||||
|
} else {
|
||||||
|
// The +Inf bucket should be explicitly added if there is an exemplar for it, similar to non-const histogram logic in https://github.com/prometheus/client_golang/blob/main/prometheus/histogram.go#L357-L365.
|
||||||
|
b := &dto.Bucket{
|
||||||
|
CumulativeCount: proto.Uint64(pb.Histogram.GetSampleCount()),
|
||||||
|
UpperBound: proto.Float64(math.Inf(1)),
|
||||||
|
Exemplar: e,
|
||||||
|
}
|
||||||
|
pb.Histogram.Bucket = append(pb.Histogram.Bucket, b)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
// TODO(bwplotka): Implement Gauge?
|
||||||
|
return errors.New("cannot inject exemplar into Gauge, Summary or Untyped")
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Exemplar is easier to use, user-facing representation of *dto.Exemplar.
|
||||||
|
type Exemplar struct {
|
||||||
|
Value float64
|
||||||
|
Labels Labels
|
||||||
|
// Optional.
|
||||||
|
// Default value (time.Time{}) indicates its empty, which should be
|
||||||
|
// understood as time.Now() time at the moment of creation of metric.
|
||||||
|
Timestamp time.Time
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewMetricWithExemplars returns a new Metric wrapping the provided Metric with given
|
||||||
|
// exemplars. Exemplars are validated.
|
||||||
|
//
|
||||||
|
// Only last applicable exemplar is injected from the list.
|
||||||
|
// For example for Counter it means last exemplar is injected.
|
||||||
|
// For Histogram, it means last applicable exemplar for each bucket is injected.
|
||||||
|
//
|
||||||
|
// NewMetricWithExemplars works best with MustNewConstMetric and
|
||||||
|
// MustNewConstHistogram, see example.
|
||||||
|
func NewMetricWithExemplars(m Metric, exemplars ...Exemplar) (Metric, error) {
|
||||||
|
if len(exemplars) == 0 {
|
||||||
|
return nil, errors.New("no exemplar was passed for NewMetricWithExemplars")
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
now = time.Now()
|
||||||
|
exs = make([]*dto.Exemplar, len(exemplars))
|
||||||
|
err error
|
||||||
|
)
|
||||||
|
for i, e := range exemplars {
|
||||||
|
ts := e.Timestamp
|
||||||
|
if ts == (time.Time{}) {
|
||||||
|
ts = now
|
||||||
|
}
|
||||||
|
exs[i], err = newExemplar(e.Value, ts, e.Labels)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &withExemplarsMetric{Metric: m, exemplars: exs}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// MustNewMetricWithExemplars is a version of NewMetricWithExemplars that panics where
|
||||||
|
// NewMetricWithExemplars would have returned an error.
|
||||||
|
func MustNewMetricWithExemplars(m Metric, exemplars ...Exemplar) Metric {
|
||||||
|
ret, err := NewMetricWithExemplars(m, exemplars...)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
return ret
|
||||||
|
}
|
||||||
|
25
vendor/github.com/prometheus/client_golang/prometheus/num_threads.go
generated
vendored
Normal file
25
vendor/github.com/prometheus/client_golang/prometheus/num_threads.go
generated
vendored
Normal file
@ -0,0 +1,25 @@
|
|||||||
|
// Copyright 2018 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build !js || wasm
|
||||||
|
// +build !js wasm
|
||||||
|
|
||||||
|
package prometheus
|
||||||
|
|
||||||
|
import "runtime"
|
||||||
|
|
||||||
|
// getRuntimeNumThreads returns the number of open OS threads.
|
||||||
|
func getRuntimeNumThreads() float64 {
|
||||||
|
n, _ := runtime.ThreadCreateProfile(nil)
|
||||||
|
return float64(n)
|
||||||
|
}
|
22
vendor/github.com/prometheus/client_golang/prometheus/num_threads_gopherjs.go
generated
vendored
Normal file
22
vendor/github.com/prometheus/client_golang/prometheus/num_threads_gopherjs.go
generated
vendored
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
// Copyright 2018 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build js && !wasm
|
||||||
|
// +build js,!wasm
|
||||||
|
|
||||||
|
package prometheus
|
||||||
|
|
||||||
|
// getRuntimeNumThreads returns the number of open OS threads.
|
||||||
|
func getRuntimeNumThreads() float64 {
|
||||||
|
return 1
|
||||||
|
}
|
2
vendor/github.com/prometheus/client_golang/prometheus/observer.go
generated
vendored
2
vendor/github.com/prometheus/client_golang/prometheus/observer.go
generated
vendored
@ -58,7 +58,7 @@ type ObserverVec interface {
|
|||||||
// current time as timestamp, and the provided Labels. Empty Labels will lead to
|
// current time as timestamp, and the provided Labels. Empty Labels will lead to
|
||||||
// a valid (label-less) exemplar. But if Labels is nil, the current exemplar is
|
// a valid (label-less) exemplar. But if Labels is nil, the current exemplar is
|
||||||
// left in place. ObserveWithExemplar panics if any of the provided labels are
|
// left in place. ObserveWithExemplar panics if any of the provided labels are
|
||||||
// invalid or if the provided labels contain more than 64 runes in total.
|
// invalid or if the provided labels contain more than 128 runes in total.
|
||||||
type ExemplarObserver interface {
|
type ExemplarObserver interface {
|
||||||
ObserveWithExemplar(value float64, exemplar Labels)
|
ObserveWithExemplar(value float64, exemplar Labels)
|
||||||
}
|
}
|
||||||
|
10
vendor/github.com/prometheus/client_golang/prometheus/process_collector.go
generated
vendored
10
vendor/github.com/prometheus/client_golang/prometheus/process_collector.go
generated
vendored
@ -16,7 +16,6 @@ package prometheus
|
|||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
|
||||||
"os"
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
@ -104,8 +103,7 @@ func NewProcessCollector(opts ProcessCollectorOpts) Collector {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if opts.PidFn == nil {
|
if opts.PidFn == nil {
|
||||||
pid := os.Getpid()
|
c.pidFn = getPIDFn()
|
||||||
c.pidFn = func() (int, error) { return pid, nil }
|
|
||||||
} else {
|
} else {
|
||||||
c.pidFn = opts.PidFn
|
c.pidFn = opts.PidFn
|
||||||
}
|
}
|
||||||
@ -152,13 +150,13 @@ func (c *processCollector) reportError(ch chan<- Metric, desc *Desc, err error)
|
|||||||
// It is meant to be used for the PidFn field in ProcessCollectorOpts.
|
// It is meant to be used for the PidFn field in ProcessCollectorOpts.
|
||||||
func NewPidFileFn(pidFilePath string) func() (int, error) {
|
func NewPidFileFn(pidFilePath string) func() (int, error) {
|
||||||
return func() (int, error) {
|
return func() (int, error) {
|
||||||
content, err := ioutil.ReadFile(pidFilePath)
|
content, err := os.ReadFile(pidFilePath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, fmt.Errorf("can't read pid file %q: %+v", pidFilePath, err)
|
return 0, fmt.Errorf("can't read pid file %q: %w", pidFilePath, err)
|
||||||
}
|
}
|
||||||
pid, err := strconv.Atoi(strings.TrimSpace(string(content)))
|
pid, err := strconv.Atoi(strings.TrimSpace(string(content)))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, fmt.Errorf("can't parse pid file %q: %+v", pidFilePath, err)
|
return 0, fmt.Errorf("can't parse pid file %q: %w", pidFilePath, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
return pid, nil
|
return pid, nil
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
// Copyright 2021 The Prometheus Authors
|
// Copyright 2019 The Prometheus Authors
|
||||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
// you may not use this file except in compliance with the License.
|
// you may not use this file except in compliance with the License.
|
||||||
// You may obtain a copy of the License at
|
// You may obtain a copy of the License at
|
||||||
@ -11,21 +11,16 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
//go:build go1.15
|
//go:build js
|
||||||
// +build go1.15
|
// +build js
|
||||||
|
|
||||||
package collectors
|
package prometheus
|
||||||
|
|
||||||
import (
|
func canCollectProcess() bool {
|
||||||
"database/sql"
|
return false
|
||||||
|
|
||||||
"github.com/prometheus/client_golang/prometheus"
|
|
||||||
)
|
|
||||||
|
|
||||||
func (c *dbStatsCollector) describeNewInGo115(ch chan<- *prometheus.Desc) {
|
|
||||||
ch <- c.maxIdleTimeClosed
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *dbStatsCollector) collectNewInGo115(ch chan<- prometheus.Metric, stats sql.DBStats) {
|
func (c *processCollector) processCollect(ch chan<- Metric) {
|
||||||
ch <- prometheus.MustNewConstMetric(c.maxIdleTimeClosed, prometheus.CounterValue, float64(stats.MaxIdleTimeClosed))
|
// noop on this platform
|
||||||
|
return
|
||||||
}
|
}
|
4
vendor/github.com/prometheus/client_golang/prometheus/process_collector_other.go
generated
vendored
4
vendor/github.com/prometheus/client_golang/prometheus/process_collector_other.go
generated
vendored
@ -11,8 +11,8 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
//go:build !windows
|
//go:build !windows && !js
|
||||||
// +build !windows
|
// +build !windows,!js
|
||||||
|
|
||||||
package prometheus
|
package prometheus
|
||||||
|
|
||||||
|
18
vendor/github.com/prometheus/client_golang/prometheus/promhttp/delegator.go
generated
vendored
18
vendor/github.com/prometheus/client_golang/prometheus/promhttp/delegator.go
generated
vendored
@ -76,16 +76,19 @@ func (r *responseWriterDelegator) Write(b []byte) (int, error) {
|
|||||||
return n, err
|
return n, err
|
||||||
}
|
}
|
||||||
|
|
||||||
type closeNotifierDelegator struct{ *responseWriterDelegator }
|
type (
|
||||||
type flusherDelegator struct{ *responseWriterDelegator }
|
closeNotifierDelegator struct{ *responseWriterDelegator }
|
||||||
type hijackerDelegator struct{ *responseWriterDelegator }
|
flusherDelegator struct{ *responseWriterDelegator }
|
||||||
type readerFromDelegator struct{ *responseWriterDelegator }
|
hijackerDelegator struct{ *responseWriterDelegator }
|
||||||
type pusherDelegator struct{ *responseWriterDelegator }
|
readerFromDelegator struct{ *responseWriterDelegator }
|
||||||
|
pusherDelegator struct{ *responseWriterDelegator }
|
||||||
|
)
|
||||||
|
|
||||||
func (d closeNotifierDelegator) CloseNotify() <-chan bool {
|
func (d closeNotifierDelegator) CloseNotify() <-chan bool {
|
||||||
//nolint:staticcheck // Ignore SA1019. http.CloseNotifier is deprecated but we keep it here to not break existing users.
|
//nolint:staticcheck // Ignore SA1019. http.CloseNotifier is deprecated but we keep it here to not break existing users.
|
||||||
return d.ResponseWriter.(http.CloseNotifier).CloseNotify()
|
return d.ResponseWriter.(http.CloseNotifier).CloseNotify()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d flusherDelegator) Flush() {
|
func (d flusherDelegator) Flush() {
|
||||||
// If applicable, call WriteHeader here so that observeWriteHeader is
|
// If applicable, call WriteHeader here so that observeWriteHeader is
|
||||||
// handled appropriately.
|
// handled appropriately.
|
||||||
@ -94,9 +97,11 @@ func (d flusherDelegator) Flush() {
|
|||||||
}
|
}
|
||||||
d.ResponseWriter.(http.Flusher).Flush()
|
d.ResponseWriter.(http.Flusher).Flush()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d hijackerDelegator) Hijack() (net.Conn, *bufio.ReadWriter, error) {
|
func (d hijackerDelegator) Hijack() (net.Conn, *bufio.ReadWriter, error) {
|
||||||
return d.ResponseWriter.(http.Hijacker).Hijack()
|
return d.ResponseWriter.(http.Hijacker).Hijack()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d readerFromDelegator) ReadFrom(re io.Reader) (int64, error) {
|
func (d readerFromDelegator) ReadFrom(re io.Reader) (int64, error) {
|
||||||
// If applicable, call WriteHeader here so that observeWriteHeader is
|
// If applicable, call WriteHeader here so that observeWriteHeader is
|
||||||
// handled appropriately.
|
// handled appropriately.
|
||||||
@ -107,6 +112,7 @@ func (d readerFromDelegator) ReadFrom(re io.Reader) (int64, error) {
|
|||||||
d.written += n
|
d.written += n
|
||||||
return n, err
|
return n, err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d pusherDelegator) Push(target string, opts *http.PushOptions) error {
|
func (d pusherDelegator) Push(target string, opts *http.PushOptions) error {
|
||||||
return d.ResponseWriter.(http.Pusher).Push(target, opts)
|
return d.ResponseWriter.(http.Pusher).Push(target, opts)
|
||||||
}
|
}
|
||||||
@ -261,7 +267,7 @@ func init() {
|
|||||||
http.Flusher
|
http.Flusher
|
||||||
}{d, pusherDelegator{d}, hijackerDelegator{d}, flusherDelegator{d}}
|
}{d, pusherDelegator{d}, hijackerDelegator{d}, flusherDelegator{d}}
|
||||||
}
|
}
|
||||||
pickDelegator[pusher+hijacker+flusher+closeNotifier] = func(d *responseWriterDelegator) delegator { //23
|
pickDelegator[pusher+hijacker+flusher+closeNotifier] = func(d *responseWriterDelegator) delegator { // 23
|
||||||
return struct {
|
return struct {
|
||||||
*responseWriterDelegator
|
*responseWriterDelegator
|
||||||
http.Pusher
|
http.Pusher
|
||||||
|
20
vendor/github.com/prometheus/client_golang/prometheus/promhttp/http.go
generated
vendored
20
vendor/github.com/prometheus/client_golang/prometheus/promhttp/http.go
generated
vendored
@ -33,6 +33,7 @@ package promhttp
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"compress/gzip"
|
"compress/gzip"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
@ -84,6 +85,13 @@ func Handler() http.Handler {
|
|||||||
// instrumentation. Use the InstrumentMetricHandler function to apply the same
|
// instrumentation. Use the InstrumentMetricHandler function to apply the same
|
||||||
// kind of instrumentation as it is used by the Handler function.
|
// kind of instrumentation as it is used by the Handler function.
|
||||||
func HandlerFor(reg prometheus.Gatherer, opts HandlerOpts) http.Handler {
|
func HandlerFor(reg prometheus.Gatherer, opts HandlerOpts) http.Handler {
|
||||||
|
return HandlerForTransactional(prometheus.ToTransactionalGatherer(reg), opts)
|
||||||
|
}
|
||||||
|
|
||||||
|
// HandlerForTransactional is like HandlerFor, but it uses transactional gather, which
|
||||||
|
// can safely change in-place returned *dto.MetricFamily before call to `Gather` and after
|
||||||
|
// call to `done` of that `Gather`.
|
||||||
|
func HandlerForTransactional(reg prometheus.TransactionalGatherer, opts HandlerOpts) http.Handler {
|
||||||
var (
|
var (
|
||||||
inFlightSem chan struct{}
|
inFlightSem chan struct{}
|
||||||
errCnt = prometheus.NewCounterVec(
|
errCnt = prometheus.NewCounterVec(
|
||||||
@ -103,7 +111,8 @@ func HandlerFor(reg prometheus.Gatherer, opts HandlerOpts) http.Handler {
|
|||||||
errCnt.WithLabelValues("gathering")
|
errCnt.WithLabelValues("gathering")
|
||||||
errCnt.WithLabelValues("encoding")
|
errCnt.WithLabelValues("encoding")
|
||||||
if err := opts.Registry.Register(errCnt); err != nil {
|
if err := opts.Registry.Register(errCnt); err != nil {
|
||||||
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
|
are := &prometheus.AlreadyRegisteredError{}
|
||||||
|
if errors.As(err, are) {
|
||||||
errCnt = are.ExistingCollector.(*prometheus.CounterVec)
|
errCnt = are.ExistingCollector.(*prometheus.CounterVec)
|
||||||
} else {
|
} else {
|
||||||
panic(err)
|
panic(err)
|
||||||
@ -123,7 +132,8 @@ func HandlerFor(reg prometheus.Gatherer, opts HandlerOpts) http.Handler {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
mfs, err := reg.Gather()
|
mfs, done, err := reg.Gather()
|
||||||
|
defer done()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if opts.ErrorLog != nil {
|
if opts.ErrorLog != nil {
|
||||||
opts.ErrorLog.Println("error gathering metrics:", err)
|
opts.ErrorLog.Println("error gathering metrics:", err)
|
||||||
@ -242,7 +252,8 @@ func InstrumentMetricHandler(reg prometheus.Registerer, handler http.Handler) ht
|
|||||||
cnt.WithLabelValues("500")
|
cnt.WithLabelValues("500")
|
||||||
cnt.WithLabelValues("503")
|
cnt.WithLabelValues("503")
|
||||||
if err := reg.Register(cnt); err != nil {
|
if err := reg.Register(cnt); err != nil {
|
||||||
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
|
are := &prometheus.AlreadyRegisteredError{}
|
||||||
|
if errors.As(err, are) {
|
||||||
cnt = are.ExistingCollector.(*prometheus.CounterVec)
|
cnt = are.ExistingCollector.(*prometheus.CounterVec)
|
||||||
} else {
|
} else {
|
||||||
panic(err)
|
panic(err)
|
||||||
@ -254,7 +265,8 @@ func InstrumentMetricHandler(reg prometheus.Registerer, handler http.Handler) ht
|
|||||||
Help: "Current number of scrapes being served.",
|
Help: "Current number of scrapes being served.",
|
||||||
})
|
})
|
||||||
if err := reg.Register(gge); err != nil {
|
if err := reg.Register(gge); err != nil {
|
||||||
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
|
are := &prometheus.AlreadyRegisteredError{}
|
||||||
|
if errors.As(err, are) {
|
||||||
gge = are.ExistingCollector.(prometheus.Gauge)
|
gge = are.ExistingCollector.(prometheus.Gauge)
|
||||||
} else {
|
} else {
|
||||||
panic(err)
|
panic(err)
|
||||||
|
40
vendor/github.com/prometheus/client_golang/prometheus/promhttp/instrument_client.go
generated
vendored
40
vendor/github.com/prometheus/client_golang/prometheus/promhttp/instrument_client.go
generated
vendored
@ -38,11 +38,11 @@ func (rt RoundTripperFunc) RoundTrip(r *http.Request) (*http.Response, error) {
|
|||||||
//
|
//
|
||||||
// See the example for ExampleInstrumentRoundTripperDuration for example usage.
|
// See the example for ExampleInstrumentRoundTripperDuration for example usage.
|
||||||
func InstrumentRoundTripperInFlight(gauge prometheus.Gauge, next http.RoundTripper) RoundTripperFunc {
|
func InstrumentRoundTripperInFlight(gauge prometheus.Gauge, next http.RoundTripper) RoundTripperFunc {
|
||||||
return RoundTripperFunc(func(r *http.Request) (*http.Response, error) {
|
return func(r *http.Request) (*http.Response, error) {
|
||||||
gauge.Inc()
|
gauge.Inc()
|
||||||
defer gauge.Dec()
|
defer gauge.Dec()
|
||||||
return next.RoundTrip(r)
|
return next.RoundTrip(r)
|
||||||
})
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// InstrumentRoundTripperCounter is a middleware that wraps the provided
|
// InstrumentRoundTripperCounter is a middleware that wraps the provided
|
||||||
@ -59,22 +59,28 @@ func InstrumentRoundTripperInFlight(gauge prometheus.Gauge, next http.RoundTripp
|
|||||||
// If the wrapped RoundTripper panics or returns a non-nil error, the Counter
|
// If the wrapped RoundTripper panics or returns a non-nil error, the Counter
|
||||||
// is not incremented.
|
// is not incremented.
|
||||||
//
|
//
|
||||||
|
// Use with WithExemplarFromContext to instrument the exemplars on the counter of requests.
|
||||||
|
//
|
||||||
// See the example for ExampleInstrumentRoundTripperDuration for example usage.
|
// See the example for ExampleInstrumentRoundTripperDuration for example usage.
|
||||||
func InstrumentRoundTripperCounter(counter *prometheus.CounterVec, next http.RoundTripper, opts ...Option) RoundTripperFunc {
|
func InstrumentRoundTripperCounter(counter *prometheus.CounterVec, next http.RoundTripper, opts ...Option) RoundTripperFunc {
|
||||||
rtOpts := &option{}
|
rtOpts := defaultOptions()
|
||||||
for _, o := range opts {
|
for _, o := range opts {
|
||||||
o(rtOpts)
|
o.apply(rtOpts)
|
||||||
}
|
}
|
||||||
|
|
||||||
code, method := checkLabels(counter)
|
code, method := checkLabels(counter)
|
||||||
|
|
||||||
return RoundTripperFunc(func(r *http.Request) (*http.Response, error) {
|
return func(r *http.Request) (*http.Response, error) {
|
||||||
resp, err := next.RoundTrip(r)
|
resp, err := next.RoundTrip(r)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
counter.With(labels(code, method, r.Method, resp.StatusCode, rtOpts.extraMethods...)).Inc()
|
addWithExemplar(
|
||||||
|
counter.With(labels(code, method, r.Method, resp.StatusCode, rtOpts.extraMethods...)),
|
||||||
|
1,
|
||||||
|
rtOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
}
|
}
|
||||||
return resp, err
|
return resp, err
|
||||||
})
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// InstrumentRoundTripperDuration is a middleware that wraps the provided
|
// InstrumentRoundTripperDuration is a middleware that wraps the provided
|
||||||
@ -94,24 +100,30 @@ func InstrumentRoundTripperCounter(counter *prometheus.CounterVec, next http.Rou
|
|||||||
// If the wrapped RoundTripper panics or returns a non-nil error, no values are
|
// If the wrapped RoundTripper panics or returns a non-nil error, no values are
|
||||||
// reported.
|
// reported.
|
||||||
//
|
//
|
||||||
|
// Use with WithExemplarFromContext to instrument the exemplars on the duration histograms.
|
||||||
|
//
|
||||||
// Note that this method is only guaranteed to never observe negative durations
|
// Note that this method is only guaranteed to never observe negative durations
|
||||||
// if used with Go1.9+.
|
// if used with Go1.9+.
|
||||||
func InstrumentRoundTripperDuration(obs prometheus.ObserverVec, next http.RoundTripper, opts ...Option) RoundTripperFunc {
|
func InstrumentRoundTripperDuration(obs prometheus.ObserverVec, next http.RoundTripper, opts ...Option) RoundTripperFunc {
|
||||||
rtOpts := &option{}
|
rtOpts := defaultOptions()
|
||||||
for _, o := range opts {
|
for _, o := range opts {
|
||||||
o(rtOpts)
|
o.apply(rtOpts)
|
||||||
}
|
}
|
||||||
|
|
||||||
code, method := checkLabels(obs)
|
code, method := checkLabels(obs)
|
||||||
|
|
||||||
return RoundTripperFunc(func(r *http.Request) (*http.Response, error) {
|
return func(r *http.Request) (*http.Response, error) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
resp, err := next.RoundTrip(r)
|
resp, err := next.RoundTrip(r)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
obs.With(labels(code, method, r.Method, resp.StatusCode, rtOpts.extraMethods...)).Observe(time.Since(start).Seconds())
|
observeWithExemplar(
|
||||||
|
obs.With(labels(code, method, r.Method, resp.StatusCode, rtOpts.extraMethods...)),
|
||||||
|
time.Since(start).Seconds(),
|
||||||
|
rtOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
}
|
}
|
||||||
return resp, err
|
return resp, err
|
||||||
})
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// InstrumentTrace is used to offer flexibility in instrumenting the available
|
// InstrumentTrace is used to offer flexibility in instrumenting the available
|
||||||
@ -149,7 +161,7 @@ type InstrumentTrace struct {
|
|||||||
//
|
//
|
||||||
// See the example for ExampleInstrumentRoundTripperDuration for example usage.
|
// See the example for ExampleInstrumentRoundTripperDuration for example usage.
|
||||||
func InstrumentRoundTripperTrace(it *InstrumentTrace, next http.RoundTripper) RoundTripperFunc {
|
func InstrumentRoundTripperTrace(it *InstrumentTrace, next http.RoundTripper) RoundTripperFunc {
|
||||||
return RoundTripperFunc(func(r *http.Request) (*http.Response, error) {
|
return func(r *http.Request) (*http.Response, error) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
|
|
||||||
trace := &httptrace.ClientTrace{
|
trace := &httptrace.ClientTrace{
|
||||||
@ -231,5 +243,5 @@ func InstrumentRoundTripperTrace(it *InstrumentTrace, next http.RoundTripper) Ro
|
|||||||
r = r.WithContext(httptrace.WithClientTrace(r.Context(), trace))
|
r = r.WithContext(httptrace.WithClientTrace(r.Context(), trace))
|
||||||
|
|
||||||
return next.RoundTrip(r)
|
return next.RoundTrip(r)
|
||||||
})
|
}
|
||||||
}
|
}
|
||||||
|
123
vendor/github.com/prometheus/client_golang/prometheus/promhttp/instrument_server.go
generated
vendored
123
vendor/github.com/prometheus/client_golang/prometheus/promhttp/instrument_server.go
generated
vendored
@ -28,6 +28,26 @@ import (
|
|||||||
// magicString is used for the hacky label test in checkLabels. Remove once fixed.
|
// magicString is used for the hacky label test in checkLabels. Remove once fixed.
|
||||||
const magicString = "zZgWfBxLqvG8kc8IMv3POi2Bb0tZI3vAnBx+gBaFi9FyPzB/CzKUer1yufDa"
|
const magicString = "zZgWfBxLqvG8kc8IMv3POi2Bb0tZI3vAnBx+gBaFi9FyPzB/CzKUer1yufDa"
|
||||||
|
|
||||||
|
// observeWithExemplar is a wrapper for [prometheus.ExemplarAdder.ExemplarObserver],
|
||||||
|
// which falls back to [prometheus.Observer.Observe] if no labels are provided.
|
||||||
|
func observeWithExemplar(obs prometheus.Observer, val float64, labels map[string]string) {
|
||||||
|
if labels == nil {
|
||||||
|
obs.Observe(val)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
obs.(prometheus.ExemplarObserver).ObserveWithExemplar(val, labels)
|
||||||
|
}
|
||||||
|
|
||||||
|
// addWithExemplar is a wrapper for [prometheus.ExemplarAdder.AddWithExemplar],
|
||||||
|
// which falls back to [prometheus.Counter.Add] if no labels are provided.
|
||||||
|
func addWithExemplar(obs prometheus.Counter, val float64, labels map[string]string) {
|
||||||
|
if labels == nil {
|
||||||
|
obs.Add(val)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
obs.(prometheus.ExemplarAdder).AddWithExemplar(val, labels)
|
||||||
|
}
|
||||||
|
|
||||||
// InstrumentHandlerInFlight is a middleware that wraps the provided
|
// InstrumentHandlerInFlight is a middleware that wraps the provided
|
||||||
// http.Handler. It sets the provided prometheus.Gauge to the number of
|
// http.Handler. It sets the provided prometheus.Gauge to the number of
|
||||||
// requests currently handled by the wrapped http.Handler.
|
// requests currently handled by the wrapped http.Handler.
|
||||||
@ -48,7 +68,7 @@ func InstrumentHandlerInFlight(g prometheus.Gauge, next http.Handler) http.Handl
|
|||||||
// names are "code" and "method". The function panics otherwise. For the "method"
|
// names are "code" and "method". The function panics otherwise. For the "method"
|
||||||
// label a predefined default label value set is used to filter given values.
|
// label a predefined default label value set is used to filter given values.
|
||||||
// Values besides predefined values will count as `unknown` method.
|
// Values besides predefined values will count as `unknown` method.
|
||||||
//`WithExtraMethods` can be used to add more methods to the set. The Observe
|
// `WithExtraMethods` can be used to add more methods to the set. The Observe
|
||||||
// method of the Observer in the ObserverVec is called with the request duration
|
// method of the Observer in the ObserverVec is called with the request duration
|
||||||
// in seconds. Partitioning happens by HTTP status code and/or HTTP method if
|
// in seconds. Partitioning happens by HTTP status code and/or HTTP method if
|
||||||
// the respective instance label names are present in the ObserverVec. For
|
// the respective instance label names are present in the ObserverVec. For
|
||||||
@ -62,28 +82,37 @@ func InstrumentHandlerInFlight(g prometheus.Gauge, next http.Handler) http.Handl
|
|||||||
// Note that this method is only guaranteed to never observe negative durations
|
// Note that this method is only guaranteed to never observe negative durations
|
||||||
// if used with Go1.9+.
|
// if used with Go1.9+.
|
||||||
func InstrumentHandlerDuration(obs prometheus.ObserverVec, next http.Handler, opts ...Option) http.HandlerFunc {
|
func InstrumentHandlerDuration(obs prometheus.ObserverVec, next http.Handler, opts ...Option) http.HandlerFunc {
|
||||||
mwOpts := &option{}
|
hOpts := defaultOptions()
|
||||||
for _, o := range opts {
|
for _, o := range opts {
|
||||||
o(mwOpts)
|
o.apply(hOpts)
|
||||||
}
|
}
|
||||||
|
|
||||||
code, method := checkLabels(obs)
|
code, method := checkLabels(obs)
|
||||||
|
|
||||||
if code {
|
if code {
|
||||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
return func(w http.ResponseWriter, r *http.Request) {
|
||||||
now := time.Now()
|
now := time.Now()
|
||||||
d := newDelegator(w, nil)
|
d := newDelegator(w, nil)
|
||||||
next.ServeHTTP(d, r)
|
next.ServeHTTP(d, r)
|
||||||
|
|
||||||
obs.With(labels(code, method, r.Method, d.Status(), mwOpts.extraMethods...)).Observe(time.Since(now).Seconds())
|
observeWithExemplar(
|
||||||
})
|
obs.With(labels(code, method, r.Method, d.Status(), hOpts.extraMethods...)),
|
||||||
|
time.Since(now).Seconds(),
|
||||||
|
hOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
return func(w http.ResponseWriter, r *http.Request) {
|
||||||
now := time.Now()
|
now := time.Now()
|
||||||
next.ServeHTTP(w, r)
|
next.ServeHTTP(w, r)
|
||||||
obs.With(labels(code, method, r.Method, 0, mwOpts.extraMethods...)).Observe(time.Since(now).Seconds())
|
|
||||||
})
|
observeWithExemplar(
|
||||||
|
obs.With(labels(code, method, r.Method, 0, hOpts.extraMethods...)),
|
||||||
|
time.Since(now).Seconds(),
|
||||||
|
hOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// InstrumentHandlerCounter is a middleware that wraps the provided http.Handler
|
// InstrumentHandlerCounter is a middleware that wraps the provided http.Handler
|
||||||
@ -104,25 +133,34 @@ func InstrumentHandlerDuration(obs prometheus.ObserverVec, next http.Handler, op
|
|||||||
//
|
//
|
||||||
// See the example for InstrumentHandlerDuration for example usage.
|
// See the example for InstrumentHandlerDuration for example usage.
|
||||||
func InstrumentHandlerCounter(counter *prometheus.CounterVec, next http.Handler, opts ...Option) http.HandlerFunc {
|
func InstrumentHandlerCounter(counter *prometheus.CounterVec, next http.Handler, opts ...Option) http.HandlerFunc {
|
||||||
mwOpts := &option{}
|
hOpts := defaultOptions()
|
||||||
for _, o := range opts {
|
for _, o := range opts {
|
||||||
o(mwOpts)
|
o.apply(hOpts)
|
||||||
}
|
}
|
||||||
|
|
||||||
code, method := checkLabels(counter)
|
code, method := checkLabels(counter)
|
||||||
|
|
||||||
if code {
|
if code {
|
||||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
return func(w http.ResponseWriter, r *http.Request) {
|
||||||
d := newDelegator(w, nil)
|
d := newDelegator(w, nil)
|
||||||
next.ServeHTTP(d, r)
|
next.ServeHTTP(d, r)
|
||||||
counter.With(labels(code, method, r.Method, d.Status(), mwOpts.extraMethods...)).Inc()
|
|
||||||
})
|
addWithExemplar(
|
||||||
|
counter.With(labels(code, method, r.Method, d.Status(), hOpts.extraMethods...)),
|
||||||
|
1,
|
||||||
|
hOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
return func(w http.ResponseWriter, r *http.Request) {
|
||||||
next.ServeHTTP(w, r)
|
next.ServeHTTP(w, r)
|
||||||
counter.With(labels(code, method, r.Method, 0, mwOpts.extraMethods...)).Inc()
|
addWithExemplar(
|
||||||
})
|
counter.With(labels(code, method, r.Method, 0, hOpts.extraMethods...)),
|
||||||
|
1,
|
||||||
|
hOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// InstrumentHandlerTimeToWriteHeader is a middleware that wraps the provided
|
// InstrumentHandlerTimeToWriteHeader is a middleware that wraps the provided
|
||||||
@ -148,20 +186,24 @@ func InstrumentHandlerCounter(counter *prometheus.CounterVec, next http.Handler,
|
|||||||
//
|
//
|
||||||
// See the example for InstrumentHandlerDuration for example usage.
|
// See the example for InstrumentHandlerDuration for example usage.
|
||||||
func InstrumentHandlerTimeToWriteHeader(obs prometheus.ObserverVec, next http.Handler, opts ...Option) http.HandlerFunc {
|
func InstrumentHandlerTimeToWriteHeader(obs prometheus.ObserverVec, next http.Handler, opts ...Option) http.HandlerFunc {
|
||||||
mwOpts := &option{}
|
hOpts := defaultOptions()
|
||||||
for _, o := range opts {
|
for _, o := range opts {
|
||||||
o(mwOpts)
|
o.apply(hOpts)
|
||||||
}
|
}
|
||||||
|
|
||||||
code, method := checkLabels(obs)
|
code, method := checkLabels(obs)
|
||||||
|
|
||||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
return func(w http.ResponseWriter, r *http.Request) {
|
||||||
now := time.Now()
|
now := time.Now()
|
||||||
d := newDelegator(w, func(status int) {
|
d := newDelegator(w, func(status int) {
|
||||||
obs.With(labels(code, method, r.Method, status, mwOpts.extraMethods...)).Observe(time.Since(now).Seconds())
|
observeWithExemplar(
|
||||||
|
obs.With(labels(code, method, r.Method, status, hOpts.extraMethods...)),
|
||||||
|
time.Since(now).Seconds(),
|
||||||
|
hOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
})
|
})
|
||||||
next.ServeHTTP(d, r)
|
next.ServeHTTP(d, r)
|
||||||
})
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// InstrumentHandlerRequestSize is a middleware that wraps the provided
|
// InstrumentHandlerRequestSize is a middleware that wraps the provided
|
||||||
@ -184,27 +226,34 @@ func InstrumentHandlerTimeToWriteHeader(obs prometheus.ObserverVec, next http.Ha
|
|||||||
//
|
//
|
||||||
// See the example for InstrumentHandlerDuration for example usage.
|
// See the example for InstrumentHandlerDuration for example usage.
|
||||||
func InstrumentHandlerRequestSize(obs prometheus.ObserverVec, next http.Handler, opts ...Option) http.HandlerFunc {
|
func InstrumentHandlerRequestSize(obs prometheus.ObserverVec, next http.Handler, opts ...Option) http.HandlerFunc {
|
||||||
mwOpts := &option{}
|
hOpts := defaultOptions()
|
||||||
for _, o := range opts {
|
for _, o := range opts {
|
||||||
o(mwOpts)
|
o.apply(hOpts)
|
||||||
}
|
}
|
||||||
|
|
||||||
code, method := checkLabels(obs)
|
code, method := checkLabels(obs)
|
||||||
|
|
||||||
if code {
|
if code {
|
||||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
return func(w http.ResponseWriter, r *http.Request) {
|
||||||
d := newDelegator(w, nil)
|
d := newDelegator(w, nil)
|
||||||
next.ServeHTTP(d, r)
|
next.ServeHTTP(d, r)
|
||||||
size := computeApproximateRequestSize(r)
|
size := computeApproximateRequestSize(r)
|
||||||
obs.With(labels(code, method, r.Method, d.Status(), mwOpts.extraMethods...)).Observe(float64(size))
|
observeWithExemplar(
|
||||||
})
|
obs.With(labels(code, method, r.Method, d.Status(), hOpts.extraMethods...)),
|
||||||
|
float64(size),
|
||||||
|
hOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
return func(w http.ResponseWriter, r *http.Request) {
|
||||||
next.ServeHTTP(w, r)
|
next.ServeHTTP(w, r)
|
||||||
size := computeApproximateRequestSize(r)
|
size := computeApproximateRequestSize(r)
|
||||||
obs.With(labels(code, method, r.Method, 0, mwOpts.extraMethods...)).Observe(float64(size))
|
observeWithExemplar(
|
||||||
})
|
obs.With(labels(code, method, r.Method, 0, hOpts.extraMethods...)),
|
||||||
|
float64(size),
|
||||||
|
hOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// InstrumentHandlerResponseSize is a middleware that wraps the provided
|
// InstrumentHandlerResponseSize is a middleware that wraps the provided
|
||||||
@ -227,9 +276,9 @@ func InstrumentHandlerRequestSize(obs prometheus.ObserverVec, next http.Handler,
|
|||||||
//
|
//
|
||||||
// See the example for InstrumentHandlerDuration for example usage.
|
// See the example for InstrumentHandlerDuration for example usage.
|
||||||
func InstrumentHandlerResponseSize(obs prometheus.ObserverVec, next http.Handler, opts ...Option) http.Handler {
|
func InstrumentHandlerResponseSize(obs prometheus.ObserverVec, next http.Handler, opts ...Option) http.Handler {
|
||||||
mwOpts := &option{}
|
hOpts := defaultOptions()
|
||||||
for _, o := range opts {
|
for _, o := range opts {
|
||||||
o(mwOpts)
|
o.apply(hOpts)
|
||||||
}
|
}
|
||||||
|
|
||||||
code, method := checkLabels(obs)
|
code, method := checkLabels(obs)
|
||||||
@ -237,7 +286,11 @@ func InstrumentHandlerResponseSize(obs prometheus.ObserverVec, next http.Handler
|
|||||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
d := newDelegator(w, nil)
|
d := newDelegator(w, nil)
|
||||||
next.ServeHTTP(d, r)
|
next.ServeHTTP(d, r)
|
||||||
obs.With(labels(code, method, r.Method, d.Status(), mwOpts.extraMethods...)).Observe(float64(d.Written()))
|
observeWithExemplar(
|
||||||
|
obs.With(labels(code, method, r.Method, d.Status(), hOpts.extraMethods...)),
|
||||||
|
float64(d.Written()),
|
||||||
|
hOpts.getExemplarFn(r.Context()),
|
||||||
|
)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -246,7 +299,7 @@ func InstrumentHandlerResponseSize(obs prometheus.ObserverVec, next http.Handler
|
|||||||
// Collector does not have a Desc or has more than one Desc or its Desc is
|
// Collector does not have a Desc or has more than one Desc or its Desc is
|
||||||
// invalid. It also panics if the Collector has any non-const, non-curried
|
// invalid. It also panics if the Collector has any non-const, non-curried
|
||||||
// labels that are not named "code" or "method".
|
// labels that are not named "code" or "method".
|
||||||
func checkLabels(c prometheus.Collector) (code bool, method bool) {
|
func checkLabels(c prometheus.Collector) (code, method bool) {
|
||||||
// TODO(beorn7): Remove this hacky way to check for instance labels
|
// TODO(beorn7): Remove this hacky way to check for instance labels
|
||||||
// once Descriptors can have their dimensionality queried.
|
// once Descriptors can have their dimensionality queried.
|
||||||
var (
|
var (
|
||||||
|
39
vendor/github.com/prometheus/client_golang/prometheus/promhttp/option.go
generated
vendored
39
vendor/github.com/prometheus/client_golang/prometheus/promhttp/option.go
generated
vendored
@ -13,19 +13,46 @@
|
|||||||
|
|
||||||
package promhttp
|
package promhttp
|
||||||
|
|
||||||
// Option are used to configure a middleware or round tripper..
|
import (
|
||||||
type Option func(*option)
|
"context"
|
||||||
|
|
||||||
type option struct {
|
"github.com/prometheus/client_golang/prometheus"
|
||||||
extraMethods []string
|
)
|
||||||
|
|
||||||
|
// Option are used to configure both handler (middleware) or round tripper.
|
||||||
|
type Option interface {
|
||||||
|
apply(*options)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// options store options for both a handler or round tripper.
|
||||||
|
type options struct {
|
||||||
|
extraMethods []string
|
||||||
|
getExemplarFn func(requestCtx context.Context) prometheus.Labels
|
||||||
|
}
|
||||||
|
|
||||||
|
func defaultOptions() *options {
|
||||||
|
return &options{getExemplarFn: func(ctx context.Context) prometheus.Labels { return nil }}
|
||||||
|
}
|
||||||
|
|
||||||
|
type optionApplyFunc func(*options)
|
||||||
|
|
||||||
|
func (o optionApplyFunc) apply(opt *options) { o(opt) }
|
||||||
|
|
||||||
// WithExtraMethods adds additional HTTP methods to the list of allowed methods.
|
// WithExtraMethods adds additional HTTP methods to the list of allowed methods.
|
||||||
// See https://developer.mozilla.org/en-US/docs/Web/HTTP/Methods for the default list.
|
// See https://developer.mozilla.org/en-US/docs/Web/HTTP/Methods for the default list.
|
||||||
//
|
//
|
||||||
// See the example for ExampleInstrumentHandlerWithExtraMethods for example usage.
|
// See the example for ExampleInstrumentHandlerWithExtraMethods for example usage.
|
||||||
func WithExtraMethods(methods ...string) Option {
|
func WithExtraMethods(methods ...string) Option {
|
||||||
return func(o *option) {
|
return optionApplyFunc(func(o *options) {
|
||||||
o.extraMethods = methods
|
o.extraMethods = methods
|
||||||
}
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithExemplarFromContext adds allows to put a hook to all counter and histogram metrics.
|
||||||
|
// If the hook function returns non-nil labels, exemplars will be added for that request, otherwise metric
|
||||||
|
// will get instrumented without exemplar.
|
||||||
|
func WithExemplarFromContext(getExemplarFn func(requestCtx context.Context) prometheus.Labels) Option {
|
||||||
|
return optionApplyFunc(func(o *options) {
|
||||||
|
o.getExemplarFn = getExemplarFn
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
152
vendor/github.com/prometheus/client_golang/prometheus/registry.go
generated
vendored
152
vendor/github.com/prometheus/client_golang/prometheus/registry.go
generated
vendored
@ -15,8 +15,8 @@ package prometheus
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"runtime"
|
"runtime"
|
||||||
@ -252,9 +252,12 @@ func (errs MultiError) MaybeUnwrap() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Registry registers Prometheus collectors, collects their metrics, and gathers
|
// Registry registers Prometheus collectors, collects their metrics, and gathers
|
||||||
// them into MetricFamilies for exposition. It implements both Registerer and
|
// them into MetricFamilies for exposition. It implements Registerer, Gatherer,
|
||||||
// Gatherer. The zero value is not usable. Create instances with NewRegistry or
|
// and Collector. The zero value is not usable. Create instances with
|
||||||
// NewPedanticRegistry.
|
// NewRegistry or NewPedanticRegistry.
|
||||||
|
//
|
||||||
|
// Registry implements Collector to allow it to be used for creating groups of
|
||||||
|
// metrics. See the Grouping example for how this can be done.
|
||||||
type Registry struct {
|
type Registry struct {
|
||||||
mtx sync.RWMutex
|
mtx sync.RWMutex
|
||||||
collectorsByID map[uint64]Collector // ID is a hash of the descIDs.
|
collectorsByID map[uint64]Collector // ID is a hash of the descIDs.
|
||||||
@ -289,7 +292,7 @@ func (r *Registry) Register(c Collector) error {
|
|||||||
|
|
||||||
// Is the descriptor valid at all?
|
// Is the descriptor valid at all?
|
||||||
if desc.err != nil {
|
if desc.err != nil {
|
||||||
return fmt.Errorf("descriptor %s is invalid: %s", desc, desc.err)
|
return fmt.Errorf("descriptor %s is invalid: %w", desc, desc.err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Is the descID unique?
|
// Is the descID unique?
|
||||||
@ -407,6 +410,14 @@ func (r *Registry) MustRegister(cs ...Collector) {
|
|||||||
|
|
||||||
// Gather implements Gatherer.
|
// Gather implements Gatherer.
|
||||||
func (r *Registry) Gather() ([]*dto.MetricFamily, error) {
|
func (r *Registry) Gather() ([]*dto.MetricFamily, error) {
|
||||||
|
r.mtx.RLock()
|
||||||
|
|
||||||
|
if len(r.collectorsByID) == 0 && len(r.uncheckedCollectors) == 0 {
|
||||||
|
// Fast path.
|
||||||
|
r.mtx.RUnlock()
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
var (
|
var (
|
||||||
checkedMetricChan = make(chan Metric, capMetricChan)
|
checkedMetricChan = make(chan Metric, capMetricChan)
|
||||||
uncheckedMetricChan = make(chan Metric, capMetricChan)
|
uncheckedMetricChan = make(chan Metric, capMetricChan)
|
||||||
@ -416,7 +427,6 @@ func (r *Registry) Gather() ([]*dto.MetricFamily, error) {
|
|||||||
registeredDescIDs map[uint64]struct{} // Only used for pedantic checks
|
registeredDescIDs map[uint64]struct{} // Only used for pedantic checks
|
||||||
)
|
)
|
||||||
|
|
||||||
r.mtx.RLock()
|
|
||||||
goroutineBudget := len(r.collectorsByID) + len(r.uncheckedCollectors)
|
goroutineBudget := len(r.collectorsByID) + len(r.uncheckedCollectors)
|
||||||
metricFamiliesByName := make(map[string]*dto.MetricFamily, len(r.dimHashesByName))
|
metricFamiliesByName := make(map[string]*dto.MetricFamily, len(r.dimHashesByName))
|
||||||
checkedCollectors := make(chan Collector, len(r.collectorsByID))
|
checkedCollectors := make(chan Collector, len(r.collectorsByID))
|
||||||
@ -549,6 +559,31 @@ func (r *Registry) Gather() ([]*dto.MetricFamily, error) {
|
|||||||
return internal.NormalizeMetricFamilies(metricFamiliesByName), errs.MaybeUnwrap()
|
return internal.NormalizeMetricFamilies(metricFamiliesByName), errs.MaybeUnwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Describe implements Collector.
|
||||||
|
func (r *Registry) Describe(ch chan<- *Desc) {
|
||||||
|
r.mtx.RLock()
|
||||||
|
defer r.mtx.RUnlock()
|
||||||
|
|
||||||
|
// Only report the checked Collectors; unchecked collectors don't report any
|
||||||
|
// Desc.
|
||||||
|
for _, c := range r.collectorsByID {
|
||||||
|
c.Describe(ch)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Collect implements Collector.
|
||||||
|
func (r *Registry) Collect(ch chan<- Metric) {
|
||||||
|
r.mtx.RLock()
|
||||||
|
defer r.mtx.RUnlock()
|
||||||
|
|
||||||
|
for _, c := range r.collectorsByID {
|
||||||
|
c.Collect(ch)
|
||||||
|
}
|
||||||
|
for _, c := range r.uncheckedCollectors {
|
||||||
|
c.Collect(ch)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// WriteToTextfile calls Gather on the provided Gatherer, encodes the result in the
|
// WriteToTextfile calls Gather on the provided Gatherer, encodes the result in the
|
||||||
// Prometheus text format, and writes it to a temporary file. Upon success, the
|
// Prometheus text format, and writes it to a temporary file. Upon success, the
|
||||||
// temporary file is renamed to the provided filename.
|
// temporary file is renamed to the provided filename.
|
||||||
@ -556,7 +591,7 @@ func (r *Registry) Gather() ([]*dto.MetricFamily, error) {
|
|||||||
// This is intended for use with the textfile collector of the node exporter.
|
// This is intended for use with the textfile collector of the node exporter.
|
||||||
// Note that the node exporter expects the filename to be suffixed with ".prom".
|
// Note that the node exporter expects the filename to be suffixed with ".prom".
|
||||||
func WriteToTextfile(filename string, g Gatherer) error {
|
func WriteToTextfile(filename string, g Gatherer) error {
|
||||||
tmp, err := ioutil.TempFile(filepath.Dir(filename), filepath.Base(filename))
|
tmp, err := os.CreateTemp(filepath.Dir(filename), filepath.Base(filename))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -575,7 +610,7 @@ func WriteToTextfile(filename string, g Gatherer) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := os.Chmod(tmp.Name(), 0644); err != nil {
|
if err := os.Chmod(tmp.Name(), 0o644); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
return os.Rename(tmp.Name(), filename)
|
return os.Rename(tmp.Name(), filename)
|
||||||
@ -596,7 +631,7 @@ func processMetric(
|
|||||||
}
|
}
|
||||||
dtoMetric := &dto.Metric{}
|
dtoMetric := &dto.Metric{}
|
||||||
if err := metric.Write(dtoMetric); err != nil {
|
if err := metric.Write(dtoMetric); err != nil {
|
||||||
return fmt.Errorf("error collecting metric %v: %s", desc, err)
|
return fmt.Errorf("error collecting metric %v: %w", desc, err)
|
||||||
}
|
}
|
||||||
metricFamily, ok := metricFamiliesByName[desc.fqName]
|
metricFamily, ok := metricFamiliesByName[desc.fqName]
|
||||||
if ok { // Existing name.
|
if ok { // Existing name.
|
||||||
@ -718,12 +753,13 @@ func (gs Gatherers) Gather() ([]*dto.MetricFamily, error) {
|
|||||||
for i, g := range gs {
|
for i, g := range gs {
|
||||||
mfs, err := g.Gather()
|
mfs, err := g.Gather()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if multiErr, ok := err.(MultiError); ok {
|
multiErr := MultiError{}
|
||||||
|
if errors.As(err, &multiErr) {
|
||||||
for _, err := range multiErr {
|
for _, err := range multiErr {
|
||||||
errs = append(errs, fmt.Errorf("[from Gatherer #%d] %s", i+1, err))
|
errs = append(errs, fmt.Errorf("[from Gatherer #%d] %w", i+1, err))
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
errs = append(errs, fmt.Errorf("[from Gatherer #%d] %s", i+1, err))
|
errs = append(errs, fmt.Errorf("[from Gatherer #%d] %w", i+1, err))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for _, mf := range mfs {
|
for _, mf := range mfs {
|
||||||
@ -884,11 +920,11 @@ func checkMetricConsistency(
|
|||||||
h.Write(separatorByteSlice)
|
h.Write(separatorByteSlice)
|
||||||
// Make sure label pairs are sorted. We depend on it for the consistency
|
// Make sure label pairs are sorted. We depend on it for the consistency
|
||||||
// check.
|
// check.
|
||||||
if !sort.IsSorted(labelPairSorter(dtoMetric.Label)) {
|
if !sort.IsSorted(internal.LabelPairSorter(dtoMetric.Label)) {
|
||||||
// We cannot sort dtoMetric.Label in place as it is immutable by contract.
|
// We cannot sort dtoMetric.Label in place as it is immutable by contract.
|
||||||
copiedLabels := make([]*dto.LabelPair, len(dtoMetric.Label))
|
copiedLabels := make([]*dto.LabelPair, len(dtoMetric.Label))
|
||||||
copy(copiedLabels, dtoMetric.Label)
|
copy(copiedLabels, dtoMetric.Label)
|
||||||
sort.Sort(labelPairSorter(copiedLabels))
|
sort.Sort(internal.LabelPairSorter(copiedLabels))
|
||||||
dtoMetric.Label = copiedLabels
|
dtoMetric.Label = copiedLabels
|
||||||
}
|
}
|
||||||
for _, lp := range dtoMetric.Label {
|
for _, lp := range dtoMetric.Label {
|
||||||
@ -935,7 +971,7 @@ func checkDescConsistency(
|
|||||||
metricFamily.GetName(), dtoMetric, desc,
|
metricFamily.GetName(), dtoMetric, desc,
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
sort.Sort(labelPairSorter(lpsFromDesc))
|
sort.Sort(internal.LabelPairSorter(lpsFromDesc))
|
||||||
for i, lpFromDesc := range lpsFromDesc {
|
for i, lpFromDesc := range lpsFromDesc {
|
||||||
lpFromMetric := dtoMetric.Label[i]
|
lpFromMetric := dtoMetric.Label[i]
|
||||||
if lpFromDesc.GetName() != lpFromMetric.GetName() ||
|
if lpFromDesc.GetName() != lpFromMetric.GetName() ||
|
||||||
@ -948,3 +984,89 @@ func checkDescConsistency(
|
|||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var _ TransactionalGatherer = &MultiTRegistry{}
|
||||||
|
|
||||||
|
// MultiTRegistry is a TransactionalGatherer that joins gathered metrics from multiple
|
||||||
|
// transactional gatherers.
|
||||||
|
//
|
||||||
|
// It is caller responsibility to ensure two registries have mutually exclusive metric families,
|
||||||
|
// no deduplication will happen.
|
||||||
|
type MultiTRegistry struct {
|
||||||
|
tGatherers []TransactionalGatherer
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewMultiTRegistry creates MultiTRegistry.
|
||||||
|
func NewMultiTRegistry(tGatherers ...TransactionalGatherer) *MultiTRegistry {
|
||||||
|
return &MultiTRegistry{
|
||||||
|
tGatherers: tGatherers,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Gather implements TransactionalGatherer interface.
|
||||||
|
func (r *MultiTRegistry) Gather() (mfs []*dto.MetricFamily, done func(), err error) {
|
||||||
|
errs := MultiError{}
|
||||||
|
|
||||||
|
dFns := make([]func(), 0, len(r.tGatherers))
|
||||||
|
// TODO(bwplotka): Implement concurrency for those?
|
||||||
|
for _, g := range r.tGatherers {
|
||||||
|
// TODO(bwplotka): Check for duplicates?
|
||||||
|
m, d, err := g.Gather()
|
||||||
|
errs.Append(err)
|
||||||
|
|
||||||
|
mfs = append(mfs, m...)
|
||||||
|
dFns = append(dFns, d)
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO(bwplotka): Consider sort in place, given metric family in gather is sorted already.
|
||||||
|
sort.Slice(mfs, func(i, j int) bool {
|
||||||
|
return *mfs[i].Name < *mfs[j].Name
|
||||||
|
})
|
||||||
|
return mfs, func() {
|
||||||
|
for _, d := range dFns {
|
||||||
|
d()
|
||||||
|
}
|
||||||
|
}, errs.MaybeUnwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
// TransactionalGatherer represents transactional gatherer that can be triggered to notify gatherer that memory
|
||||||
|
// used by metric family is no longer used by a caller. This allows implementations with cache.
|
||||||
|
type TransactionalGatherer interface {
|
||||||
|
// Gather returns metrics in a lexicographically sorted slice
|
||||||
|
// of uniquely named MetricFamily protobufs. Gather ensures that the
|
||||||
|
// returned slice is valid and self-consistent so that it can be used
|
||||||
|
// for valid exposition. As an exception to the strict consistency
|
||||||
|
// requirements described for metric.Desc, Gather will tolerate
|
||||||
|
// different sets of label names for metrics of the same metric family.
|
||||||
|
//
|
||||||
|
// Even if an error occurs, Gather attempts to gather as many metrics as
|
||||||
|
// possible. Hence, if a non-nil error is returned, the returned
|
||||||
|
// MetricFamily slice could be nil (in case of a fatal error that
|
||||||
|
// prevented any meaningful metric collection) or contain a number of
|
||||||
|
// MetricFamily protobufs, some of which might be incomplete, and some
|
||||||
|
// might be missing altogether. The returned error (which might be a
|
||||||
|
// MultiError) explains the details. Note that this is mostly useful for
|
||||||
|
// debugging purposes. If the gathered protobufs are to be used for
|
||||||
|
// exposition in actual monitoring, it is almost always better to not
|
||||||
|
// expose an incomplete result and instead disregard the returned
|
||||||
|
// MetricFamily protobufs in case the returned error is non-nil.
|
||||||
|
//
|
||||||
|
// Important: done is expected to be triggered (even if the error occurs!)
|
||||||
|
// once caller does not need returned slice of dto.MetricFamily.
|
||||||
|
Gather() (_ []*dto.MetricFamily, done func(), err error)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ToTransactionalGatherer transforms Gatherer to transactional one with noop as done function.
|
||||||
|
func ToTransactionalGatherer(g Gatherer) TransactionalGatherer {
|
||||||
|
return &noTransactionGatherer{g: g}
|
||||||
|
}
|
||||||
|
|
||||||
|
type noTransactionGatherer struct {
|
||||||
|
g Gatherer
|
||||||
|
}
|
||||||
|
|
||||||
|
// Gather implements TransactionalGatherer interface.
|
||||||
|
func (g *noTransactionGatherer) Gather() (_ []*dto.MetricFamily, done func(), err error) {
|
||||||
|
mfs, err := g.g.Gather()
|
||||||
|
return mfs, func() {}, err
|
||||||
|
}
|
||||||
|
9
vendor/github.com/prometheus/client_golang/prometheus/summary.go
generated
vendored
9
vendor/github.com/prometheus/client_golang/prometheus/summary.go
generated
vendored
@ -603,7 +603,8 @@ func (v *SummaryVec) GetMetricWith(labels Labels) (Observer, error) {
|
|||||||
// WithLabelValues works as GetMetricWithLabelValues, but panics where
|
// WithLabelValues works as GetMetricWithLabelValues, but panics where
|
||||||
// GetMetricWithLabelValues would have returned an error. Not returning an
|
// GetMetricWithLabelValues would have returned an error. Not returning an
|
||||||
// error allows shortcuts like
|
// error allows shortcuts like
|
||||||
// myVec.WithLabelValues("404", "GET").Observe(42.21)
|
//
|
||||||
|
// myVec.WithLabelValues("404", "GET").Observe(42.21)
|
||||||
func (v *SummaryVec) WithLabelValues(lvs ...string) Observer {
|
func (v *SummaryVec) WithLabelValues(lvs ...string) Observer {
|
||||||
s, err := v.GetMetricWithLabelValues(lvs...)
|
s, err := v.GetMetricWithLabelValues(lvs...)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -614,7 +615,8 @@ func (v *SummaryVec) WithLabelValues(lvs ...string) Observer {
|
|||||||
|
|
||||||
// With works as GetMetricWith, but panics where GetMetricWithLabels would have
|
// With works as GetMetricWith, but panics where GetMetricWithLabels would have
|
||||||
// returned an error. Not returning an error allows shortcuts like
|
// returned an error. Not returning an error allows shortcuts like
|
||||||
// myVec.With(prometheus.Labels{"code": "404", "method": "GET"}).Observe(42.21)
|
//
|
||||||
|
// myVec.With(prometheus.Labels{"code": "404", "method": "GET"}).Observe(42.21)
|
||||||
func (v *SummaryVec) With(labels Labels) Observer {
|
func (v *SummaryVec) With(labels Labels) Observer {
|
||||||
s, err := v.GetMetricWith(labels)
|
s, err := v.GetMetricWith(labels)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -701,7 +703,8 @@ func (s *constSummary) Write(out *dto.Metric) error {
|
|||||||
//
|
//
|
||||||
// quantiles maps ranks to quantile values. For example, a median latency of
|
// quantiles maps ranks to quantile values. For example, a median latency of
|
||||||
// 0.23s and a 99th percentile latency of 0.56s would be expressed as:
|
// 0.23s and a 99th percentile latency of 0.56s would be expressed as:
|
||||||
// map[float64]float64{0.5: 0.23, 0.99: 0.56}
|
//
|
||||||
|
// map[float64]float64{0.5: 0.23, 0.99: 0.56}
|
||||||
//
|
//
|
||||||
// NewConstSummary returns an error if the length of labelValues is not
|
// NewConstSummary returns an error if the length of labelValues is not
|
||||||
// consistent with the variable labels in Desc or if Desc is invalid.
|
// consistent with the variable labels in Desc or if Desc is invalid.
|
||||||
|
4
vendor/github.com/prometheus/client_golang/prometheus/testutil/lint.go
generated
vendored
4
vendor/github.com/prometheus/client_golang/prometheus/testutil/lint.go
generated
vendored
@ -26,7 +26,7 @@ import (
|
|||||||
func CollectAndLint(c prometheus.Collector, metricNames ...string) ([]promlint.Problem, error) {
|
func CollectAndLint(c prometheus.Collector, metricNames ...string) ([]promlint.Problem, error) {
|
||||||
reg := prometheus.NewPedanticRegistry()
|
reg := prometheus.NewPedanticRegistry()
|
||||||
if err := reg.Register(c); err != nil {
|
if err := reg.Register(c); err != nil {
|
||||||
return nil, fmt.Errorf("registering collector failed: %s", err)
|
return nil, fmt.Errorf("registering collector failed: %w", err)
|
||||||
}
|
}
|
||||||
return GatherAndLint(reg, metricNames...)
|
return GatherAndLint(reg, metricNames...)
|
||||||
}
|
}
|
||||||
@ -37,7 +37,7 @@ func CollectAndLint(c prometheus.Collector, metricNames ...string) ([]promlint.P
|
|||||||
func GatherAndLint(g prometheus.Gatherer, metricNames ...string) ([]promlint.Problem, error) {
|
func GatherAndLint(g prometheus.Gatherer, metricNames ...string) ([]promlint.Problem, error) {
|
||||||
got, err := g.Gather()
|
got, err := g.Gather()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("gathering metrics failed: %s", err)
|
return nil, fmt.Errorf("gathering metrics failed: %w", err)
|
||||||
}
|
}
|
||||||
if metricNames != nil {
|
if metricNames != nil {
|
||||||
got = filterMetrics(got, metricNames)
|
got = filterMetrics(got, metricNames)
|
||||||
|
5
vendor/github.com/prometheus/client_golang/prometheus/testutil/promlint/promlint.go
generated
vendored
5
vendor/github.com/prometheus/client_golang/prometheus/testutil/promlint/promlint.go
generated
vendored
@ -15,6 +15,7 @@
|
|||||||
package promlint
|
package promlint
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"regexp"
|
"regexp"
|
||||||
@ -83,7 +84,7 @@ func (l *Linter) Lint() ([]Problem, error) {
|
|||||||
mf := &dto.MetricFamily{}
|
mf := &dto.MetricFamily{}
|
||||||
for {
|
for {
|
||||||
if err := d.Decode(mf); err != nil {
|
if err := d.Decode(mf); err != nil {
|
||||||
if err == io.EOF {
|
if errors.Is(err, io.EOF) {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -283,7 +284,7 @@ func lintUnitAbbreviations(mf *dto.MetricFamily) []Problem {
|
|||||||
|
|
||||||
// metricUnits attempts to detect known unit types used as part of a metric name,
|
// metricUnits attempts to detect known unit types used as part of a metric name,
|
||||||
// e.g. "foo_bytes_total" or "bar_baz_milligrams".
|
// e.g. "foo_bytes_total" or "bar_baz_milligrams".
|
||||||
func metricUnits(m string) (unit string, base string, ok bool) {
|
func metricUnits(m string) (unit, base string, ok bool) {
|
||||||
ss := strings.Split(m, "_")
|
ss := strings.Split(m, "_")
|
||||||
|
|
||||||
for unit, base := range units {
|
for unit, base := range units {
|
||||||
|
166
vendor/github.com/prometheus/client_golang/prometheus/testutil/testutil.go
generated
vendored
166
vendor/github.com/prometheus/client_golang/prometheus/testutil/testutil.go
generated
vendored
@ -41,10 +41,12 @@ import (
|
|||||||
"bytes"
|
"bytes"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
|
"net/http"
|
||||||
|
"reflect"
|
||||||
|
|
||||||
"github.com/prometheus/common/expfmt"
|
"github.com/davecgh/go-spew/spew"
|
||||||
|
|
||||||
dto "github.com/prometheus/client_model/go"
|
dto "github.com/prometheus/client_model/go"
|
||||||
|
"github.com/prometheus/common/expfmt"
|
||||||
|
|
||||||
"github.com/prometheus/client_golang/prometheus"
|
"github.com/prometheus/client_golang/prometheus"
|
||||||
"github.com/prometheus/client_golang/prometheus/internal"
|
"github.com/prometheus/client_golang/prometheus/internal"
|
||||||
@ -99,7 +101,9 @@ func ToFloat64(c prometheus.Collector) float64 {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pb := &dto.Metric{}
|
pb := &dto.Metric{}
|
||||||
m.Write(pb)
|
if err := m.Write(pb); err != nil {
|
||||||
|
panic(fmt.Errorf("error happened while collecting metrics: %w", err))
|
||||||
|
}
|
||||||
if pb.Gauge != nil {
|
if pb.Gauge != nil {
|
||||||
return pb.Gauge.GetValue()
|
return pb.Gauge.GetValue()
|
||||||
}
|
}
|
||||||
@ -122,7 +126,7 @@ func ToFloat64(c prometheus.Collector) float64 {
|
|||||||
func CollectAndCount(c prometheus.Collector, metricNames ...string) int {
|
func CollectAndCount(c prometheus.Collector, metricNames ...string) int {
|
||||||
reg := prometheus.NewPedanticRegistry()
|
reg := prometheus.NewPedanticRegistry()
|
||||||
if err := reg.Register(c); err != nil {
|
if err := reg.Register(c); err != nil {
|
||||||
panic(fmt.Errorf("registering collector failed: %s", err))
|
panic(fmt.Errorf("registering collector failed: %w", err))
|
||||||
}
|
}
|
||||||
result, err := GatherAndCount(reg, metricNames...)
|
result, err := GatherAndCount(reg, metricNames...)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -138,7 +142,7 @@ func CollectAndCount(c prometheus.Collector, metricNames ...string) int {
|
|||||||
func GatherAndCount(g prometheus.Gatherer, metricNames ...string) (int, error) {
|
func GatherAndCount(g prometheus.Gatherer, metricNames ...string) (int, error) {
|
||||||
got, err := g.Gather()
|
got, err := g.Gather()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, fmt.Errorf("gathering metrics failed: %s", err)
|
return 0, fmt.Errorf("gathering metrics failed: %w", err)
|
||||||
}
|
}
|
||||||
if metricNames != nil {
|
if metricNames != nil {
|
||||||
got = filterMetrics(got, metricNames)
|
got = filterMetrics(got, metricNames)
|
||||||
@ -151,13 +155,41 @@ func GatherAndCount(g prometheus.Gatherer, metricNames ...string) (int, error) {
|
|||||||
return result, nil
|
return result, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ScrapeAndCompare calls a remote exporter's endpoint which is expected to return some metrics in
|
||||||
|
// plain text format. Then it compares it with the results that the `expected` would return.
|
||||||
|
// If the `metricNames` is not empty it would filter the comparison only to the given metric names.
|
||||||
|
func ScrapeAndCompare(url string, expected io.Reader, metricNames ...string) error {
|
||||||
|
resp, err := http.Get(url)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("scraping metrics failed: %w", err)
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
if resp.StatusCode != http.StatusOK {
|
||||||
|
return fmt.Errorf("the scraping target returned a status code other than 200: %d",
|
||||||
|
resp.StatusCode)
|
||||||
|
}
|
||||||
|
|
||||||
|
scraped, err := convertReaderToMetricFamily(resp.Body)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
wanted, err := convertReaderToMetricFamily(expected)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
return compareMetricFamilies(scraped, wanted, metricNames...)
|
||||||
|
}
|
||||||
|
|
||||||
// CollectAndCompare registers the provided Collector with a newly created
|
// CollectAndCompare registers the provided Collector with a newly created
|
||||||
// pedantic Registry. It then calls GatherAndCompare with that Registry and with
|
// pedantic Registry. It then calls GatherAndCompare with that Registry and with
|
||||||
// the provided metricNames.
|
// the provided metricNames.
|
||||||
func CollectAndCompare(c prometheus.Collector, expected io.Reader, metricNames ...string) error {
|
func CollectAndCompare(c prometheus.Collector, expected io.Reader, metricNames ...string) error {
|
||||||
reg := prometheus.NewPedanticRegistry()
|
reg := prometheus.NewPedanticRegistry()
|
||||||
if err := reg.Register(c); err != nil {
|
if err := reg.Register(c); err != nil {
|
||||||
return fmt.Errorf("registering collector failed: %s", err)
|
return fmt.Errorf("registering collector failed: %w", err)
|
||||||
}
|
}
|
||||||
return GatherAndCompare(reg, expected, metricNames...)
|
return GatherAndCompare(reg, expected, metricNames...)
|
||||||
}
|
}
|
||||||
@ -167,21 +199,48 @@ func CollectAndCompare(c prometheus.Collector, expected io.Reader, metricNames .
|
|||||||
// exposition format. If any metricNames are provided, only metrics with those
|
// exposition format. If any metricNames are provided, only metrics with those
|
||||||
// names are compared.
|
// names are compared.
|
||||||
func GatherAndCompare(g prometheus.Gatherer, expected io.Reader, metricNames ...string) error {
|
func GatherAndCompare(g prometheus.Gatherer, expected io.Reader, metricNames ...string) error {
|
||||||
got, err := g.Gather()
|
return TransactionalGatherAndCompare(prometheus.ToTransactionalGatherer(g), expected, metricNames...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// TransactionalGatherAndCompare gathers all metrics from the provided Gatherer and compares
|
||||||
|
// it to an expected output read from the provided Reader in the Prometheus text
|
||||||
|
// exposition format. If any metricNames are provided, only metrics with those
|
||||||
|
// names are compared.
|
||||||
|
func TransactionalGatherAndCompare(g prometheus.TransactionalGatherer, expected io.Reader, metricNames ...string) error {
|
||||||
|
got, done, err := g.Gather()
|
||||||
|
defer done()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("gathering metrics failed: %s", err)
|
return fmt.Errorf("gathering metrics failed: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
wanted, err := convertReaderToMetricFamily(expected)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
return compareMetricFamilies(got, wanted, metricNames...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// convertReaderToMetricFamily would read from a io.Reader object and convert it to a slice of
|
||||||
|
// dto.MetricFamily.
|
||||||
|
func convertReaderToMetricFamily(reader io.Reader) ([]*dto.MetricFamily, error) {
|
||||||
|
var tp expfmt.TextParser
|
||||||
|
notNormalized, err := tp.TextToMetricFamilies(reader)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("converting reader to metric families failed: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return internal.NormalizeMetricFamilies(notNormalized), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// compareMetricFamilies would compare 2 slices of metric families, and optionally filters both of
|
||||||
|
// them to the `metricNames` provided.
|
||||||
|
func compareMetricFamilies(got, expected []*dto.MetricFamily, metricNames ...string) error {
|
||||||
if metricNames != nil {
|
if metricNames != nil {
|
||||||
got = filterMetrics(got, metricNames)
|
got = filterMetrics(got, metricNames)
|
||||||
}
|
}
|
||||||
var tp expfmt.TextParser
|
|
||||||
wantRaw, err := tp.TextToMetricFamilies(expected)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("parsing expected metrics failed: %s", err)
|
|
||||||
}
|
|
||||||
want := internal.NormalizeMetricFamilies(wantRaw)
|
|
||||||
|
|
||||||
return compare(got, want)
|
return compare(got, expected)
|
||||||
}
|
}
|
||||||
|
|
||||||
// compare encodes both provided slices of metric families into the text format,
|
// compare encodes both provided slices of metric families into the text format,
|
||||||
@ -193,29 +252,82 @@ func compare(got, want []*dto.MetricFamily) error {
|
|||||||
enc := expfmt.NewEncoder(&gotBuf, expfmt.FmtText)
|
enc := expfmt.NewEncoder(&gotBuf, expfmt.FmtText)
|
||||||
for _, mf := range got {
|
for _, mf := range got {
|
||||||
if err := enc.Encode(mf); err != nil {
|
if err := enc.Encode(mf); err != nil {
|
||||||
return fmt.Errorf("encoding gathered metrics failed: %s", err)
|
return fmt.Errorf("encoding gathered metrics failed: %w", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
enc = expfmt.NewEncoder(&wantBuf, expfmt.FmtText)
|
enc = expfmt.NewEncoder(&wantBuf, expfmt.FmtText)
|
||||||
for _, mf := range want {
|
for _, mf := range want {
|
||||||
if err := enc.Encode(mf); err != nil {
|
if err := enc.Encode(mf); err != nil {
|
||||||
return fmt.Errorf("encoding expected metrics failed: %s", err)
|
return fmt.Errorf("encoding expected metrics failed: %w", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
if diffErr := diff(wantBuf, gotBuf); diffErr != "" {
|
||||||
if wantBuf.String() != gotBuf.String() {
|
return fmt.Errorf(diffErr)
|
||||||
return fmt.Errorf(`
|
|
||||||
metric output does not match expectation; want:
|
|
||||||
|
|
||||||
%s
|
|
||||||
got:
|
|
||||||
|
|
||||||
%s`, wantBuf.String(), gotBuf.String())
|
|
||||||
|
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// diff returns a diff of both values as long as both are of the same type and
|
||||||
|
// are a struct, map, slice, array or string. Otherwise it returns an empty string.
|
||||||
|
func diff(expected, actual interface{}) string {
|
||||||
|
if expected == nil || actual == nil {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
et, ek := typeAndKind(expected)
|
||||||
|
at, _ := typeAndKind(actual)
|
||||||
|
if et != at {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
if ek != reflect.Struct && ek != reflect.Map && ek != reflect.Slice && ek != reflect.Array && ek != reflect.String {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
var e, a string
|
||||||
|
c := spew.ConfigState{
|
||||||
|
Indent: " ",
|
||||||
|
DisablePointerAddresses: true,
|
||||||
|
DisableCapacities: true,
|
||||||
|
SortKeys: true,
|
||||||
|
}
|
||||||
|
if et != reflect.TypeOf("") {
|
||||||
|
e = c.Sdump(expected)
|
||||||
|
a = c.Sdump(actual)
|
||||||
|
} else {
|
||||||
|
e = reflect.ValueOf(expected).String()
|
||||||
|
a = reflect.ValueOf(actual).String()
|
||||||
|
}
|
||||||
|
|
||||||
|
diff, _ := internal.GetUnifiedDiffString(internal.UnifiedDiff{
|
||||||
|
A: internal.SplitLines(e),
|
||||||
|
B: internal.SplitLines(a),
|
||||||
|
FromFile: "metric output does not match expectation; want",
|
||||||
|
FromDate: "",
|
||||||
|
ToFile: "got:",
|
||||||
|
ToDate: "",
|
||||||
|
Context: 1,
|
||||||
|
})
|
||||||
|
|
||||||
|
if diff == "" {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
return "\n\nDiff:\n" + diff
|
||||||
|
}
|
||||||
|
|
||||||
|
// typeAndKind returns the type and kind of the given interface{}
|
||||||
|
func typeAndKind(v interface{}) (reflect.Type, reflect.Kind) {
|
||||||
|
t := reflect.TypeOf(v)
|
||||||
|
k := t.Kind()
|
||||||
|
|
||||||
|
if k == reflect.Ptr {
|
||||||
|
t = t.Elem()
|
||||||
|
k = t.Kind()
|
||||||
|
}
|
||||||
|
return t, k
|
||||||
|
}
|
||||||
|
|
||||||
func filterMetrics(metrics []*dto.MetricFamily, names []string) []*dto.MetricFamily {
|
func filterMetrics(metrics []*dto.MetricFamily, names []string) []*dto.MetricFamily {
|
||||||
var filtered []*dto.MetricFamily
|
var filtered []*dto.MetricFamily
|
||||||
for _, m := range metrics {
|
for _, m := range metrics {
|
||||||
|
11
vendor/github.com/prometheus/client_golang/prometheus/timer.go
generated
vendored
11
vendor/github.com/prometheus/client_golang/prometheus/timer.go
generated
vendored
@ -25,11 +25,12 @@ type Timer struct {
|
|||||||
// NewTimer creates a new Timer. The provided Observer is used to observe a
|
// NewTimer creates a new Timer. The provided Observer is used to observe a
|
||||||
// duration in seconds. Timer is usually used to time a function call in the
|
// duration in seconds. Timer is usually used to time a function call in the
|
||||||
// following way:
|
// following way:
|
||||||
// func TimeMe() {
|
//
|
||||||
// timer := NewTimer(myHistogram)
|
// func TimeMe() {
|
||||||
// defer timer.ObserveDuration()
|
// timer := NewTimer(myHistogram)
|
||||||
// // Do actual work.
|
// defer timer.ObserveDuration()
|
||||||
// }
|
// // Do actual work.
|
||||||
|
// }
|
||||||
func NewTimer(o Observer) *Timer {
|
func NewTimer(o Observer) *Timer {
|
||||||
return &Timer{
|
return &Timer{
|
||||||
begin: time.Now(),
|
begin: time.Now(),
|
||||||
|
47
vendor/github.com/prometheus/client_golang/prometheus/value.go
generated
vendored
47
vendor/github.com/prometheus/client_golang/prometheus/value.go
generated
vendored
@ -23,6 +23,8 @@ import (
|
|||||||
"github.com/golang/protobuf/proto"
|
"github.com/golang/protobuf/proto"
|
||||||
"google.golang.org/protobuf/types/known/timestamppb"
|
"google.golang.org/protobuf/types/known/timestamppb"
|
||||||
|
|
||||||
|
"github.com/prometheus/client_golang/prometheus/internal"
|
||||||
|
|
||||||
dto "github.com/prometheus/client_model/go"
|
dto "github.com/prometheus/client_model/go"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -38,6 +40,23 @@ const (
|
|||||||
UntypedValue
|
UntypedValue
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
CounterMetricTypePtr = func() *dto.MetricType { d := dto.MetricType_COUNTER; return &d }()
|
||||||
|
GaugeMetricTypePtr = func() *dto.MetricType { d := dto.MetricType_GAUGE; return &d }()
|
||||||
|
UntypedMetricTypePtr = func() *dto.MetricType { d := dto.MetricType_UNTYPED; return &d }()
|
||||||
|
)
|
||||||
|
|
||||||
|
func (v ValueType) ToDTO() *dto.MetricType {
|
||||||
|
switch v {
|
||||||
|
case CounterValue:
|
||||||
|
return CounterMetricTypePtr
|
||||||
|
case GaugeValue:
|
||||||
|
return GaugeMetricTypePtr
|
||||||
|
default:
|
||||||
|
return UntypedMetricTypePtr
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// valueFunc is a generic metric for simple values retrieved on collect time
|
// valueFunc is a generic metric for simple values retrieved on collect time
|
||||||
// from a function. It implements Metric and Collector. Its effective type is
|
// from a function. It implements Metric and Collector. Its effective type is
|
||||||
// determined by ValueType. This is a low-level building block used by the
|
// determined by ValueType. This is a low-level building block used by the
|
||||||
@ -91,11 +110,15 @@ func NewConstMetric(desc *Desc, valueType ValueType, value float64, labelValues
|
|||||||
if err := validateLabelValues(labelValues, len(desc.variableLabels)); err != nil {
|
if err := validateLabelValues(labelValues, len(desc.variableLabels)); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
metric := &dto.Metric{}
|
||||||
|
if err := populateMetric(valueType, value, MakeLabelPairs(desc, labelValues), nil, metric); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
return &constMetric{
|
return &constMetric{
|
||||||
desc: desc,
|
desc: desc,
|
||||||
valType: valueType,
|
metric: metric,
|
||||||
val: value,
|
|
||||||
labelPairs: MakeLabelPairs(desc, labelValues),
|
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -110,10 +133,8 @@ func MustNewConstMetric(desc *Desc, valueType ValueType, value float64, labelVal
|
|||||||
}
|
}
|
||||||
|
|
||||||
type constMetric struct {
|
type constMetric struct {
|
||||||
desc *Desc
|
desc *Desc
|
||||||
valType ValueType
|
metric *dto.Metric
|
||||||
val float64
|
|
||||||
labelPairs []*dto.LabelPair
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *constMetric) Desc() *Desc {
|
func (m *constMetric) Desc() *Desc {
|
||||||
@ -121,7 +142,11 @@ func (m *constMetric) Desc() *Desc {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (m *constMetric) Write(out *dto.Metric) error {
|
func (m *constMetric) Write(out *dto.Metric) error {
|
||||||
return populateMetric(m.valType, m.val, m.labelPairs, nil, out)
|
out.Label = m.metric.Label
|
||||||
|
out.Counter = m.metric.Counter
|
||||||
|
out.Gauge = m.metric.Gauge
|
||||||
|
out.Untyped = m.metric.Untyped
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func populateMetric(
|
func populateMetric(
|
||||||
@ -170,12 +195,12 @@ func MakeLabelPairs(desc *Desc, labelValues []string) []*dto.LabelPair {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
labelPairs = append(labelPairs, desc.constLabelPairs...)
|
labelPairs = append(labelPairs, desc.constLabelPairs...)
|
||||||
sort.Sort(labelPairSorter(labelPairs))
|
sort.Sort(internal.LabelPairSorter(labelPairs))
|
||||||
return labelPairs
|
return labelPairs
|
||||||
}
|
}
|
||||||
|
|
||||||
// ExemplarMaxRunes is the max total number of runes allowed in exemplar labels.
|
// ExemplarMaxRunes is the max total number of runes allowed in exemplar labels.
|
||||||
const ExemplarMaxRunes = 64
|
const ExemplarMaxRunes = 128
|
||||||
|
|
||||||
// newExemplar creates a new dto.Exemplar from the provided values. An error is
|
// newExemplar creates a new dto.Exemplar from the provided values. An error is
|
||||||
// returned if any of the label names or values are invalid or if the total
|
// returned if any of the label names or values are invalid or if the total
|
||||||
|
88
vendor/github.com/prometheus/client_golang/prometheus/vec.go
generated
vendored
88
vendor/github.com/prometheus/client_golang/prometheus/vec.go
generated
vendored
@ -99,6 +99,16 @@ func (m *MetricVec) Delete(labels Labels) bool {
|
|||||||
return m.metricMap.deleteByHashWithLabels(h, labels, m.curry)
|
return m.metricMap.deleteByHashWithLabels(h, labels, m.curry)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// DeletePartialMatch deletes all metrics where the variable labels contain all of those
|
||||||
|
// passed in as labels. The order of the labels does not matter.
|
||||||
|
// It returns the number of metrics deleted.
|
||||||
|
//
|
||||||
|
// Note that curried labels will never be matched if deleting from the curried vector.
|
||||||
|
// To match curried labels with DeletePartialMatch, it must be called on the base vector.
|
||||||
|
func (m *MetricVec) DeletePartialMatch(labels Labels) int {
|
||||||
|
return m.metricMap.deleteByLabels(labels, m.curry)
|
||||||
|
}
|
||||||
|
|
||||||
// Without explicit forwarding of Describe, Collect, Reset, those methods won't
|
// Without explicit forwarding of Describe, Collect, Reset, those methods won't
|
||||||
// show up in GoDoc.
|
// show up in GoDoc.
|
||||||
|
|
||||||
@ -381,6 +391,82 @@ func (m *metricMap) deleteByHashWithLabels(
|
|||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// deleteByLabels deletes a metric if the given labels are present in the metric.
|
||||||
|
func (m *metricMap) deleteByLabels(labels Labels, curry []curriedLabelValue) int {
|
||||||
|
m.mtx.Lock()
|
||||||
|
defer m.mtx.Unlock()
|
||||||
|
|
||||||
|
var numDeleted int
|
||||||
|
|
||||||
|
for h, metrics := range m.metrics {
|
||||||
|
i := findMetricWithPartialLabels(m.desc, metrics, labels, curry)
|
||||||
|
if i >= len(metrics) {
|
||||||
|
// Didn't find matching labels in this metric slice.
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
delete(m.metrics, h)
|
||||||
|
numDeleted++
|
||||||
|
}
|
||||||
|
|
||||||
|
return numDeleted
|
||||||
|
}
|
||||||
|
|
||||||
|
// findMetricWithPartialLabel returns the index of the matching metric or
|
||||||
|
// len(metrics) if not found.
|
||||||
|
func findMetricWithPartialLabels(
|
||||||
|
desc *Desc, metrics []metricWithLabelValues, labels Labels, curry []curriedLabelValue,
|
||||||
|
) int {
|
||||||
|
for i, metric := range metrics {
|
||||||
|
if matchPartialLabels(desc, metric.values, labels, curry) {
|
||||||
|
return i
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return len(metrics)
|
||||||
|
}
|
||||||
|
|
||||||
|
// indexOf searches the given slice of strings for the target string and returns
|
||||||
|
// the index or len(items) as well as a boolean whether the search succeeded.
|
||||||
|
func indexOf(target string, items []string) (int, bool) {
|
||||||
|
for i, l := range items {
|
||||||
|
if l == target {
|
||||||
|
return i, true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return len(items), false
|
||||||
|
}
|
||||||
|
|
||||||
|
// valueMatchesVariableOrCurriedValue determines if a value was previously curried,
|
||||||
|
// and returns whether it matches either the "base" value or the curried value accordingly.
|
||||||
|
// It also indicates whether the match is against a curried or uncurried value.
|
||||||
|
func valueMatchesVariableOrCurriedValue(targetValue string, index int, values []string, curry []curriedLabelValue) (bool, bool) {
|
||||||
|
for _, curriedValue := range curry {
|
||||||
|
if curriedValue.index == index {
|
||||||
|
// This label was curried. See if the curried value matches our target.
|
||||||
|
return curriedValue.value == targetValue, true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// This label was not curried. See if the current value matches our target label.
|
||||||
|
return values[index] == targetValue, false
|
||||||
|
}
|
||||||
|
|
||||||
|
// matchPartialLabels searches the current metric and returns whether all of the target label:value pairs are present.
|
||||||
|
func matchPartialLabels(desc *Desc, values []string, labels Labels, curry []curriedLabelValue) bool {
|
||||||
|
for l, v := range labels {
|
||||||
|
// Check if the target label exists in our metrics and get the index.
|
||||||
|
varLabelIndex, validLabel := indexOf(l, desc.variableLabels)
|
||||||
|
if validLabel {
|
||||||
|
// Check the value of that label against the target value.
|
||||||
|
// We don't consider curried values in partial matches.
|
||||||
|
matches, curried := valueMatchesVariableOrCurriedValue(v, varLabelIndex, values, curry)
|
||||||
|
if matches && !curried {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
// getOrCreateMetricWithLabelValues retrieves the metric by hash and label value
|
// getOrCreateMetricWithLabelValues retrieves the metric by hash and label value
|
||||||
// or creates it and returns the new one.
|
// or creates it and returns the new one.
|
||||||
//
|
//
|
||||||
@ -485,7 +571,7 @@ func findMetricWithLabels(
|
|||||||
return len(metrics)
|
return len(metrics)
|
||||||
}
|
}
|
||||||
|
|
||||||
func matchLabelValues(values []string, lvs []string, curry []curriedLabelValue) bool {
|
func matchLabelValues(values, lvs []string, curry []curriedLabelValue) bool {
|
||||||
if len(values) != len(lvs)+len(curry) {
|
if len(values) != len(lvs)+len(curry) {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
4
vendor/github.com/prometheus/client_golang/prometheus/wrap.go
generated
vendored
4
vendor/github.com/prometheus/client_golang/prometheus/wrap.go
generated
vendored
@ -21,6 +21,8 @@ import (
|
|||||||
"github.com/golang/protobuf/proto"
|
"github.com/golang/protobuf/proto"
|
||||||
|
|
||||||
dto "github.com/prometheus/client_model/go"
|
dto "github.com/prometheus/client_model/go"
|
||||||
|
|
||||||
|
"github.com/prometheus/client_golang/prometheus/internal"
|
||||||
)
|
)
|
||||||
|
|
||||||
// WrapRegistererWith returns a Registerer wrapping the provided
|
// WrapRegistererWith returns a Registerer wrapping the provided
|
||||||
@ -182,7 +184,7 @@ func (m *wrappingMetric) Write(out *dto.Metric) error {
|
|||||||
Value: proto.String(lv),
|
Value: proto.String(lv),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
sort.Sort(labelPairSorter(out.Label))
|
sort.Sort(internal.LabelPairSorter(out.Label))
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
333
vendor/github.com/prometheus/client_model/go/metrics.pb.go
generated
vendored
333
vendor/github.com/prometheus/client_model/go/metrics.pb.go
generated
vendored
@ -1,5 +1,5 @@
|
|||||||
// Code generated by protoc-gen-go. DO NOT EDIT.
|
// Code generated by protoc-gen-go. DO NOT EDIT.
|
||||||
// source: metrics.proto
|
// source: io/prometheus/client/metrics.proto
|
||||||
|
|
||||||
package io_prometheus_client
|
package io_prometheus_client
|
||||||
|
|
||||||
@ -24,11 +24,18 @@ const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package
|
|||||||
type MetricType int32
|
type MetricType int32
|
||||||
|
|
||||||
const (
|
const (
|
||||||
MetricType_COUNTER MetricType = 0
|
// COUNTER must use the Metric field "counter".
|
||||||
MetricType_GAUGE MetricType = 1
|
MetricType_COUNTER MetricType = 0
|
||||||
MetricType_SUMMARY MetricType = 2
|
// GAUGE must use the Metric field "gauge".
|
||||||
MetricType_UNTYPED MetricType = 3
|
MetricType_GAUGE MetricType = 1
|
||||||
|
// SUMMARY must use the Metric field "summary".
|
||||||
|
MetricType_SUMMARY MetricType = 2
|
||||||
|
// UNTYPED must use the Metric field "untyped".
|
||||||
|
MetricType_UNTYPED MetricType = 3
|
||||||
|
// HISTOGRAM must use the Metric field "histogram".
|
||||||
MetricType_HISTOGRAM MetricType = 4
|
MetricType_HISTOGRAM MetricType = 4
|
||||||
|
// GAUGE_HISTOGRAM must use the Metric field "histogram".
|
||||||
|
MetricType_GAUGE_HISTOGRAM MetricType = 5
|
||||||
)
|
)
|
||||||
|
|
||||||
var MetricType_name = map[int32]string{
|
var MetricType_name = map[int32]string{
|
||||||
@ -37,14 +44,16 @@ var MetricType_name = map[int32]string{
|
|||||||
2: "SUMMARY",
|
2: "SUMMARY",
|
||||||
3: "UNTYPED",
|
3: "UNTYPED",
|
||||||
4: "HISTOGRAM",
|
4: "HISTOGRAM",
|
||||||
|
5: "GAUGE_HISTOGRAM",
|
||||||
}
|
}
|
||||||
|
|
||||||
var MetricType_value = map[string]int32{
|
var MetricType_value = map[string]int32{
|
||||||
"COUNTER": 0,
|
"COUNTER": 0,
|
||||||
"GAUGE": 1,
|
"GAUGE": 1,
|
||||||
"SUMMARY": 2,
|
"SUMMARY": 2,
|
||||||
"UNTYPED": 3,
|
"UNTYPED": 3,
|
||||||
"HISTOGRAM": 4,
|
"HISTOGRAM": 4,
|
||||||
|
"GAUGE_HISTOGRAM": 5,
|
||||||
}
|
}
|
||||||
|
|
||||||
func (x MetricType) Enum() *MetricType {
|
func (x MetricType) Enum() *MetricType {
|
||||||
@ -67,7 +76,7 @@ func (x *MetricType) UnmarshalJSON(data []byte) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (MetricType) EnumDescriptor() ([]byte, []int) {
|
func (MetricType) EnumDescriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{0}
|
return fileDescriptor_d1e5ddb18987a258, []int{0}
|
||||||
}
|
}
|
||||||
|
|
||||||
type LabelPair struct {
|
type LabelPair struct {
|
||||||
@ -82,7 +91,7 @@ func (m *LabelPair) Reset() { *m = LabelPair{} }
|
|||||||
func (m *LabelPair) String() string { return proto.CompactTextString(m) }
|
func (m *LabelPair) String() string { return proto.CompactTextString(m) }
|
||||||
func (*LabelPair) ProtoMessage() {}
|
func (*LabelPair) ProtoMessage() {}
|
||||||
func (*LabelPair) Descriptor() ([]byte, []int) {
|
func (*LabelPair) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{0}
|
return fileDescriptor_d1e5ddb18987a258, []int{0}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *LabelPair) XXX_Unmarshal(b []byte) error {
|
func (m *LabelPair) XXX_Unmarshal(b []byte) error {
|
||||||
@ -128,7 +137,7 @@ func (m *Gauge) Reset() { *m = Gauge{} }
|
|||||||
func (m *Gauge) String() string { return proto.CompactTextString(m) }
|
func (m *Gauge) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Gauge) ProtoMessage() {}
|
func (*Gauge) ProtoMessage() {}
|
||||||
func (*Gauge) Descriptor() ([]byte, []int) {
|
func (*Gauge) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{1}
|
return fileDescriptor_d1e5ddb18987a258, []int{1}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Gauge) XXX_Unmarshal(b []byte) error {
|
func (m *Gauge) XXX_Unmarshal(b []byte) error {
|
||||||
@ -168,7 +177,7 @@ func (m *Counter) Reset() { *m = Counter{} }
|
|||||||
func (m *Counter) String() string { return proto.CompactTextString(m) }
|
func (m *Counter) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Counter) ProtoMessage() {}
|
func (*Counter) ProtoMessage() {}
|
||||||
func (*Counter) Descriptor() ([]byte, []int) {
|
func (*Counter) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{2}
|
return fileDescriptor_d1e5ddb18987a258, []int{2}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Counter) XXX_Unmarshal(b []byte) error {
|
func (m *Counter) XXX_Unmarshal(b []byte) error {
|
||||||
@ -215,7 +224,7 @@ func (m *Quantile) Reset() { *m = Quantile{} }
|
|||||||
func (m *Quantile) String() string { return proto.CompactTextString(m) }
|
func (m *Quantile) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Quantile) ProtoMessage() {}
|
func (*Quantile) ProtoMessage() {}
|
||||||
func (*Quantile) Descriptor() ([]byte, []int) {
|
func (*Quantile) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{3}
|
return fileDescriptor_d1e5ddb18987a258, []int{3}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Quantile) XXX_Unmarshal(b []byte) error {
|
func (m *Quantile) XXX_Unmarshal(b []byte) error {
|
||||||
@ -263,7 +272,7 @@ func (m *Summary) Reset() { *m = Summary{} }
|
|||||||
func (m *Summary) String() string { return proto.CompactTextString(m) }
|
func (m *Summary) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Summary) ProtoMessage() {}
|
func (*Summary) ProtoMessage() {}
|
||||||
func (*Summary) Descriptor() ([]byte, []int) {
|
func (*Summary) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{4}
|
return fileDescriptor_d1e5ddb18987a258, []int{4}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Summary) XXX_Unmarshal(b []byte) error {
|
func (m *Summary) XXX_Unmarshal(b []byte) error {
|
||||||
@ -316,7 +325,7 @@ func (m *Untyped) Reset() { *m = Untyped{} }
|
|||||||
func (m *Untyped) String() string { return proto.CompactTextString(m) }
|
func (m *Untyped) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Untyped) ProtoMessage() {}
|
func (*Untyped) ProtoMessage() {}
|
||||||
func (*Untyped) Descriptor() ([]byte, []int) {
|
func (*Untyped) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{5}
|
return fileDescriptor_d1e5ddb18987a258, []int{5}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Untyped) XXX_Unmarshal(b []byte) error {
|
func (m *Untyped) XXX_Unmarshal(b []byte) error {
|
||||||
@ -345,9 +354,34 @@ func (m *Untyped) GetValue() float64 {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type Histogram struct {
|
type Histogram struct {
|
||||||
SampleCount *uint64 `protobuf:"varint,1,opt,name=sample_count,json=sampleCount" json:"sample_count,omitempty"`
|
SampleCount *uint64 `protobuf:"varint,1,opt,name=sample_count,json=sampleCount" json:"sample_count,omitempty"`
|
||||||
SampleSum *float64 `protobuf:"fixed64,2,opt,name=sample_sum,json=sampleSum" json:"sample_sum,omitempty"`
|
SampleCountFloat *float64 `protobuf:"fixed64,4,opt,name=sample_count_float,json=sampleCountFloat" json:"sample_count_float,omitempty"`
|
||||||
Bucket []*Bucket `protobuf:"bytes,3,rep,name=bucket" json:"bucket,omitempty"`
|
SampleSum *float64 `protobuf:"fixed64,2,opt,name=sample_sum,json=sampleSum" json:"sample_sum,omitempty"`
|
||||||
|
// Buckets for the conventional histogram.
|
||||||
|
Bucket []*Bucket `protobuf:"bytes,3,rep,name=bucket" json:"bucket,omitempty"`
|
||||||
|
// schema defines the bucket schema. Currently, valid numbers are -4 <= n <= 8.
|
||||||
|
// They are all for base-2 bucket schemas, where 1 is a bucket boundary in each case, and
|
||||||
|
// then each power of two is divided into 2^n logarithmic buckets.
|
||||||
|
// Or in other words, each bucket boundary is the previous boundary times 2^(2^-n).
|
||||||
|
// In the future, more bucket schemas may be added using numbers < -4 or > 8.
|
||||||
|
Schema *int32 `protobuf:"zigzag32,5,opt,name=schema" json:"schema,omitempty"`
|
||||||
|
ZeroThreshold *float64 `protobuf:"fixed64,6,opt,name=zero_threshold,json=zeroThreshold" json:"zero_threshold,omitempty"`
|
||||||
|
ZeroCount *uint64 `protobuf:"varint,7,opt,name=zero_count,json=zeroCount" json:"zero_count,omitempty"`
|
||||||
|
ZeroCountFloat *float64 `protobuf:"fixed64,8,opt,name=zero_count_float,json=zeroCountFloat" json:"zero_count_float,omitempty"`
|
||||||
|
// Negative buckets for the native histogram.
|
||||||
|
NegativeSpan []*BucketSpan `protobuf:"bytes,9,rep,name=negative_span,json=negativeSpan" json:"negative_span,omitempty"`
|
||||||
|
// Use either "negative_delta" or "negative_count", the former for
|
||||||
|
// regular histograms with integer counts, the latter for float
|
||||||
|
// histograms.
|
||||||
|
NegativeDelta []int64 `protobuf:"zigzag64,10,rep,name=negative_delta,json=negativeDelta" json:"negative_delta,omitempty"`
|
||||||
|
NegativeCount []float64 `protobuf:"fixed64,11,rep,name=negative_count,json=negativeCount" json:"negative_count,omitempty"`
|
||||||
|
// Positive buckets for the native histogram.
|
||||||
|
PositiveSpan []*BucketSpan `protobuf:"bytes,12,rep,name=positive_span,json=positiveSpan" json:"positive_span,omitempty"`
|
||||||
|
// Use either "positive_delta" or "positive_count", the former for
|
||||||
|
// regular histograms with integer counts, the latter for float
|
||||||
|
// histograms.
|
||||||
|
PositiveDelta []int64 `protobuf:"zigzag64,13,rep,name=positive_delta,json=positiveDelta" json:"positive_delta,omitempty"`
|
||||||
|
PositiveCount []float64 `protobuf:"fixed64,14,rep,name=positive_count,json=positiveCount" json:"positive_count,omitempty"`
|
||||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||||
XXX_unrecognized []byte `json:"-"`
|
XXX_unrecognized []byte `json:"-"`
|
||||||
XXX_sizecache int32 `json:"-"`
|
XXX_sizecache int32 `json:"-"`
|
||||||
@ -357,7 +391,7 @@ func (m *Histogram) Reset() { *m = Histogram{} }
|
|||||||
func (m *Histogram) String() string { return proto.CompactTextString(m) }
|
func (m *Histogram) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Histogram) ProtoMessage() {}
|
func (*Histogram) ProtoMessage() {}
|
||||||
func (*Histogram) Descriptor() ([]byte, []int) {
|
func (*Histogram) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{6}
|
return fileDescriptor_d1e5ddb18987a258, []int{6}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Histogram) XXX_Unmarshal(b []byte) error {
|
func (m *Histogram) XXX_Unmarshal(b []byte) error {
|
||||||
@ -385,6 +419,13 @@ func (m *Histogram) GetSampleCount() uint64 {
|
|||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetSampleCountFloat() float64 {
|
||||||
|
if m != nil && m.SampleCountFloat != nil {
|
||||||
|
return *m.SampleCountFloat
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
func (m *Histogram) GetSampleSum() float64 {
|
func (m *Histogram) GetSampleSum() float64 {
|
||||||
if m != nil && m.SampleSum != nil {
|
if m != nil && m.SampleSum != nil {
|
||||||
return *m.SampleSum
|
return *m.SampleSum
|
||||||
@ -399,8 +440,81 @@ func (m *Histogram) GetBucket() []*Bucket {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetSchema() int32 {
|
||||||
|
if m != nil && m.Schema != nil {
|
||||||
|
return *m.Schema
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetZeroThreshold() float64 {
|
||||||
|
if m != nil && m.ZeroThreshold != nil {
|
||||||
|
return *m.ZeroThreshold
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetZeroCount() uint64 {
|
||||||
|
if m != nil && m.ZeroCount != nil {
|
||||||
|
return *m.ZeroCount
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetZeroCountFloat() float64 {
|
||||||
|
if m != nil && m.ZeroCountFloat != nil {
|
||||||
|
return *m.ZeroCountFloat
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetNegativeSpan() []*BucketSpan {
|
||||||
|
if m != nil {
|
||||||
|
return m.NegativeSpan
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetNegativeDelta() []int64 {
|
||||||
|
if m != nil {
|
||||||
|
return m.NegativeDelta
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetNegativeCount() []float64 {
|
||||||
|
if m != nil {
|
||||||
|
return m.NegativeCount
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetPositiveSpan() []*BucketSpan {
|
||||||
|
if m != nil {
|
||||||
|
return m.PositiveSpan
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetPositiveDelta() []int64 {
|
||||||
|
if m != nil {
|
||||||
|
return m.PositiveDelta
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetPositiveCount() []float64 {
|
||||||
|
if m != nil {
|
||||||
|
return m.PositiveCount
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// A Bucket of a conventional histogram, each of which is treated as
|
||||||
|
// an individual counter-like time series by Prometheus.
|
||||||
type Bucket struct {
|
type Bucket struct {
|
||||||
CumulativeCount *uint64 `protobuf:"varint,1,opt,name=cumulative_count,json=cumulativeCount" json:"cumulative_count,omitempty"`
|
CumulativeCount *uint64 `protobuf:"varint,1,opt,name=cumulative_count,json=cumulativeCount" json:"cumulative_count,omitempty"`
|
||||||
|
CumulativeCountFloat *float64 `protobuf:"fixed64,4,opt,name=cumulative_count_float,json=cumulativeCountFloat" json:"cumulative_count_float,omitempty"`
|
||||||
UpperBound *float64 `protobuf:"fixed64,2,opt,name=upper_bound,json=upperBound" json:"upper_bound,omitempty"`
|
UpperBound *float64 `protobuf:"fixed64,2,opt,name=upper_bound,json=upperBound" json:"upper_bound,omitempty"`
|
||||||
Exemplar *Exemplar `protobuf:"bytes,3,opt,name=exemplar" json:"exemplar,omitempty"`
|
Exemplar *Exemplar `protobuf:"bytes,3,opt,name=exemplar" json:"exemplar,omitempty"`
|
||||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||||
@ -412,7 +526,7 @@ func (m *Bucket) Reset() { *m = Bucket{} }
|
|||||||
func (m *Bucket) String() string { return proto.CompactTextString(m) }
|
func (m *Bucket) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Bucket) ProtoMessage() {}
|
func (*Bucket) ProtoMessage() {}
|
||||||
func (*Bucket) Descriptor() ([]byte, []int) {
|
func (*Bucket) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{7}
|
return fileDescriptor_d1e5ddb18987a258, []int{7}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Bucket) XXX_Unmarshal(b []byte) error {
|
func (m *Bucket) XXX_Unmarshal(b []byte) error {
|
||||||
@ -440,6 +554,13 @@ func (m *Bucket) GetCumulativeCount() uint64 {
|
|||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (m *Bucket) GetCumulativeCountFloat() float64 {
|
||||||
|
if m != nil && m.CumulativeCountFloat != nil {
|
||||||
|
return *m.CumulativeCountFloat
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
func (m *Bucket) GetUpperBound() float64 {
|
func (m *Bucket) GetUpperBound() float64 {
|
||||||
if m != nil && m.UpperBound != nil {
|
if m != nil && m.UpperBound != nil {
|
||||||
return *m.UpperBound
|
return *m.UpperBound
|
||||||
@ -454,6 +575,59 @@ func (m *Bucket) GetExemplar() *Exemplar {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// A BucketSpan defines a number of consecutive buckets in a native
|
||||||
|
// histogram with their offset. Logically, it would be more
|
||||||
|
// straightforward to include the bucket counts in the Span. However,
|
||||||
|
// the protobuf representation is more compact in the way the data is
|
||||||
|
// structured here (with all the buckets in a single array separate
|
||||||
|
// from the Spans).
|
||||||
|
type BucketSpan struct {
|
||||||
|
Offset *int32 `protobuf:"zigzag32,1,opt,name=offset" json:"offset,omitempty"`
|
||||||
|
Length *uint32 `protobuf:"varint,2,opt,name=length" json:"length,omitempty"`
|
||||||
|
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||||
|
XXX_unrecognized []byte `json:"-"`
|
||||||
|
XXX_sizecache int32 `json:"-"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *BucketSpan) Reset() { *m = BucketSpan{} }
|
||||||
|
func (m *BucketSpan) String() string { return proto.CompactTextString(m) }
|
||||||
|
func (*BucketSpan) ProtoMessage() {}
|
||||||
|
func (*BucketSpan) Descriptor() ([]byte, []int) {
|
||||||
|
return fileDescriptor_d1e5ddb18987a258, []int{8}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *BucketSpan) XXX_Unmarshal(b []byte) error {
|
||||||
|
return xxx_messageInfo_BucketSpan.Unmarshal(m, b)
|
||||||
|
}
|
||||||
|
func (m *BucketSpan) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||||
|
return xxx_messageInfo_BucketSpan.Marshal(b, m, deterministic)
|
||||||
|
}
|
||||||
|
func (m *BucketSpan) XXX_Merge(src proto.Message) {
|
||||||
|
xxx_messageInfo_BucketSpan.Merge(m, src)
|
||||||
|
}
|
||||||
|
func (m *BucketSpan) XXX_Size() int {
|
||||||
|
return xxx_messageInfo_BucketSpan.Size(m)
|
||||||
|
}
|
||||||
|
func (m *BucketSpan) XXX_DiscardUnknown() {
|
||||||
|
xxx_messageInfo_BucketSpan.DiscardUnknown(m)
|
||||||
|
}
|
||||||
|
|
||||||
|
var xxx_messageInfo_BucketSpan proto.InternalMessageInfo
|
||||||
|
|
||||||
|
func (m *BucketSpan) GetOffset() int32 {
|
||||||
|
if m != nil && m.Offset != nil {
|
||||||
|
return *m.Offset
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *BucketSpan) GetLength() uint32 {
|
||||||
|
if m != nil && m.Length != nil {
|
||||||
|
return *m.Length
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
type Exemplar struct {
|
type Exemplar struct {
|
||||||
Label []*LabelPair `protobuf:"bytes,1,rep,name=label" json:"label,omitempty"`
|
Label []*LabelPair `protobuf:"bytes,1,rep,name=label" json:"label,omitempty"`
|
||||||
Value *float64 `protobuf:"fixed64,2,opt,name=value" json:"value,omitempty"`
|
Value *float64 `protobuf:"fixed64,2,opt,name=value" json:"value,omitempty"`
|
||||||
@ -467,7 +641,7 @@ func (m *Exemplar) Reset() { *m = Exemplar{} }
|
|||||||
func (m *Exemplar) String() string { return proto.CompactTextString(m) }
|
func (m *Exemplar) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Exemplar) ProtoMessage() {}
|
func (*Exemplar) ProtoMessage() {}
|
||||||
func (*Exemplar) Descriptor() ([]byte, []int) {
|
func (*Exemplar) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{8}
|
return fileDescriptor_d1e5ddb18987a258, []int{9}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Exemplar) XXX_Unmarshal(b []byte) error {
|
func (m *Exemplar) XXX_Unmarshal(b []byte) error {
|
||||||
@ -526,7 +700,7 @@ func (m *Metric) Reset() { *m = Metric{} }
|
|||||||
func (m *Metric) String() string { return proto.CompactTextString(m) }
|
func (m *Metric) String() string { return proto.CompactTextString(m) }
|
||||||
func (*Metric) ProtoMessage() {}
|
func (*Metric) ProtoMessage() {}
|
||||||
func (*Metric) Descriptor() ([]byte, []int) {
|
func (*Metric) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{9}
|
return fileDescriptor_d1e5ddb18987a258, []int{10}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Metric) XXX_Unmarshal(b []byte) error {
|
func (m *Metric) XXX_Unmarshal(b []byte) error {
|
||||||
@ -610,7 +784,7 @@ func (m *MetricFamily) Reset() { *m = MetricFamily{} }
|
|||||||
func (m *MetricFamily) String() string { return proto.CompactTextString(m) }
|
func (m *MetricFamily) String() string { return proto.CompactTextString(m) }
|
||||||
func (*MetricFamily) ProtoMessage() {}
|
func (*MetricFamily) ProtoMessage() {}
|
||||||
func (*MetricFamily) Descriptor() ([]byte, []int) {
|
func (*MetricFamily) Descriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_6039342a2ba47b72, []int{10}
|
return fileDescriptor_d1e5ddb18987a258, []int{11}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *MetricFamily) XXX_Unmarshal(b []byte) error {
|
func (m *MetricFamily) XXX_Unmarshal(b []byte) error {
|
||||||
@ -669,55 +843,72 @@ func init() {
|
|||||||
proto.RegisterType((*Untyped)(nil), "io.prometheus.client.Untyped")
|
proto.RegisterType((*Untyped)(nil), "io.prometheus.client.Untyped")
|
||||||
proto.RegisterType((*Histogram)(nil), "io.prometheus.client.Histogram")
|
proto.RegisterType((*Histogram)(nil), "io.prometheus.client.Histogram")
|
||||||
proto.RegisterType((*Bucket)(nil), "io.prometheus.client.Bucket")
|
proto.RegisterType((*Bucket)(nil), "io.prometheus.client.Bucket")
|
||||||
|
proto.RegisterType((*BucketSpan)(nil), "io.prometheus.client.BucketSpan")
|
||||||
proto.RegisterType((*Exemplar)(nil), "io.prometheus.client.Exemplar")
|
proto.RegisterType((*Exemplar)(nil), "io.prometheus.client.Exemplar")
|
||||||
proto.RegisterType((*Metric)(nil), "io.prometheus.client.Metric")
|
proto.RegisterType((*Metric)(nil), "io.prometheus.client.Metric")
|
||||||
proto.RegisterType((*MetricFamily)(nil), "io.prometheus.client.MetricFamily")
|
proto.RegisterType((*MetricFamily)(nil), "io.prometheus.client.MetricFamily")
|
||||||
}
|
}
|
||||||
|
|
||||||
func init() { proto.RegisterFile("metrics.proto", fileDescriptor_6039342a2ba47b72) }
|
func init() {
|
||||||
|
proto.RegisterFile("io/prometheus/client/metrics.proto", fileDescriptor_d1e5ddb18987a258)
|
||||||
var fileDescriptor_6039342a2ba47b72 = []byte{
|
}
|
||||||
// 665 bytes of a gzipped FileDescriptorProto
|
|
||||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x54, 0xcd, 0x6e, 0xd3, 0x4c,
|
var fileDescriptor_d1e5ddb18987a258 = []byte{
|
||||||
0x14, 0xfd, 0xdc, 0x38, 0x3f, 0xbe, 0x69, 0x3f, 0xa2, 0x51, 0x17, 0x56, 0xa1, 0x24, 0x78, 0x55,
|
// 896 bytes of a gzipped FileDescriptorProto
|
||||||
0x58, 0x38, 0xa2, 0x6a, 0x05, 0x2a, 0xb0, 0x68, 0x4b, 0x48, 0x91, 0x48, 0x5b, 0x26, 0xc9, 0xa2,
|
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x56, 0xdd, 0x8e, 0xdb, 0x44,
|
||||||
0xb0, 0x88, 0x1c, 0x77, 0x70, 0x2c, 0x3c, 0xb1, 0xb1, 0x67, 0x2a, 0xb2, 0x66, 0xc1, 0x16, 0x5e,
|
0x18, 0xc5, 0x9b, 0x5f, 0x7f, 0xd9, 0x6c, 0xd3, 0x61, 0x55, 0x59, 0x0b, 0xcb, 0x06, 0x4b, 0x48,
|
||||||
0x81, 0x17, 0x05, 0xcd, 0x8f, 0x6d, 0x2a, 0xb9, 0x95, 0x40, 0xec, 0x66, 0xee, 0x3d, 0xe7, 0xfa,
|
0x0b, 0x42, 0x8e, 0x40, 0x5b, 0x81, 0x0a, 0x5c, 0xec, 0xb6, 0xe9, 0x16, 0x89, 0xb4, 0x65, 0x92,
|
||||||
0xcc, 0xf8, 0x9c, 0x81, 0x0d, 0x4a, 0x58, 0x1a, 0xfa, 0x99, 0x9b, 0xa4, 0x31, 0x8b, 0xd1, 0x66,
|
0x5c, 0x14, 0x2e, 0xac, 0x49, 0x32, 0xeb, 0x58, 0x78, 0x3c, 0xc6, 0x1e, 0x57, 0x2c, 0x2f, 0xc0,
|
||||||
0x18, 0x8b, 0x15, 0x25, 0x6c, 0x41, 0x78, 0xe6, 0xfa, 0x51, 0x48, 0x96, 0x6c, 0xab, 0x1b, 0xc4,
|
0x35, 0xaf, 0xc0, 0xc3, 0xf0, 0x22, 0x3c, 0x08, 0x68, 0xfe, 0xec, 0xdd, 0xe2, 0x94, 0xd2, 0x3b,
|
||||||
0x71, 0x10, 0x91, 0xbe, 0xc4, 0xcc, 0xf9, 0x87, 0x3e, 0x0b, 0x29, 0xc9, 0x98, 0x47, 0x13, 0x45,
|
0x7f, 0x67, 0xce, 0xf7, 0xcd, 0x39, 0xe3, 0xc9, 0x71, 0xc0, 0x8f, 0xf9, 0x24, 0xcb, 0x39, 0xa3,
|
||||||
0x73, 0xf6, 0xc1, 0x7a, 0xe3, 0xcd, 0x49, 0x74, 0xee, 0x85, 0x29, 0x42, 0x60, 0x2e, 0x3d, 0x4a,
|
0x62, 0x4b, 0xcb, 0x62, 0xb2, 0x4e, 0x62, 0x9a, 0x8a, 0x09, 0xa3, 0x22, 0x8f, 0xd7, 0x45, 0x90,
|
||||||
0x6c, 0xa3, 0x67, 0xec, 0x58, 0x58, 0xae, 0xd1, 0x26, 0xd4, 0xaf, 0xbc, 0x88, 0x13, 0x7b, 0x4d,
|
0xe5, 0x5c, 0x70, 0x74, 0x18, 0xf3, 0xa0, 0xe6, 0x04, 0x9a, 0x73, 0x74, 0x12, 0x71, 0x1e, 0x25,
|
||||||
0x16, 0xd5, 0xc6, 0xd9, 0x86, 0xfa, 0xd0, 0xe3, 0xc1, 0x6f, 0x6d, 0xc1, 0x31, 0xf2, 0xf6, 0x7b,
|
0x74, 0xa2, 0x38, 0xab, 0xf2, 0x6a, 0x22, 0x62, 0x46, 0x0b, 0x41, 0x58, 0xa6, 0xdb, 0xfc, 0xfb,
|
||||||
0x68, 0x1e, 0xc7, 0x7c, 0xc9, 0x48, 0x5a, 0x0d, 0x40, 0x07, 0xd0, 0x22, 0x9f, 0x09, 0x4d, 0x22,
|
0xe0, 0x7e, 0x47, 0x56, 0x34, 0x79, 0x4e, 0xe2, 0x1c, 0x21, 0x68, 0xa7, 0x84, 0x51, 0xcf, 0x19,
|
||||||
0x2f, 0x95, 0x83, 0xdb, 0xbb, 0xf7, 0xdd, 0xaa, 0x03, 0xb8, 0x03, 0x8d, 0xc2, 0x05, 0xde, 0x79,
|
0x3b, 0xa7, 0x2e, 0x56, 0xcf, 0xe8, 0x10, 0x3a, 0x2f, 0x49, 0x52, 0x52, 0x6f, 0x4f, 0x81, 0xba,
|
||||||
0x0e, 0xad, 0xb7, 0xdc, 0x5b, 0xb2, 0x30, 0x22, 0x68, 0x0b, 0x5a, 0x9f, 0xf4, 0x5a, 0x7f, 0xa0,
|
0xf0, 0x8f, 0xa1, 0x73, 0x49, 0xca, 0xe8, 0xc6, 0xb2, 0xec, 0x71, 0xec, 0xf2, 0x8f, 0xd0, 0x7b,
|
||||||
0xd8, 0x5f, 0x57, 0x5e, 0x48, 0xfb, 0x6a, 0x40, 0x73, 0xcc, 0x29, 0xf5, 0xd2, 0x15, 0x7a, 0x00,
|
0xc8, 0xcb, 0x54, 0xd0, 0xbc, 0x99, 0x80, 0x1e, 0x40, 0x9f, 0xfe, 0x42, 0x59, 0x96, 0x90, 0x5c,
|
||||||
0xeb, 0x99, 0x47, 0x93, 0x88, 0xcc, 0x7c, 0xa1, 0x56, 0x4e, 0x30, 0x71, 0x5b, 0xd5, 0xe4, 0x01,
|
0x0d, 0x1e, 0x7c, 0xfe, 0x41, 0xd0, 0x64, 0x20, 0x98, 0x1a, 0x16, 0xae, 0xf8, 0xfe, 0xd7, 0xd0,
|
||||||
0xd0, 0x36, 0x80, 0x86, 0x64, 0x9c, 0xea, 0x49, 0x96, 0xaa, 0x8c, 0x39, 0x15, 0xe7, 0x28, 0xbe,
|
0xff, 0xbe, 0x24, 0xa9, 0x88, 0x13, 0x8a, 0x8e, 0xa0, 0xff, 0xb3, 0x79, 0x36, 0x1b, 0x54, 0xf5,
|
||||||
0x5f, 0xeb, 0xd5, 0x6e, 0x3e, 0x47, 0xae, 0xb8, 0xd4, 0xe7, 0x74, 0xa1, 0x39, 0x5d, 0xb2, 0x55,
|
0x6d, 0xe5, 0x95, 0xb4, 0xdf, 0x1c, 0xe8, 0xcd, 0x4b, 0xc6, 0x48, 0x7e, 0x8d, 0x3e, 0x84, 0xfd,
|
||||||
0x42, 0x2e, 0x6f, 0xb8, 0xc5, 0x2f, 0x06, 0x58, 0x27, 0x61, 0xc6, 0xe2, 0x20, 0xf5, 0xe8, 0x3f,
|
0x82, 0xb0, 0x2c, 0xa1, 0xe1, 0x5a, 0xaa, 0x55, 0x13, 0xda, 0x78, 0xa0, 0x31, 0x65, 0x00, 0x1d,
|
||||||
0x10, 0xbb, 0x07, 0x8d, 0x39, 0xf7, 0x3f, 0x12, 0xa6, 0xa5, 0xde, 0xab, 0x96, 0x7a, 0x24, 0x31,
|
0x03, 0x18, 0x4a, 0x51, 0x32, 0x33, 0xc9, 0xd5, 0xc8, 0xbc, 0x64, 0xd2, 0x47, 0xb5, 0x7f, 0x6b,
|
||||||
0x58, 0x63, 0x9d, 0x6f, 0x06, 0x34, 0x54, 0x09, 0x3d, 0x84, 0x8e, 0xcf, 0x29, 0x8f, 0x3c, 0x16,
|
0xdc, 0xda, 0xed, 0xc3, 0x2a, 0xae, 0xf5, 0xf9, 0x27, 0xd0, 0x5b, 0xa6, 0xe2, 0x3a, 0xa3, 0x9b,
|
||||||
0x5e, 0x5d, 0x97, 0x71, 0xa7, 0xac, 0x2b, 0x29, 0x5d, 0x68, 0xf3, 0x24, 0x21, 0xe9, 0x6c, 0x1e,
|
0x1d, 0xa7, 0xf8, 0x57, 0x1b, 0xdc, 0x27, 0x71, 0x21, 0x78, 0x94, 0x13, 0xf6, 0x26, 0x62, 0x3f,
|
||||||
0xf3, 0xe5, 0xa5, 0xd6, 0x02, 0xb2, 0x74, 0x24, 0x2a, 0xd7, 0x1c, 0x50, 0xfb, 0x43, 0x07, 0x7c,
|
0x05, 0x74, 0x93, 0x12, 0x5e, 0x25, 0x9c, 0x08, 0xaf, 0xad, 0x66, 0x8e, 0x6e, 0x10, 0x1f, 0x4b,
|
||||||
0x37, 0xa0, 0x95, 0x97, 0xd1, 0x3e, 0xd4, 0x23, 0xe1, 0x60, 0xdb, 0x90, 0x87, 0xea, 0x56, 0x4f,
|
0xfc, 0xbf, 0xac, 0x9d, 0x41, 0x77, 0x55, 0xae, 0x7f, 0xa2, 0xc2, 0x18, 0x7b, 0xbf, 0xd9, 0xd8,
|
||||||
0x29, 0x4c, 0x8e, 0x15, 0xba, 0xda, 0x1d, 0xe8, 0x29, 0x58, 0x45, 0x42, 0xb4, 0xac, 0x2d, 0x57,
|
0x85, 0xe2, 0x60, 0xc3, 0x45, 0xf7, 0xa0, 0x5b, 0xac, 0xb7, 0x94, 0x11, 0xaf, 0x33, 0x76, 0x4e,
|
||||||
0x65, 0xc8, 0xcd, 0x33, 0xe4, 0x4e, 0x72, 0x04, 0x2e, 0xc1, 0xce, 0xcf, 0x35, 0x68, 0x8c, 0x64,
|
0xef, 0x62, 0x53, 0xa1, 0x8f, 0xe0, 0xe0, 0x57, 0x9a, 0xf3, 0x50, 0x6c, 0x73, 0x5a, 0x6c, 0x79,
|
||||||
0x22, 0xff, 0x56, 0xd1, 0x63, 0xa8, 0x07, 0x22, 0x53, 0x3a, 0x10, 0x77, 0xab, 0x69, 0x32, 0x76,
|
0xb2, 0xf1, 0xba, 0x6a, 0xc3, 0xa1, 0x44, 0x17, 0x16, 0x94, 0x9a, 0x14, 0x4d, 0x5b, 0xec, 0x29,
|
||||||
0x58, 0x21, 0xd1, 0x13, 0x68, 0xfa, 0x2a, 0x67, 0x5a, 0xec, 0x76, 0x35, 0x49, 0x87, 0x11, 0xe7,
|
0x8b, 0xae, 0x44, 0xb4, 0xc1, 0x53, 0x18, 0xd5, 0xcb, 0xc6, 0x5e, 0x5f, 0xcd, 0x39, 0xa8, 0x48,
|
||||||
0x68, 0x41, 0xcc, 0x54, 0x08, 0x6c, 0xf3, 0x36, 0xa2, 0x4e, 0x0a, 0xce, 0xd1, 0x82, 0xc8, 0x95,
|
0xda, 0xdc, 0x14, 0x86, 0x29, 0x8d, 0x88, 0x88, 0x5f, 0xd2, 0xb0, 0xc8, 0x48, 0xea, 0xb9, 0xca,
|
||||||
0x69, 0xed, 0xfa, 0x6d, 0x44, 0xed, 0x6c, 0x9c, 0xa3, 0xd1, 0x0b, 0xb0, 0x16, 0xb9, 0x97, 0xed,
|
0xc4, 0xf8, 0x75, 0x26, 0xe6, 0x19, 0x49, 0xf1, 0xbe, 0x6d, 0x93, 0x95, 0x94, 0x5d, 0x8d, 0xd9,
|
||||||
0xa6, 0xa4, 0xde, 0x70, 0x31, 0x85, 0xe5, 0x71, 0xc9, 0x10, 0xee, 0x2f, 0xee, 0x7a, 0x46, 0x33,
|
0xd0, 0x44, 0x10, 0x0f, 0xc6, 0xad, 0x53, 0x84, 0xab, 0xe1, 0x8f, 0x24, 0x78, 0x8b, 0xa6, 0xa5,
|
||||||
0xbb, 0xd1, 0x33, 0x76, 0x6a, 0xb8, 0x5d, 0xd4, 0x46, 0x99, 0xf3, 0xc3, 0x80, 0x75, 0xf5, 0x07,
|
0x0f, 0xc6, 0x2d, 0xe9, 0xce, 0xa2, 0x5a, 0xfe, 0x14, 0x86, 0x19, 0x2f, 0xe2, 0x5a, 0xd4, 0xfe,
|
||||||
0x5e, 0x79, 0x34, 0x8c, 0x56, 0x95, 0xcf, 0x19, 0x02, 0x73, 0x41, 0xa2, 0x44, 0xbf, 0x66, 0x72,
|
0x9b, 0x8a, 0xb2, 0x6d, 0x56, 0x54, 0x35, 0x46, 0x8b, 0x1a, 0x6a, 0x51, 0x16, 0xad, 0x44, 0x55,
|
||||||
0x8d, 0xf6, 0xc0, 0x14, 0x1a, 0xe5, 0x15, 0xfe, 0xbf, 0xdb, 0xab, 0x56, 0xa5, 0x26, 0x4f, 0x56,
|
0x34, 0x2d, 0xea, 0x40, 0x8b, 0xb2, 0xa8, 0x12, 0xe5, 0xff, 0xe9, 0x40, 0x57, 0x6f, 0x85, 0x3e,
|
||||||
0x09, 0xc1, 0x12, 0x2d, 0xd2, 0xa4, 0x5e, 0x60, 0xdb, 0xbc, 0x2d, 0x4d, 0x8a, 0x87, 0x35, 0xf6,
|
0x86, 0xd1, 0xba, 0x64, 0x65, 0x72, 0xd3, 0x88, 0xbe, 0x66, 0x77, 0x6a, 0x5c, 0x5b, 0x39, 0x83,
|
||||||
0xd1, 0x08, 0xa0, 0x9c, 0x84, 0xda, 0xd0, 0x3c, 0x3e, 0x9b, 0x9e, 0x4e, 0x06, 0xb8, 0xf3, 0x1f,
|
0x7b, 0xaf, 0x52, 0x6f, 0x5d, 0xb7, 0xc3, 0x57, 0x1a, 0xf4, 0x5b, 0x39, 0x81, 0x41, 0x99, 0x65,
|
||||||
0xb2, 0xa0, 0x3e, 0x3c, 0x9c, 0x0e, 0x07, 0x1d, 0x43, 0xd4, 0xc7, 0xd3, 0xd1, 0xe8, 0x10, 0x5f,
|
0x34, 0x0f, 0x57, 0xbc, 0x4c, 0x37, 0xe6, 0xce, 0x81, 0x82, 0x2e, 0x24, 0x72, 0x2b, 0x17, 0x5a,
|
||||||
0x74, 0xd6, 0xc4, 0x66, 0x7a, 0x3a, 0xb9, 0x38, 0x1f, 0xbc, 0xec, 0xd4, 0xd0, 0x06, 0x58, 0x27,
|
0xff, 0x3b, 0x17, 0xa0, 0x3e, 0x32, 0x79, 0x11, 0xf9, 0xd5, 0x55, 0x41, 0xb5, 0x83, 0xbb, 0xd8,
|
||||||
0xaf, 0xc7, 0x93, 0xb3, 0x21, 0x3e, 0x1c, 0x75, 0xcc, 0x23, 0x0c, 0x95, 0xef, 0xfe, 0xbb, 0x83,
|
0x54, 0x12, 0x4f, 0x68, 0x1a, 0x89, 0xad, 0xda, 0x7d, 0x88, 0x4d, 0xe5, 0xff, 0xee, 0x40, 0xdf,
|
||||||
0x20, 0x64, 0x0b, 0x3e, 0x77, 0xfd, 0x98, 0xf6, 0xcb, 0x6e, 0x5f, 0x75, 0x67, 0x34, 0xbe, 0x24,
|
0x0e, 0x45, 0xf7, 0xa1, 0x93, 0xc8, 0x54, 0xf4, 0x1c, 0xf5, 0x82, 0x4e, 0x9a, 0x35, 0x54, 0xc1,
|
||||||
0x51, 0x3f, 0x88, 0x9f, 0x85, 0xf1, 0xac, 0xec, 0xce, 0x54, 0xf7, 0x57, 0x00, 0x00, 0x00, 0xff,
|
0x89, 0x35, 0xbb, 0x39, 0x71, 0xd0, 0x97, 0xe0, 0x56, 0xa9, 0x6b, 0x4c, 0x1d, 0x05, 0x3a, 0x97,
|
||||||
0xff, 0xd0, 0x84, 0x91, 0x73, 0x59, 0x06, 0x00, 0x00,
|
0x03, 0x9b, 0xcb, 0xc1, 0xc2, 0x32, 0x70, 0x4d, 0xf6, 0xff, 0xde, 0x83, 0xee, 0x4c, 0xa5, 0xfc,
|
||||||
|
0xdb, 0x2a, 0xfa, 0x0c, 0x3a, 0x91, 0xcc, 0x69, 0x13, 0xb2, 0xef, 0x35, 0xb7, 0xa9, 0x28, 0xc7,
|
||||||
|
0x9a, 0x89, 0xbe, 0x80, 0xde, 0x5a, 0x67, 0xb7, 0x11, 0x7b, 0xdc, 0xdc, 0x64, 0x02, 0x1e, 0x5b,
|
||||||
|
0xb6, 0x6c, 0x2c, 0x74, 0xb0, 0xaa, 0x3b, 0xb0, 0xb3, 0xd1, 0xa4, 0x2f, 0xb6, 0x6c, 0xd9, 0x58,
|
||||||
|
0xea, 0x20, 0x54, 0xa1, 0xb1, 0xb3, 0xd1, 0xa4, 0x25, 0xb6, 0x6c, 0xf4, 0x0d, 0xb8, 0x5b, 0x9b,
|
||||||
|
0x8f, 0x2a, 0x2c, 0x76, 0x1e, 0x4c, 0x15, 0xa3, 0xb8, 0xee, 0x90, 0x89, 0x5a, 0x9d, 0x75, 0xc8,
|
||||||
|
0x0a, 0x95, 0x48, 0x2d, 0x3c, 0xa8, 0xb0, 0x59, 0xe1, 0xff, 0xe1, 0xc0, 0xbe, 0x7e, 0x03, 0x8f,
|
||||||
|
0x09, 0x8b, 0x93, 0xeb, 0xc6, 0x4f, 0x24, 0x82, 0xf6, 0x96, 0x26, 0x99, 0xf9, 0x42, 0xaa, 0x67,
|
||||||
|
0x74, 0x06, 0x6d, 0xa9, 0x51, 0x1d, 0xe1, 0xc1, 0xae, 0x5f, 0xb8, 0x9e, 0xbc, 0xb8, 0xce, 0x28,
|
||||||
|
0x56, 0x6c, 0x99, 0xb9, 0xfa, 0xab, 0xee, 0xb5, 0x5f, 0x97, 0xb9, 0xba, 0x0f, 0x1b, 0xee, 0x27,
|
||||||
|
0x2b, 0x80, 0x7a, 0x12, 0x1a, 0x40, 0xef, 0xe1, 0xb3, 0xe5, 0xd3, 0xc5, 0x14, 0x8f, 0xde, 0x41,
|
||||||
|
0x2e, 0x74, 0x2e, 0xcf, 0x97, 0x97, 0xd3, 0x91, 0x23, 0xf1, 0xf9, 0x72, 0x36, 0x3b, 0xc7, 0x2f,
|
||||||
|
0x46, 0x7b, 0xb2, 0x58, 0x3e, 0x5d, 0xbc, 0x78, 0x3e, 0x7d, 0x34, 0x6a, 0xa1, 0x21, 0xb8, 0x4f,
|
||||||
|
0xbe, 0x9d, 0x2f, 0x9e, 0x5d, 0xe2, 0xf3, 0xd9, 0xa8, 0x8d, 0xde, 0x85, 0x3b, 0xaa, 0x27, 0xac,
|
||||||
|
0xc1, 0xce, 0x05, 0x86, 0xc6, 0x3f, 0x18, 0x3f, 0x3c, 0x88, 0x62, 0xb1, 0x2d, 0x57, 0xc1, 0x9a,
|
||||||
|
0xb3, 0x7f, 0xff, 0x45, 0x09, 0x19, 0xdf, 0xd0, 0x64, 0x12, 0xf1, 0xaf, 0x62, 0x1e, 0xd6, 0xab,
|
||||||
|
0xa1, 0x5e, 0xfd, 0x27, 0x00, 0x00, 0xff, 0xff, 0x16, 0x77, 0x81, 0x98, 0xd7, 0x08, 0x00, 0x00,
|
||||||
}
|
}
|
||||||
|
1
vendor/github.com/prometheus/common/expfmt/fuzz.go
generated
vendored
1
vendor/github.com/prometheus/common/expfmt/fuzz.go
generated
vendored
@ -12,6 +12,7 @@
|
|||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
// Build only when actually fuzzing
|
// Build only when actually fuzzing
|
||||||
|
//go:build gofuzz
|
||||||
// +build gofuzz
|
// +build gofuzz
|
||||||
|
|
||||||
package expfmt
|
package expfmt
|
||||||
|
4
vendor/github.com/prometheus/common/expfmt/openmetrics_create.go
generated
vendored
4
vendor/github.com/prometheus/common/expfmt/openmetrics_create.go
generated
vendored
@ -22,7 +22,6 @@ import (
|
|||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/golang/protobuf/ptypes"
|
|
||||||
"github.com/prometheus/common/model"
|
"github.com/prometheus/common/model"
|
||||||
|
|
||||||
dto "github.com/prometheus/client_model/go"
|
dto "github.com/prometheus/client_model/go"
|
||||||
@ -473,10 +472,11 @@ func writeExemplar(w enhancedWriter, e *dto.Exemplar) (int, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return written, err
|
return written, err
|
||||||
}
|
}
|
||||||
ts, err := ptypes.Timestamp((*e).Timestamp)
|
err = (*e).Timestamp.CheckValid()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return written, err
|
return written, err
|
||||||
}
|
}
|
||||||
|
ts := (*e).Timestamp.AsTime()
|
||||||
// TODO(beorn7): Format this directly from components of ts to
|
// TODO(beorn7): Format this directly from components of ts to
|
||||||
// avoid overflow/underflow and precision issues of the float
|
// avoid overflow/underflow and precision issues of the float
|
||||||
// conversion.
|
// conversion.
|
||||||
|
2
vendor/github.com/prometheus/common/model/time.go
generated
vendored
2
vendor/github.com/prometheus/common/model/time.go
generated
vendored
@ -193,7 +193,7 @@ func ParseDuration(durationStr string) (Duration, error) {
|
|||||||
// Allow 0 without a unit.
|
// Allow 0 without a unit.
|
||||||
return 0, nil
|
return 0, nil
|
||||||
case "":
|
case "":
|
||||||
return 0, fmt.Errorf("empty duration string")
|
return 0, errors.New("empty duration string")
|
||||||
}
|
}
|
||||||
matches := durationRE.FindStringSubmatch(durationStr)
|
matches := durationRE.FindStringSubmatch(durationStr)
|
||||||
if matches == nil {
|
if matches == nil {
|
||||||
|
3
vendor/github.com/prometheus/procfs/.gitignore
generated
vendored
3
vendor/github.com/prometheus/procfs/.gitignore
generated
vendored
@ -1 +1,2 @@
|
|||||||
/fixtures/
|
/testdata/fixtures/
|
||||||
|
/fixtures
|
||||||
|
10
vendor/github.com/prometheus/procfs/.golangci.yml
generated
vendored
10
vendor/github.com/prometheus/procfs/.golangci.yml
generated
vendored
@ -1,4 +1,12 @@
|
|||||||
---
|
---
|
||||||
linters:
|
linters:
|
||||||
enable:
|
enable:
|
||||||
- golint
|
- godot
|
||||||
|
- revive
|
||||||
|
|
||||||
|
linter-settings:
|
||||||
|
godot:
|
||||||
|
capital: true
|
||||||
|
exclude:
|
||||||
|
# Ignore "See: URL"
|
||||||
|
- 'See:'
|
||||||
|
4
vendor/github.com/prometheus/procfs/CODE_OF_CONDUCT.md
generated
vendored
4
vendor/github.com/prometheus/procfs/CODE_OF_CONDUCT.md
generated
vendored
@ -1,3 +1,3 @@
|
|||||||
## Prometheus Community Code of Conduct
|
# Prometheus Community Code of Conduct
|
||||||
|
|
||||||
Prometheus follows the [CNCF Code of Conduct](https://github.com/cncf/foundation/blob/master/code-of-conduct.md).
|
Prometheus follows the [CNCF Code of Conduct](https://github.com/cncf/foundation/blob/main/code-of-conduct.md).
|
||||||
|
4
vendor/github.com/prometheus/procfs/CONTRIBUTING.md
generated
vendored
4
vendor/github.com/prometheus/procfs/CONTRIBUTING.md
generated
vendored
@ -97,7 +97,7 @@ Many of the files are changing continuously and the data being read can in some
|
|||||||
reads in the same file. Also, most of the files are relatively small (less than a few KBs), and system calls
|
reads in the same file. Also, most of the files are relatively small (less than a few KBs), and system calls
|
||||||
to the `stat` function will often return the wrong size. Therefore, for most files it's recommended to read the
|
to the `stat` function will often return the wrong size. Therefore, for most files it's recommended to read the
|
||||||
full file in a single operation using an internal utility function called `util.ReadFileNoStat`.
|
full file in a single operation using an internal utility function called `util.ReadFileNoStat`.
|
||||||
This function is similar to `ioutil.ReadFile`, but it avoids the system call to `stat` to get the current size of
|
This function is similar to `os.ReadFile`, but it avoids the system call to `stat` to get the current size of
|
||||||
the file.
|
the file.
|
||||||
|
|
||||||
Note that parsing the file's contents can still be performed one line at a time. This is done by first reading
|
Note that parsing the file's contents can still be performed one line at a time. This is done by first reading
|
||||||
@ -113,7 +113,7 @@ the full file, and then using a scanner on the `[]byte` or `string` containing t
|
|||||||
```
|
```
|
||||||
|
|
||||||
The `/sys` filesystem contains many very small files which contain only a single numeric or text value. These files
|
The `/sys` filesystem contains many very small files which contain only a single numeric or text value. These files
|
||||||
can be read using an internal function called `util.SysReadFile` which is similar to `ioutil.ReadFile` but does
|
can be read using an internal function called `util.SysReadFile` which is similar to `os.ReadFile` but does
|
||||||
not bother to check the size of the file before reading.
|
not bother to check the size of the file before reading.
|
||||||
```
|
```
|
||||||
data, err := util.SysReadFile("/sys/class/power_supply/BAT0/capacity")
|
data, err := util.SysReadFile("/sys/class/power_supply/BAT0/capacity")
|
||||||
|
10
vendor/github.com/prometheus/procfs/Makefile
generated
vendored
10
vendor/github.com/prometheus/procfs/Makefile
generated
vendored
@ -14,18 +14,18 @@
|
|||||||
include Makefile.common
|
include Makefile.common
|
||||||
|
|
||||||
%/.unpacked: %.ttar
|
%/.unpacked: %.ttar
|
||||||
@echo ">> extracting fixtures"
|
@echo ">> extracting fixtures $*"
|
||||||
./ttar -C $(dir $*) -x -f $*.ttar
|
./ttar -C $(dir $*) -x -f $*.ttar
|
||||||
touch $@
|
touch $@
|
||||||
|
|
||||||
fixtures: fixtures/.unpacked
|
fixtures: testdata/fixtures/.unpacked
|
||||||
|
|
||||||
update_fixtures:
|
update_fixtures:
|
||||||
rm -vf fixtures/.unpacked
|
rm -vf testdata/fixtures/.unpacked
|
||||||
./ttar -c -f fixtures.ttar fixtures/
|
./ttar -c -f testdata/fixtures.ttar -C testdata/ fixtures/
|
||||||
|
|
||||||
.PHONY: build
|
.PHONY: build
|
||||||
build:
|
build:
|
||||||
|
|
||||||
.PHONY: test
|
.PHONY: test
|
||||||
test: fixtures/.unpacked common-test
|
test: testdata/fixtures/.unpacked common-test
|
||||||
|
89
vendor/github.com/prometheus/procfs/Makefile.common
generated
vendored
89
vendor/github.com/prometheus/procfs/Makefile.common
generated
vendored
@ -36,29 +36,6 @@ GO_VERSION ?= $(shell $(GO) version)
|
|||||||
GO_VERSION_NUMBER ?= $(word 3, $(GO_VERSION))
|
GO_VERSION_NUMBER ?= $(word 3, $(GO_VERSION))
|
||||||
PRE_GO_111 ?= $(shell echo $(GO_VERSION_NUMBER) | grep -E 'go1\.(10|[0-9])\.')
|
PRE_GO_111 ?= $(shell echo $(GO_VERSION_NUMBER) | grep -E 'go1\.(10|[0-9])\.')
|
||||||
|
|
||||||
GOVENDOR :=
|
|
||||||
GO111MODULE :=
|
|
||||||
ifeq (, $(PRE_GO_111))
|
|
||||||
ifneq (,$(wildcard go.mod))
|
|
||||||
# Enforce Go modules support just in case the directory is inside GOPATH (and for Travis CI).
|
|
||||||
GO111MODULE := on
|
|
||||||
|
|
||||||
ifneq (,$(wildcard vendor))
|
|
||||||
# Always use the local vendor/ directory to satisfy the dependencies.
|
|
||||||
GOOPTS := $(GOOPTS) -mod=vendor
|
|
||||||
endif
|
|
||||||
endif
|
|
||||||
else
|
|
||||||
ifneq (,$(wildcard go.mod))
|
|
||||||
ifneq (,$(wildcard vendor))
|
|
||||||
$(warning This repository requires Go >= 1.11 because of Go modules)
|
|
||||||
$(warning Some recipes may not work as expected as the current Go runtime is '$(GO_VERSION_NUMBER)')
|
|
||||||
endif
|
|
||||||
else
|
|
||||||
# This repository isn't using Go modules (yet).
|
|
||||||
GOVENDOR := $(FIRST_GOPATH)/bin/govendor
|
|
||||||
endif
|
|
||||||
endif
|
|
||||||
PROMU := $(FIRST_GOPATH)/bin/promu
|
PROMU := $(FIRST_GOPATH)/bin/promu
|
||||||
pkgs = ./...
|
pkgs = ./...
|
||||||
|
|
||||||
@ -78,17 +55,23 @@ ifneq ($(shell which gotestsum),)
|
|||||||
endif
|
endif
|
||||||
endif
|
endif
|
||||||
|
|
||||||
PROMU_VERSION ?= 0.12.0
|
PROMU_VERSION ?= 0.13.0
|
||||||
PROMU_URL := https://github.com/prometheus/promu/releases/download/v$(PROMU_VERSION)/promu-$(PROMU_VERSION).$(GO_BUILD_PLATFORM).tar.gz
|
PROMU_URL := https://github.com/prometheus/promu/releases/download/v$(PROMU_VERSION)/promu-$(PROMU_VERSION).$(GO_BUILD_PLATFORM).tar.gz
|
||||||
|
|
||||||
GOLANGCI_LINT :=
|
GOLANGCI_LINT :=
|
||||||
GOLANGCI_LINT_OPTS ?=
|
GOLANGCI_LINT_OPTS ?=
|
||||||
GOLANGCI_LINT_VERSION ?= v1.39.0
|
GOLANGCI_LINT_VERSION ?= v1.45.2
|
||||||
# golangci-lint only supports linux, darwin and windows platforms on i386/amd64.
|
# golangci-lint only supports linux, darwin and windows platforms on i386/amd64.
|
||||||
# windows isn't included here because of the path separator being different.
|
# windows isn't included here because of the path separator being different.
|
||||||
ifeq ($(GOHOSTOS),$(filter $(GOHOSTOS),linux darwin))
|
ifeq ($(GOHOSTOS),$(filter $(GOHOSTOS),linux darwin))
|
||||||
ifeq ($(GOHOSTARCH),$(filter $(GOHOSTARCH),amd64 i386))
|
ifeq ($(GOHOSTARCH),$(filter $(GOHOSTARCH),amd64 i386))
|
||||||
GOLANGCI_LINT := $(FIRST_GOPATH)/bin/golangci-lint
|
# If we're in CI and there is an Actions file, that means the linter
|
||||||
|
# is being run in Actions, so we don't need to run it here.
|
||||||
|
ifeq (,$(CIRCLE_JOB))
|
||||||
|
GOLANGCI_LINT := $(FIRST_GOPATH)/bin/golangci-lint
|
||||||
|
else ifeq (,$(wildcard .github/workflows/golangci-lint.yml))
|
||||||
|
GOLANGCI_LINT := $(FIRST_GOPATH)/bin/golangci-lint
|
||||||
|
endif
|
||||||
endif
|
endif
|
||||||
endif
|
endif
|
||||||
|
|
||||||
@ -144,32 +127,25 @@ common-check_license:
|
|||||||
.PHONY: common-deps
|
.PHONY: common-deps
|
||||||
common-deps:
|
common-deps:
|
||||||
@echo ">> getting dependencies"
|
@echo ">> getting dependencies"
|
||||||
ifdef GO111MODULE
|
$(GO) mod download
|
||||||
GO111MODULE=$(GO111MODULE) $(GO) mod download
|
|
||||||
else
|
|
||||||
$(GO) get $(GOOPTS) -t ./...
|
|
||||||
endif
|
|
||||||
|
|
||||||
.PHONY: update-go-deps
|
.PHONY: update-go-deps
|
||||||
update-go-deps:
|
update-go-deps:
|
||||||
@echo ">> updating Go dependencies"
|
@echo ">> updating Go dependencies"
|
||||||
@for m in $$($(GO) list -mod=readonly -m -f '{{ if and (not .Indirect) (not .Main)}}{{.Path}}{{end}}' all); do \
|
@for m in $$($(GO) list -mod=readonly -m -f '{{ if and (not .Indirect) (not .Main)}}{{.Path}}{{end}}' all); do \
|
||||||
$(GO) get $$m; \
|
$(GO) get -d $$m; \
|
||||||
done
|
done
|
||||||
GO111MODULE=$(GO111MODULE) $(GO) mod tidy
|
$(GO) mod tidy
|
||||||
ifneq (,$(wildcard vendor))
|
|
||||||
GO111MODULE=$(GO111MODULE) $(GO) mod vendor
|
|
||||||
endif
|
|
||||||
|
|
||||||
.PHONY: common-test-short
|
.PHONY: common-test-short
|
||||||
common-test-short: $(GOTEST_DIR)
|
common-test-short: $(GOTEST_DIR)
|
||||||
@echo ">> running short tests"
|
@echo ">> running short tests"
|
||||||
GO111MODULE=$(GO111MODULE) $(GOTEST) -short $(GOOPTS) $(pkgs)
|
$(GOTEST) -short $(GOOPTS) $(pkgs)
|
||||||
|
|
||||||
.PHONY: common-test
|
.PHONY: common-test
|
||||||
common-test: $(GOTEST_DIR)
|
common-test: $(GOTEST_DIR)
|
||||||
@echo ">> running all tests"
|
@echo ">> running all tests"
|
||||||
GO111MODULE=$(GO111MODULE) $(GOTEST) $(test-flags) $(GOOPTS) $(pkgs)
|
$(GOTEST) $(test-flags) $(GOOPTS) $(pkgs)
|
||||||
|
|
||||||
$(GOTEST_DIR):
|
$(GOTEST_DIR):
|
||||||
@mkdir -p $@
|
@mkdir -p $@
|
||||||
@ -177,25 +153,21 @@ $(GOTEST_DIR):
|
|||||||
.PHONY: common-format
|
.PHONY: common-format
|
||||||
common-format:
|
common-format:
|
||||||
@echo ">> formatting code"
|
@echo ">> formatting code"
|
||||||
GO111MODULE=$(GO111MODULE) $(GO) fmt $(pkgs)
|
$(GO) fmt $(pkgs)
|
||||||
|
|
||||||
.PHONY: common-vet
|
.PHONY: common-vet
|
||||||
common-vet:
|
common-vet:
|
||||||
@echo ">> vetting code"
|
@echo ">> vetting code"
|
||||||
GO111MODULE=$(GO111MODULE) $(GO) vet $(GOOPTS) $(pkgs)
|
$(GO) vet $(GOOPTS) $(pkgs)
|
||||||
|
|
||||||
.PHONY: common-lint
|
.PHONY: common-lint
|
||||||
common-lint: $(GOLANGCI_LINT)
|
common-lint: $(GOLANGCI_LINT)
|
||||||
ifdef GOLANGCI_LINT
|
ifdef GOLANGCI_LINT
|
||||||
@echo ">> running golangci-lint"
|
@echo ">> running golangci-lint"
|
||||||
ifdef GO111MODULE
|
|
||||||
# 'go list' needs to be executed before staticcheck to prepopulate the modules cache.
|
# 'go list' needs to be executed before staticcheck to prepopulate the modules cache.
|
||||||
# Otherwise staticcheck might fail randomly for some reason not yet explained.
|
# Otherwise staticcheck might fail randomly for some reason not yet explained.
|
||||||
GO111MODULE=$(GO111MODULE) $(GO) list -e -compiled -test=true -export=false -deps=true -find=false -tags= -- ./... > /dev/null
|
$(GO) list -e -compiled -test=true -export=false -deps=true -find=false -tags= -- ./... > /dev/null
|
||||||
GO111MODULE=$(GO111MODULE) $(GOLANGCI_LINT) run $(GOLANGCI_LINT_OPTS) $(pkgs)
|
$(GOLANGCI_LINT) run $(GOLANGCI_LINT_OPTS) $(pkgs)
|
||||||
else
|
|
||||||
$(GOLANGCI_LINT) run $(pkgs)
|
|
||||||
endif
|
|
||||||
endif
|
endif
|
||||||
|
|
||||||
.PHONY: common-yamllint
|
.PHONY: common-yamllint
|
||||||
@ -212,28 +184,15 @@ endif
|
|||||||
common-staticcheck: lint
|
common-staticcheck: lint
|
||||||
|
|
||||||
.PHONY: common-unused
|
.PHONY: common-unused
|
||||||
common-unused: $(GOVENDOR)
|
common-unused:
|
||||||
ifdef GOVENDOR
|
|
||||||
@echo ">> running check for unused packages"
|
|
||||||
@$(GOVENDOR) list +unused | grep . && exit 1 || echo 'No unused packages'
|
|
||||||
else
|
|
||||||
ifdef GO111MODULE
|
|
||||||
@echo ">> running check for unused/missing packages in go.mod"
|
@echo ">> running check for unused/missing packages in go.mod"
|
||||||
GO111MODULE=$(GO111MODULE) $(GO) mod tidy
|
$(GO) mod tidy
|
||||||
ifeq (,$(wildcard vendor))
|
|
||||||
@git diff --exit-code -- go.sum go.mod
|
@git diff --exit-code -- go.sum go.mod
|
||||||
else
|
|
||||||
@echo ">> running check for unused packages in vendor/"
|
|
||||||
GO111MODULE=$(GO111MODULE) $(GO) mod vendor
|
|
||||||
@git diff --exit-code -- go.sum go.mod vendor/
|
|
||||||
endif
|
|
||||||
endif
|
|
||||||
endif
|
|
||||||
|
|
||||||
.PHONY: common-build
|
.PHONY: common-build
|
||||||
common-build: promu
|
common-build: promu
|
||||||
@echo ">> building binaries"
|
@echo ">> building binaries"
|
||||||
GO111MODULE=$(GO111MODULE) $(PROMU) build --prefix $(PREFIX) $(PROMU_BINARIES)
|
$(PROMU) build --prefix $(PREFIX) $(PROMU_BINARIES)
|
||||||
|
|
||||||
.PHONY: common-tarball
|
.PHONY: common-tarball
|
||||||
common-tarball: promu
|
common-tarball: promu
|
||||||
@ -289,12 +248,6 @@ $(GOLANGCI_LINT):
|
|||||||
| sh -s -- -b $(FIRST_GOPATH)/bin $(GOLANGCI_LINT_VERSION)
|
| sh -s -- -b $(FIRST_GOPATH)/bin $(GOLANGCI_LINT_VERSION)
|
||||||
endif
|
endif
|
||||||
|
|
||||||
ifdef GOVENDOR
|
|
||||||
.PHONY: $(GOVENDOR)
|
|
||||||
$(GOVENDOR):
|
|
||||||
GOOS= GOARCH= $(GO) get -u github.com/kardianos/govendor
|
|
||||||
endif
|
|
||||||
|
|
||||||
.PHONY: precheck
|
.PHONY: precheck
|
||||||
precheck::
|
precheck::
|
||||||
|
|
||||||
|
2
vendor/github.com/prometheus/procfs/SECURITY.md
generated
vendored
2
vendor/github.com/prometheus/procfs/SECURITY.md
generated
vendored
@ -3,4 +3,4 @@
|
|||||||
The Prometheus security policy, including how to report vulnerabilities, can be
|
The Prometheus security policy, including how to report vulnerabilities, can be
|
||||||
found here:
|
found here:
|
||||||
|
|
||||||
https://prometheus.io/docs/operating/security/
|
<https://prometheus.io/docs/operating/security/>
|
||||||
|
45
vendor/github.com/prometheus/procfs/arp.go
generated
vendored
45
vendor/github.com/prometheus/procfs/arp.go
generated
vendored
@ -15,11 +15,28 @@ package procfs
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
|
||||||
"net"
|
"net"
|
||||||
|
"os"
|
||||||
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Learned from include/uapi/linux/if_arp.h.
|
||||||
|
const (
|
||||||
|
// completed entry (ha valid).
|
||||||
|
ATFComplete = 0x02
|
||||||
|
// permanent entry.
|
||||||
|
ATFPermanent = 0x04
|
||||||
|
// Publish entry.
|
||||||
|
ATFPublish = 0x08
|
||||||
|
// Has requested trailers.
|
||||||
|
ATFUseTrailers = 0x10
|
||||||
|
// Obsoleted: Want to use a netmask (only for proxy entries).
|
||||||
|
ATFNetmask = 0x20
|
||||||
|
// Don't answer this addresses.
|
||||||
|
ATFDontPublish = 0x40
|
||||||
|
)
|
||||||
|
|
||||||
// ARPEntry contains a single row of the columnar data represented in
|
// ARPEntry contains a single row of the columnar data represented in
|
||||||
// /proc/net/arp.
|
// /proc/net/arp.
|
||||||
type ARPEntry struct {
|
type ARPEntry struct {
|
||||||
@ -29,12 +46,14 @@ type ARPEntry struct {
|
|||||||
HWAddr net.HardwareAddr
|
HWAddr net.HardwareAddr
|
||||||
// Name of the device
|
// Name of the device
|
||||||
Device string
|
Device string
|
||||||
|
// Flags
|
||||||
|
Flags byte
|
||||||
}
|
}
|
||||||
|
|
||||||
// GatherARPEntries retrieves all the ARP entries, parse the relevant columns,
|
// GatherARPEntries retrieves all the ARP entries, parse the relevant columns,
|
||||||
// and then return a slice of ARPEntry's.
|
// and then return a slice of ARPEntry's.
|
||||||
func (fs FS) GatherARPEntries() ([]ARPEntry, error) {
|
func (fs FS) GatherARPEntries() ([]ARPEntry, error) {
|
||||||
data, err := ioutil.ReadFile(fs.proc.Path("net/arp"))
|
data, err := os.ReadFile(fs.proc.Path("net/arp"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("error reading arp %q: %w", fs.proc.Path("net/arp"), err)
|
return nil, fmt.Errorf("error reading arp %q: %w", fs.proc.Path("net/arp"), err)
|
||||||
}
|
}
|
||||||
@ -72,14 +91,26 @@ func parseARPEntries(data []byte) ([]ARPEntry, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func parseARPEntry(columns []string) (ARPEntry, error) {
|
func parseARPEntry(columns []string) (ARPEntry, error) {
|
||||||
|
entry := ARPEntry{Device: columns[5]}
|
||||||
ip := net.ParseIP(columns[0])
|
ip := net.ParseIP(columns[0])
|
||||||
mac := net.HardwareAddr(columns[3])
|
entry.IPAddr = ip
|
||||||
|
|
||||||
entry := ARPEntry{
|
if mac, err := net.ParseMAC(columns[3]); err == nil {
|
||||||
IPAddr: ip,
|
entry.HWAddr = mac
|
||||||
HWAddr: mac,
|
} else {
|
||||||
Device: columns[5],
|
return ARPEntry{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if flags, err := strconv.ParseUint(columns[2], 0, 8); err == nil {
|
||||||
|
entry.Flags = byte(flags)
|
||||||
|
} else {
|
||||||
|
return ARPEntry{}, err
|
||||||
}
|
}
|
||||||
|
|
||||||
return entry, nil
|
return entry, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// IsComplete returns true if ARP entry is marked with complete flag.
|
||||||
|
func (entry *ARPEntry) IsComplete() bool {
|
||||||
|
return entry.Flags&ATFComplete != 0
|
||||||
|
}
|
||||||
|
5
vendor/github.com/prometheus/procfs/cpuinfo.go
generated
vendored
5
vendor/github.com/prometheus/procfs/cpuinfo.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build linux
|
||||||
// +build linux
|
// +build linux
|
||||||
|
|
||||||
package procfs
|
package procfs
|
||||||
@ -27,7 +28,7 @@ import (
|
|||||||
"github.com/prometheus/procfs/internal/util"
|
"github.com/prometheus/procfs/internal/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// CPUInfo contains general information about a system CPU found in /proc/cpuinfo
|
// CPUInfo contains general information about a system CPU found in /proc/cpuinfo.
|
||||||
type CPUInfo struct {
|
type CPUInfo struct {
|
||||||
Processor uint
|
Processor uint
|
||||||
VendorID string
|
VendorID string
|
||||||
@ -469,7 +470,7 @@ func parseCPUInfoDummy(_ []byte) ([]CPUInfo, error) { // nolint:unused,deadcode
|
|||||||
}
|
}
|
||||||
|
|
||||||
// firstNonEmptyLine advances the scanner to the first non-empty line
|
// firstNonEmptyLine advances the scanner to the first non-empty line
|
||||||
// and returns the contents of that line
|
// and returns the contents of that line.
|
||||||
func firstNonEmptyLine(scanner *bufio.Scanner) string {
|
func firstNonEmptyLine(scanner *bufio.Scanner) string {
|
||||||
for scanner.Scan() {
|
for scanner.Scan() {
|
||||||
line := scanner.Text()
|
line := scanner.Text()
|
||||||
|
1
vendor/github.com/prometheus/procfs/cpuinfo_armx.go
generated
vendored
1
vendor/github.com/prometheus/procfs/cpuinfo_armx.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build linux && (arm || arm64)
|
||||||
// +build linux
|
// +build linux
|
||||||
// +build arm arm64
|
// +build arm arm64
|
||||||
|
|
||||||
|
1
vendor/github.com/prometheus/procfs/cpuinfo_mipsx.go
generated
vendored
1
vendor/github.com/prometheus/procfs/cpuinfo_mipsx.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build linux && (mips || mipsle || mips64 || mips64le)
|
||||||
// +build linux
|
// +build linux
|
||||||
// +build mips mipsle mips64 mips64le
|
// +build mips mipsle mips64 mips64le
|
||||||
|
|
||||||
|
4
vendor/github.com/prometheus/procfs/cpuinfo_others.go
generated
vendored
4
vendor/github.com/prometheus/procfs/cpuinfo_others.go
generated
vendored
@ -11,8 +11,8 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
// +build linux
|
//go:build linux && !386 && !amd64 && !arm && !arm64 && !mips && !mips64 && !mips64le && !mipsle && !ppc64 && !ppc64le && !riscv64 && !s390x
|
||||||
// +build !386,!amd64,!arm,!arm64,!mips,!mips64,!mips64le,!mipsle,!ppc64,!ppc64le,!riscv64,!s390x
|
// +build linux,!386,!amd64,!arm,!arm64,!mips,!mips64,!mips64le,!mipsle,!ppc64,!ppc64le,!riscv64,!s390x
|
||||||
|
|
||||||
package procfs
|
package procfs
|
||||||
|
|
||||||
|
1
vendor/github.com/prometheus/procfs/cpuinfo_ppcx.go
generated
vendored
1
vendor/github.com/prometheus/procfs/cpuinfo_ppcx.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build linux && (ppc64 || ppc64le)
|
||||||
// +build linux
|
// +build linux
|
||||||
// +build ppc64 ppc64le
|
// +build ppc64 ppc64le
|
||||||
|
|
||||||
|
1
vendor/github.com/prometheus/procfs/cpuinfo_riscvx.go
generated
vendored
1
vendor/github.com/prometheus/procfs/cpuinfo_riscvx.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build linux && (riscv || riscv64)
|
||||||
// +build linux
|
// +build linux
|
||||||
// +build riscv riscv64
|
// +build riscv riscv64
|
||||||
|
|
||||||
|
1
vendor/github.com/prometheus/procfs/cpuinfo_s390x.go
generated
vendored
1
vendor/github.com/prometheus/procfs/cpuinfo_s390x.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build linux
|
||||||
// +build linux
|
// +build linux
|
||||||
|
|
||||||
package procfs
|
package procfs
|
||||||
|
1
vendor/github.com/prometheus/procfs/cpuinfo_x86.go
generated
vendored
1
vendor/github.com/prometheus/procfs/cpuinfo_x86.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build linux && (386 || amd64)
|
||||||
// +build linux
|
// +build linux
|
||||||
// +build 386 amd64
|
// +build 386 amd64
|
||||||
|
|
||||||
|
7673
vendor/github.com/prometheus/procfs/fixtures.ttar
generated
vendored
7673
vendor/github.com/prometheus/procfs/fixtures.ttar
generated
vendored
File diff suppressed because it is too large
Load Diff
2
vendor/github.com/prometheus/procfs/internal/fs/fs.go
generated
vendored
2
vendor/github.com/prometheus/procfs/internal/fs/fs.go
generated
vendored
@ -26,7 +26,7 @@ const (
|
|||||||
// DefaultSysMountPoint is the common mount point of the sys filesystem.
|
// DefaultSysMountPoint is the common mount point of the sys filesystem.
|
||||||
DefaultSysMountPoint = "/sys"
|
DefaultSysMountPoint = "/sys"
|
||||||
|
|
||||||
// DefaultConfigfsMountPoint is the common mount point of the configfs
|
// DefaultConfigfsMountPoint is the common mount point of the configfs.
|
||||||
DefaultConfigfsMountPoint = "/sys/kernel/config"
|
DefaultConfigfsMountPoint = "/sys/kernel/config"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
6
vendor/github.com/prometheus/procfs/internal/util/parse.go
generated
vendored
6
vendor/github.com/prometheus/procfs/internal/util/parse.go
generated
vendored
@ -14,7 +14,7 @@
|
|||||||
package util
|
package util
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"io/ioutil"
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
@ -66,7 +66,7 @@ func ParsePInt64s(ss []string) ([]*int64, error) {
|
|||||||
|
|
||||||
// ReadUintFromFile reads a file and attempts to parse a uint64 from it.
|
// ReadUintFromFile reads a file and attempts to parse a uint64 from it.
|
||||||
func ReadUintFromFile(path string) (uint64, error) {
|
func ReadUintFromFile(path string) (uint64, error) {
|
||||||
data, err := ioutil.ReadFile(path)
|
data, err := os.ReadFile(path)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
@ -75,7 +75,7 @@ func ReadUintFromFile(path string) (uint64, error) {
|
|||||||
|
|
||||||
// ReadIntFromFile reads a file and attempts to parse a int64 from it.
|
// ReadIntFromFile reads a file and attempts to parse a int64 from it.
|
||||||
func ReadIntFromFile(path string) (int64, error) {
|
func ReadIntFromFile(path string) (int64, error) {
|
||||||
data, err := ioutil.ReadFile(path)
|
data, err := os.ReadFile(path)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
|
11
vendor/github.com/prometheus/procfs/internal/util/readfile.go
generated
vendored
11
vendor/github.com/prometheus/procfs/internal/util/readfile.go
generated
vendored
@ -15,17 +15,16 @@ package util
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"io"
|
"io"
|
||||||
"io/ioutil"
|
|
||||||
"os"
|
"os"
|
||||||
)
|
)
|
||||||
|
|
||||||
// ReadFileNoStat uses ioutil.ReadAll to read contents of entire file.
|
// ReadFileNoStat uses io.ReadAll to read contents of entire file.
|
||||||
// This is similar to ioutil.ReadFile but without the call to os.Stat, because
|
// This is similar to os.ReadFile but without the call to os.Stat, because
|
||||||
// many files in /proc and /sys report incorrect file sizes (either 0 or 4096).
|
// many files in /proc and /sys report incorrect file sizes (either 0 or 4096).
|
||||||
// Reads a max file size of 512kB. For files larger than this, a scanner
|
// Reads a max file size of 1024kB. For files larger than this, a scanner
|
||||||
// should be used.
|
// should be used.
|
||||||
func ReadFileNoStat(filename string) ([]byte, error) {
|
func ReadFileNoStat(filename string) ([]byte, error) {
|
||||||
const maxBufferSize = 1024 * 512
|
const maxBufferSize = 1024 * 1024
|
||||||
|
|
||||||
f, err := os.Open(filename)
|
f, err := os.Open(filename)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -34,5 +33,5 @@ func ReadFileNoStat(filename string) ([]byte, error) {
|
|||||||
defer f.Close()
|
defer f.Close()
|
||||||
|
|
||||||
reader := io.LimitReader(f, maxBufferSize)
|
reader := io.LimitReader(f, maxBufferSize)
|
||||||
return ioutil.ReadAll(reader)
|
return io.ReadAll(reader)
|
||||||
}
|
}
|
||||||
|
8
vendor/github.com/prometheus/procfs/internal/util/sysreadfile.go
generated
vendored
8
vendor/github.com/prometheus/procfs/internal/util/sysreadfile.go
generated
vendored
@ -11,7 +11,9 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
// +build linux,!appengine
|
//go:build (linux || darwin) && !appengine
|
||||||
|
// +build linux darwin
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
package util
|
package util
|
||||||
|
|
||||||
@ -21,7 +23,7 @@ import (
|
|||||||
"syscall"
|
"syscall"
|
||||||
)
|
)
|
||||||
|
|
||||||
// SysReadFile is a simplified ioutil.ReadFile that invokes syscall.Read directly.
|
// SysReadFile is a simplified os.ReadFile that invokes syscall.Read directly.
|
||||||
// https://github.com/prometheus/node_exporter/pull/728/files
|
// https://github.com/prometheus/node_exporter/pull/728/files
|
||||||
//
|
//
|
||||||
// Note that this function will not read files larger than 128 bytes.
|
// Note that this function will not read files larger than 128 bytes.
|
||||||
@ -33,7 +35,7 @@ func SysReadFile(file string) (string, error) {
|
|||||||
defer f.Close()
|
defer f.Close()
|
||||||
|
|
||||||
// On some machines, hwmon drivers are broken and return EAGAIN. This causes
|
// On some machines, hwmon drivers are broken and return EAGAIN. This causes
|
||||||
// Go's ioutil.ReadFile implementation to poll forever.
|
// Go's os.ReadFile implementation to poll forever.
|
||||||
//
|
//
|
||||||
// Since we either want to read data or bail immediately, do the simplest
|
// Since we either want to read data or bail immediately, do the simplest
|
||||||
// possible read using syscall directly.
|
// possible read using syscall directly.
|
||||||
|
3
vendor/github.com/prometheus/procfs/internal/util/sysreadfile_compat.go
generated
vendored
3
vendor/github.com/prometheus/procfs/internal/util/sysreadfile_compat.go
generated
vendored
@ -11,7 +11,8 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
// +build linux,appengine !linux
|
//go:build (linux && appengine) || (!linux && !darwin)
|
||||||
|
// +build linux,appengine !linux,!darwin
|
||||||
|
|
||||||
package util
|
package util
|
||||||
|
|
||||||
|
3
vendor/github.com/prometheus/procfs/ipvs.go
generated
vendored
3
vendor/github.com/prometheus/procfs/ipvs.go
generated
vendored
@ -20,7 +20,6 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"io/ioutil"
|
|
||||||
"net"
|
"net"
|
||||||
"os"
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
@ -84,7 +83,7 @@ func parseIPVSStats(r io.Reader) (IPVSStats, error) {
|
|||||||
stats IPVSStats
|
stats IPVSStats
|
||||||
)
|
)
|
||||||
|
|
||||||
statContent, err := ioutil.ReadAll(r)
|
statContent, err := io.ReadAll(r)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return IPVSStats{}, err
|
return IPVSStats{}, err
|
||||||
}
|
}
|
||||||
|
1
vendor/github.com/prometheus/procfs/kernel_random.go
generated
vendored
1
vendor/github.com/prometheus/procfs/kernel_random.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build !windows
|
||||||
// +build !windows
|
// +build !windows
|
||||||
|
|
||||||
package procfs
|
package procfs
|
||||||
|
2
vendor/github.com/prometheus/procfs/loadavg.go
generated
vendored
2
vendor/github.com/prometheus/procfs/loadavg.go
generated
vendored
@ -21,7 +21,7 @@ import (
|
|||||||
"github.com/prometheus/procfs/internal/util"
|
"github.com/prometheus/procfs/internal/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// LoadAvg represents an entry in /proc/loadavg
|
// LoadAvg represents an entry in /proc/loadavg.
|
||||||
type LoadAvg struct {
|
type LoadAvg struct {
|
||||||
Load1 float64
|
Load1 float64
|
||||||
Load5 float64
|
Load5 float64
|
||||||
|
10
vendor/github.com/prometheus/procfs/mdstat.go
generated
vendored
10
vendor/github.com/prometheus/procfs/mdstat.go
generated
vendored
@ -15,7 +15,7 @@ package procfs
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
"os"
|
||||||
"regexp"
|
"regexp"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
@ -64,7 +64,7 @@ type MDStat struct {
|
|||||||
// structs containing the relevant info. More information available here:
|
// structs containing the relevant info. More information available here:
|
||||||
// https://raid.wiki.kernel.org/index.php/Mdstat
|
// https://raid.wiki.kernel.org/index.php/Mdstat
|
||||||
func (fs FS) MDStat() ([]MDStat, error) {
|
func (fs FS) MDStat() ([]MDStat, error) {
|
||||||
data, err := ioutil.ReadFile(fs.proc.Path("mdstat"))
|
data, err := os.ReadFile(fs.proc.Path("mdstat"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -166,8 +166,12 @@ func parseMDStat(mdStatData []byte) ([]MDStat, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func evalStatusLine(deviceLine, statusLine string) (active, total, down, size int64, err error) {
|
func evalStatusLine(deviceLine, statusLine string) (active, total, down, size int64, err error) {
|
||||||
|
statusFields := strings.Fields(statusLine)
|
||||||
|
if len(statusFields) < 1 {
|
||||||
|
return 0, 0, 0, 0, fmt.Errorf("unexpected statusLine %q", statusLine)
|
||||||
|
}
|
||||||
|
|
||||||
sizeStr := strings.Fields(statusLine)[0]
|
sizeStr := statusFields[0]
|
||||||
size, err = strconv.ParseInt(sizeStr, 10, 64)
|
size, err = strconv.ParseInt(sizeStr, 10, 64)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, 0, 0, 0, fmt.Errorf("unexpected statusLine %q: %w", statusLine, err)
|
return 0, 0, 0, 0, fmt.Errorf("unexpected statusLine %q: %w", statusLine, err)
|
||||||
|
12
vendor/github.com/prometheus/procfs/net_conntrackstat.go
generated
vendored
12
vendor/github.com/prometheus/procfs/net_conntrackstat.go
generated
vendored
@ -25,7 +25,7 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
// A ConntrackStatEntry represents one line from net/stat/nf_conntrack
|
// A ConntrackStatEntry represents one line from net/stat/nf_conntrack
|
||||||
// and contains netfilter conntrack statistics at one CPU core
|
// and contains netfilter conntrack statistics at one CPU core.
|
||||||
type ConntrackStatEntry struct {
|
type ConntrackStatEntry struct {
|
||||||
Entries uint64
|
Entries uint64
|
||||||
Found uint64
|
Found uint64
|
||||||
@ -38,12 +38,12 @@ type ConntrackStatEntry struct {
|
|||||||
SearchRestart uint64
|
SearchRestart uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
// ConntrackStat retrieves netfilter's conntrack statistics, split by CPU cores
|
// ConntrackStat retrieves netfilter's conntrack statistics, split by CPU cores.
|
||||||
func (fs FS) ConntrackStat() ([]ConntrackStatEntry, error) {
|
func (fs FS) ConntrackStat() ([]ConntrackStatEntry, error) {
|
||||||
return readConntrackStat(fs.proc.Path("net", "stat", "nf_conntrack"))
|
return readConntrackStat(fs.proc.Path("net", "stat", "nf_conntrack"))
|
||||||
}
|
}
|
||||||
|
|
||||||
// Parses a slice of ConntrackStatEntries from the given filepath
|
// Parses a slice of ConntrackStatEntries from the given filepath.
|
||||||
func readConntrackStat(path string) ([]ConntrackStatEntry, error) {
|
func readConntrackStat(path string) ([]ConntrackStatEntry, error) {
|
||||||
// This file is small and can be read with one syscall.
|
// This file is small and can be read with one syscall.
|
||||||
b, err := util.ReadFileNoStat(path)
|
b, err := util.ReadFileNoStat(path)
|
||||||
@ -61,7 +61,7 @@ func readConntrackStat(path string) ([]ConntrackStatEntry, error) {
|
|||||||
return stat, nil
|
return stat, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Reads the contents of a conntrack statistics file and parses a slice of ConntrackStatEntries
|
// Reads the contents of a conntrack statistics file and parses a slice of ConntrackStatEntries.
|
||||||
func parseConntrackStat(r io.Reader) ([]ConntrackStatEntry, error) {
|
func parseConntrackStat(r io.Reader) ([]ConntrackStatEntry, error) {
|
||||||
var entries []ConntrackStatEntry
|
var entries []ConntrackStatEntry
|
||||||
|
|
||||||
@ -79,7 +79,7 @@ func parseConntrackStat(r io.Reader) ([]ConntrackStatEntry, error) {
|
|||||||
return entries, nil
|
return entries, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Parses a ConntrackStatEntry from given array of fields
|
// Parses a ConntrackStatEntry from given array of fields.
|
||||||
func parseConntrackStatEntry(fields []string) (*ConntrackStatEntry, error) {
|
func parseConntrackStatEntry(fields []string) (*ConntrackStatEntry, error) {
|
||||||
if len(fields) != 17 {
|
if len(fields) != 17 {
|
||||||
return nil, fmt.Errorf("invalid conntrackstat entry, missing fields")
|
return nil, fmt.Errorf("invalid conntrackstat entry, missing fields")
|
||||||
@ -143,7 +143,7 @@ func parseConntrackStatEntry(fields []string) (*ConntrackStatEntry, error) {
|
|||||||
return entry, nil
|
return entry, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Parses a uint64 from given hex in string
|
// Parses a uint64 from given hex in string.
|
||||||
func parseConntrackStatField(field string) (uint64, error) {
|
func parseConntrackStatField(field string) (uint64, error) {
|
||||||
val, err := strconv.ParseUint(field, 16, 64)
|
val, err := strconv.ParseUint(field, 16, 64)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
8
vendor/github.com/prometheus/procfs/net_dev.go
generated
vendored
8
vendor/github.com/prometheus/procfs/net_dev.go
generated
vendored
@ -87,17 +87,17 @@ func newNetDev(file string) (NetDev, error) {
|
|||||||
// parseLine parses a single line from the /proc/net/dev file. Header lines
|
// parseLine parses a single line from the /proc/net/dev file. Header lines
|
||||||
// must be filtered prior to calling this method.
|
// must be filtered prior to calling this method.
|
||||||
func (netDev NetDev) parseLine(rawLine string) (*NetDevLine, error) {
|
func (netDev NetDev) parseLine(rawLine string) (*NetDevLine, error) {
|
||||||
parts := strings.SplitN(rawLine, ":", 2)
|
idx := strings.LastIndex(rawLine, ":")
|
||||||
if len(parts) != 2 {
|
if idx == -1 {
|
||||||
return nil, errors.New("invalid net/dev line, missing colon")
|
return nil, errors.New("invalid net/dev line, missing colon")
|
||||||
}
|
}
|
||||||
fields := strings.Fields(strings.TrimSpace(parts[1]))
|
fields := strings.Fields(strings.TrimSpace(rawLine[idx+1:]))
|
||||||
|
|
||||||
var err error
|
var err error
|
||||||
line := &NetDevLine{}
|
line := &NetDevLine{}
|
||||||
|
|
||||||
// Interface Name
|
// Interface Name
|
||||||
line.Name = strings.TrimSpace(parts[0])
|
line.Name = strings.TrimSpace(rawLine[:idx])
|
||||||
if line.Name == "" {
|
if line.Name == "" {
|
||||||
return nil, errors.New("invalid net/dev line, empty interface name")
|
return nil, errors.New("invalid net/dev line, empty interface name")
|
||||||
}
|
}
|
||||||
|
2
vendor/github.com/prometheus/procfs/net_ip_socket.go
generated
vendored
2
vendor/github.com/prometheus/procfs/net_ip_socket.go
generated
vendored
@ -34,7 +34,7 @@ const (
|
|||||||
readLimit = 4294967296 // Byte -> 4 GiB
|
readLimit = 4294967296 // Byte -> 4 GiB
|
||||||
)
|
)
|
||||||
|
|
||||||
// this contains generic data structures for both udp and tcp sockets
|
// This contains generic data structures for both udp and tcp sockets.
|
||||||
type (
|
type (
|
||||||
// NetIPSocket represents the contents of /proc/net/{t,u}dp{,6} file without the header.
|
// NetIPSocket represents the contents of /proc/net/{t,u}dp{,6} file without the header.
|
||||||
NetIPSocket []*netIPSocketLine
|
NetIPSocket []*netIPSocketLine
|
||||||
|
4
vendor/github.com/prometheus/procfs/net_protocols.go
generated
vendored
4
vendor/github.com/prometheus/procfs/net_protocols.go
generated
vendored
@ -23,7 +23,7 @@ import (
|
|||||||
"github.com/prometheus/procfs/internal/util"
|
"github.com/prometheus/procfs/internal/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// NetProtocolStats stores the contents from /proc/net/protocols
|
// NetProtocolStats stores the contents from /proc/net/protocols.
|
||||||
type NetProtocolStats map[string]NetProtocolStatLine
|
type NetProtocolStats map[string]NetProtocolStatLine
|
||||||
|
|
||||||
// NetProtocolStatLine contains a single line parsed from /proc/net/protocols. We
|
// NetProtocolStatLine contains a single line parsed from /proc/net/protocols. We
|
||||||
@ -41,7 +41,7 @@ type NetProtocolStatLine struct {
|
|||||||
Capabilities NetProtocolCapabilities
|
Capabilities NetProtocolCapabilities
|
||||||
}
|
}
|
||||||
|
|
||||||
// NetProtocolCapabilities contains a list of capabilities for each protocol
|
// NetProtocolCapabilities contains a list of capabilities for each protocol.
|
||||||
type NetProtocolCapabilities struct {
|
type NetProtocolCapabilities struct {
|
||||||
Close bool // 8
|
Close bool // 8
|
||||||
Connect bool // 9
|
Connect bool // 9
|
||||||
|
8
vendor/github.com/prometheus/procfs/net_softnet.go
generated
vendored
8
vendor/github.com/prometheus/procfs/net_softnet.go
generated
vendored
@ -30,13 +30,13 @@ import (
|
|||||||
// * Linux 4.17 https://elixir.bootlin.com/linux/v4.17/source/net/core/net-procfs.c#L162
|
// * Linux 4.17 https://elixir.bootlin.com/linux/v4.17/source/net/core/net-procfs.c#L162
|
||||||
// and https://elixir.bootlin.com/linux/v4.17/source/include/linux/netdevice.h#L2810.
|
// and https://elixir.bootlin.com/linux/v4.17/source/include/linux/netdevice.h#L2810.
|
||||||
|
|
||||||
// SoftnetStat contains a single row of data from /proc/net/softnet_stat
|
// SoftnetStat contains a single row of data from /proc/net/softnet_stat.
|
||||||
type SoftnetStat struct {
|
type SoftnetStat struct {
|
||||||
// Number of processed packets
|
// Number of processed packets.
|
||||||
Processed uint32
|
Processed uint32
|
||||||
// Number of dropped packets
|
// Number of dropped packets.
|
||||||
Dropped uint32
|
Dropped uint32
|
||||||
// Number of times processing packets ran out of quota
|
// Number of times processing packets ran out of quota.
|
||||||
TimeSqueezed uint32
|
TimeSqueezed uint32
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -79,10 +79,13 @@ type XfrmStat struct {
|
|||||||
// Policy is dead
|
// Policy is dead
|
||||||
XfrmOutPolDead int
|
XfrmOutPolDead int
|
||||||
// Policy Error
|
// Policy Error
|
||||||
XfrmOutPolError int
|
XfrmOutPolError int
|
||||||
XfrmFwdHdrError int
|
// Forward routing of a packet is not allowed
|
||||||
|
XfrmFwdHdrError int
|
||||||
|
// State is invalid, perhaps expired
|
||||||
XfrmOutStateInvalid int
|
XfrmOutStateInvalid int
|
||||||
XfrmAcquireError int
|
// State hasn’t been fully acquired before use
|
||||||
|
XfrmAcquireError int
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewXfrmStat reads the xfrm_stat statistics.
|
// NewXfrmStat reads the xfrm_stat statistics.
|
8
vendor/github.com/prometheus/procfs/netstat.go
generated
vendored
8
vendor/github.com/prometheus/procfs/netstat.go
generated
vendored
@ -21,13 +21,13 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
// NetStat contains statistics for all the counters from one file
|
// NetStat contains statistics for all the counters from one file.
|
||||||
type NetStat struct {
|
type NetStat struct {
|
||||||
Filename string
|
|
||||||
Stats map[string][]uint64
|
Stats map[string][]uint64
|
||||||
|
Filename string
|
||||||
}
|
}
|
||||||
|
|
||||||
// NetStat retrieves stats from /proc/net/stat/
|
// NetStat retrieves stats from `/proc/net/stat/`.
|
||||||
func (fs FS) NetStat() ([]NetStat, error) {
|
func (fs FS) NetStat() ([]NetStat, error) {
|
||||||
statFiles, err := filepath.Glob(fs.proc.Path("net/stat/*"))
|
statFiles, err := filepath.Glob(fs.proc.Path("net/stat/*"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -55,7 +55,7 @@ func (fs FS) NetStat() ([]NetStat, error) {
|
|||||||
// Other strings represent per-CPU counters
|
// Other strings represent per-CPU counters
|
||||||
for scanner.Scan() {
|
for scanner.Scan() {
|
||||||
for num, counter := range strings.Fields(scanner.Text()) {
|
for num, counter := range strings.Fields(scanner.Text()) {
|
||||||
value, err := strconv.ParseUint(counter, 16, 32)
|
value, err := strconv.ParseUint(counter, 16, 64)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
10
vendor/github.com/prometheus/procfs/proc.go
generated
vendored
10
vendor/github.com/prometheus/procfs/proc.go
generated
vendored
@ -16,7 +16,7 @@ package procfs
|
|||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
"io"
|
||||||
"os"
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
@ -82,7 +82,7 @@ func (fs FS) Self() (Proc, error) {
|
|||||||
|
|
||||||
// NewProc returns a process for the given pid.
|
// NewProc returns a process for the given pid.
|
||||||
//
|
//
|
||||||
// Deprecated: use fs.Proc() instead
|
// Deprecated: Use fs.Proc() instead.
|
||||||
func (fs FS) NewProc(pid int) (Proc, error) {
|
func (fs FS) NewProc(pid int) (Proc, error) {
|
||||||
return fs.Proc(pid)
|
return fs.Proc(pid)
|
||||||
}
|
}
|
||||||
@ -142,7 +142,7 @@ func (p Proc) Wchan() (string, error) {
|
|||||||
}
|
}
|
||||||
defer f.Close()
|
defer f.Close()
|
||||||
|
|
||||||
data, err := ioutil.ReadAll(f)
|
data, err := io.ReadAll(f)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
@ -185,7 +185,7 @@ func (p Proc) Cwd() (string, error) {
|
|||||||
return wd, err
|
return wd, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// RootDir returns the absolute path to the process's root directory (as set by chroot)
|
// RootDir returns the absolute path to the process's root directory (as set by chroot).
|
||||||
func (p Proc) RootDir() (string, error) {
|
func (p Proc) RootDir() (string, error) {
|
||||||
rdir, err := os.Readlink(p.path("root"))
|
rdir, err := os.Readlink(p.path("root"))
|
||||||
if os.IsNotExist(err) {
|
if os.IsNotExist(err) {
|
||||||
@ -311,7 +311,7 @@ func (p Proc) FileDescriptorsInfo() (ProcFDInfos, error) {
|
|||||||
|
|
||||||
// Schedstat returns task scheduling information for the process.
|
// Schedstat returns task scheduling information for the process.
|
||||||
func (p Proc) Schedstat() (ProcSchedstat, error) {
|
func (p Proc) Schedstat() (ProcSchedstat, error) {
|
||||||
contents, err := ioutil.ReadFile(p.path("schedstat"))
|
contents, err := os.ReadFile(p.path("schedstat"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return ProcSchedstat{}, err
|
return ProcSchedstat{}, err
|
||||||
}
|
}
|
||||||
|
6
vendor/github.com/prometheus/procfs/proc_cgroup.go
generated
vendored
6
vendor/github.com/prometheus/procfs/proc_cgroup.go
generated
vendored
@ -45,7 +45,7 @@ type Cgroup struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// parseCgroupString parses each line of the /proc/[pid]/cgroup file
|
// parseCgroupString parses each line of the /proc/[pid]/cgroup file
|
||||||
// Line format is hierarchyID:[controller1,controller2]:path
|
// Line format is hierarchyID:[controller1,controller2]:path.
|
||||||
func parseCgroupString(cgroupStr string) (*Cgroup, error) {
|
func parseCgroupString(cgroupStr string) (*Cgroup, error) {
|
||||||
var err error
|
var err error
|
||||||
|
|
||||||
@ -69,7 +69,7 @@ func parseCgroupString(cgroupStr string) (*Cgroup, error) {
|
|||||||
return cgroup, nil
|
return cgroup, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// parseCgroups reads each line of the /proc/[pid]/cgroup file
|
// parseCgroups reads each line of the /proc/[pid]/cgroup file.
|
||||||
func parseCgroups(data []byte) ([]Cgroup, error) {
|
func parseCgroups(data []byte) ([]Cgroup, error) {
|
||||||
var cgroups []Cgroup
|
var cgroups []Cgroup
|
||||||
scanner := bufio.NewScanner(bytes.NewReader(data))
|
scanner := bufio.NewScanner(bytes.NewReader(data))
|
||||||
@ -88,7 +88,7 @@ func parseCgroups(data []byte) ([]Cgroup, error) {
|
|||||||
|
|
||||||
// Cgroups reads from /proc/<pid>/cgroups and returns a []*Cgroup struct locating this PID in each process
|
// Cgroups reads from /proc/<pid>/cgroups and returns a []*Cgroup struct locating this PID in each process
|
||||||
// control hierarchy running on this system. On every system (v1 and v2), all hierarchies contain all processes,
|
// control hierarchy running on this system. On every system (v1 and v2), all hierarchies contain all processes,
|
||||||
// so the len of the returned struct is equal to the number of active hierarchies on this system
|
// so the len of the returned struct is equal to the number of active hierarchies on this system.
|
||||||
func (p Proc) Cgroups() ([]Cgroup, error) {
|
func (p Proc) Cgroups() ([]Cgroup, error) {
|
||||||
data, err := util.ReadFileNoStat(p.path("cgroup"))
|
data, err := util.ReadFileNoStat(p.path("cgroup"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
98
vendor/github.com/prometheus/procfs/proc_cgroups.go
generated
vendored
Normal file
98
vendor/github.com/prometheus/procfs/proc_cgroups.go
generated
vendored
Normal file
@ -0,0 +1,98 @@
|
|||||||
|
// Copyright 2021 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package procfs
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/prometheus/procfs/internal/util"
|
||||||
|
)
|
||||||
|
|
||||||
|
// CgroupSummary models one line from /proc/cgroups.
|
||||||
|
// This file contains information about the controllers that are compiled into the kernel.
|
||||||
|
//
|
||||||
|
// Also see http://man7.org/linux/man-pages/man7/cgroups.7.html
|
||||||
|
type CgroupSummary struct {
|
||||||
|
// The name of the controller. controller is also known as subsystem.
|
||||||
|
SubsysName string
|
||||||
|
// The unique ID of the cgroup hierarchy on which this controller is mounted.
|
||||||
|
Hierarchy int
|
||||||
|
// The number of control groups in this hierarchy using this controller.
|
||||||
|
Cgroups int
|
||||||
|
// This field contains the value 1 if this controller is enabled, or 0 if it has been disabled
|
||||||
|
Enabled int
|
||||||
|
}
|
||||||
|
|
||||||
|
// parseCgroupSummary parses each line of the /proc/cgroup file
|
||||||
|
// Line format is `subsys_name hierarchy num_cgroups enabled`.
|
||||||
|
func parseCgroupSummaryString(CgroupSummaryStr string) (*CgroupSummary, error) {
|
||||||
|
var err error
|
||||||
|
|
||||||
|
fields := strings.Fields(CgroupSummaryStr)
|
||||||
|
// require at least 4 fields
|
||||||
|
if len(fields) < 4 {
|
||||||
|
return nil, fmt.Errorf("at least 4 fields required, found %d fields in cgroup info string: %s", len(fields), CgroupSummaryStr)
|
||||||
|
}
|
||||||
|
|
||||||
|
CgroupSummary := &CgroupSummary{
|
||||||
|
SubsysName: fields[0],
|
||||||
|
}
|
||||||
|
CgroupSummary.Hierarchy, err = strconv.Atoi(fields[1])
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to parse hierarchy ID")
|
||||||
|
}
|
||||||
|
CgroupSummary.Cgroups, err = strconv.Atoi(fields[2])
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to parse Cgroup Num")
|
||||||
|
}
|
||||||
|
CgroupSummary.Enabled, err = strconv.Atoi(fields[3])
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to parse Enabled")
|
||||||
|
}
|
||||||
|
return CgroupSummary, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// parseCgroupSummary reads each line of the /proc/cgroup file.
|
||||||
|
func parseCgroupSummary(data []byte) ([]CgroupSummary, error) {
|
||||||
|
var CgroupSummarys []CgroupSummary
|
||||||
|
scanner := bufio.NewScanner(bytes.NewReader(data))
|
||||||
|
for scanner.Scan() {
|
||||||
|
CgroupSummaryString := scanner.Text()
|
||||||
|
// ignore comment lines
|
||||||
|
if strings.HasPrefix(CgroupSummaryString, "#") {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
CgroupSummary, err := parseCgroupSummaryString(CgroupSummaryString)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
CgroupSummarys = append(CgroupSummarys, *CgroupSummary)
|
||||||
|
}
|
||||||
|
|
||||||
|
err := scanner.Err()
|
||||||
|
return CgroupSummarys, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// CgroupSummarys returns information about current /proc/cgroups.
|
||||||
|
func (fs FS) CgroupSummarys() ([]CgroupSummary, error) {
|
||||||
|
data, err := util.ReadFileNoStat(fs.proc.Path("cgroups"))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return parseCgroupSummary(data)
|
||||||
|
}
|
2
vendor/github.com/prometheus/procfs/proc_environ.go
generated
vendored
2
vendor/github.com/prometheus/procfs/proc_environ.go
generated
vendored
@ -19,7 +19,7 @@ import (
|
|||||||
"github.com/prometheus/procfs/internal/util"
|
"github.com/prometheus/procfs/internal/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Environ reads process environments from /proc/<pid>/environ
|
// Environ reads process environments from `/proc/<pid>/environ`.
|
||||||
func (p Proc) Environ() ([]string, error) {
|
func (p Proc) Environ() ([]string, error) {
|
||||||
environments := make([]string, 0)
|
environments := make([]string, 0)
|
||||||
|
|
||||||
|
3
vendor/github.com/prometheus/procfs/proc_fdinfo.go
generated
vendored
3
vendor/github.com/prometheus/procfs/proc_fdinfo.go
generated
vendored
@ -22,7 +22,6 @@ import (
|
|||||||
"github.com/prometheus/procfs/internal/util"
|
"github.com/prometheus/procfs/internal/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Regexp variables
|
|
||||||
var (
|
var (
|
||||||
rPos = regexp.MustCompile(`^pos:\s+(\d+)$`)
|
rPos = regexp.MustCompile(`^pos:\s+(\d+)$`)
|
||||||
rFlags = regexp.MustCompile(`^flags:\s+(\d+)$`)
|
rFlags = regexp.MustCompile(`^flags:\s+(\d+)$`)
|
||||||
@ -122,7 +121,7 @@ func (p ProcFDInfos) Len() int { return len(p) }
|
|||||||
func (p ProcFDInfos) Swap(i, j int) { p[i], p[j] = p[j], p[i] }
|
func (p ProcFDInfos) Swap(i, j int) { p[i], p[j] = p[j], p[i] }
|
||||||
func (p ProcFDInfos) Less(i, j int) bool { return p[i].FD < p[j].FD }
|
func (p ProcFDInfos) Less(i, j int) bool { return p[i].FD < p[j].FD }
|
||||||
|
|
||||||
// InotifyWatchLen returns the total number of inotify watches
|
// InotifyWatchLen returns the total number of inotify watches.
|
||||||
func (p ProcFDInfos) InotifyWatchLen() (int, error) {
|
func (p ProcFDInfos) InotifyWatchLen() (int, error) {
|
||||||
length := 0
|
length := 0
|
||||||
for _, f := range p {
|
for _, f := range p {
|
||||||
|
2
vendor/github.com/prometheus/procfs/proc_limits.go
generated
vendored
2
vendor/github.com/prometheus/procfs/proc_limits.go
generated
vendored
@ -79,7 +79,7 @@ var (
|
|||||||
|
|
||||||
// NewLimits returns the current soft limits of the process.
|
// NewLimits returns the current soft limits of the process.
|
||||||
//
|
//
|
||||||
// Deprecated: use p.Limits() instead
|
// Deprecated: Use p.Limits() instead.
|
||||||
func (p Proc) NewLimits() (ProcLimits, error) {
|
func (p Proc) NewLimits() (ProcLimits, error) {
|
||||||
return p.Limits()
|
return p.Limits()
|
||||||
}
|
}
|
||||||
|
12
vendor/github.com/prometheus/procfs/proc_maps.go
generated
vendored
12
vendor/github.com/prometheus/procfs/proc_maps.go
generated
vendored
@ -11,7 +11,9 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build (aix || darwin || dragonfly || freebsd || linux || netbsd || openbsd || solaris) && !js
|
||||||
// +build aix darwin dragonfly freebsd linux netbsd openbsd solaris
|
// +build aix darwin dragonfly freebsd linux netbsd openbsd solaris
|
||||||
|
// +build !js
|
||||||
|
|
||||||
package procfs
|
package procfs
|
||||||
|
|
||||||
@ -25,7 +27,7 @@ import (
|
|||||||
"golang.org/x/sys/unix"
|
"golang.org/x/sys/unix"
|
||||||
)
|
)
|
||||||
|
|
||||||
// ProcMapPermissions contains permission settings read from /proc/[pid]/maps
|
// ProcMapPermissions contains permission settings read from `/proc/[pid]/maps`.
|
||||||
type ProcMapPermissions struct {
|
type ProcMapPermissions struct {
|
||||||
// mapping has the [R]ead flag set
|
// mapping has the [R]ead flag set
|
||||||
Read bool
|
Read bool
|
||||||
@ -39,8 +41,8 @@ type ProcMapPermissions struct {
|
|||||||
Private bool
|
Private bool
|
||||||
}
|
}
|
||||||
|
|
||||||
// ProcMap contains the process memory-mappings of the process,
|
// ProcMap contains the process memory-mappings of the process
|
||||||
// read from /proc/[pid]/maps
|
// read from `/proc/[pid]/maps`.
|
||||||
type ProcMap struct {
|
type ProcMap struct {
|
||||||
// The start address of current mapping.
|
// The start address of current mapping.
|
||||||
StartAddr uintptr
|
StartAddr uintptr
|
||||||
@ -79,7 +81,7 @@ func parseDevice(s string) (uint64, error) {
|
|||||||
return unix.Mkdev(uint32(major), uint32(minor)), nil
|
return unix.Mkdev(uint32(major), uint32(minor)), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// parseAddress just converts a hex-string to a uintptr
|
// parseAddress converts a hex-string to a uintptr.
|
||||||
func parseAddress(s string) (uintptr, error) {
|
func parseAddress(s string) (uintptr, error) {
|
||||||
a, err := strconv.ParseUint(s, 16, 0)
|
a, err := strconv.ParseUint(s, 16, 0)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -89,7 +91,7 @@ func parseAddress(s string) (uintptr, error) {
|
|||||||
return uintptr(a), nil
|
return uintptr(a), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// parseAddresses parses the start-end address
|
// parseAddresses parses the start-end address.
|
||||||
func parseAddresses(s string) (uintptr, uintptr, error) {
|
func parseAddresses(s string) (uintptr, uintptr, error) {
|
||||||
toks := strings.Split(s, "-")
|
toks := strings.Split(s, "-")
|
||||||
if len(toks) < 2 {
|
if len(toks) < 2 {
|
||||||
|
440
vendor/github.com/prometheus/procfs/proc_netstat.go
generated
vendored
Normal file
440
vendor/github.com/prometheus/procfs/proc_netstat.go
generated
vendored
Normal file
@ -0,0 +1,440 @@
|
|||||||
|
// Copyright 2022 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package procfs
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/prometheus/procfs/internal/util"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ProcNetstat models the content of /proc/<pid>/net/netstat.
|
||||||
|
type ProcNetstat struct {
|
||||||
|
// The process ID.
|
||||||
|
PID int
|
||||||
|
TcpExt
|
||||||
|
IpExt
|
||||||
|
}
|
||||||
|
|
||||||
|
type TcpExt struct { // nolint:revive
|
||||||
|
SyncookiesSent float64
|
||||||
|
SyncookiesRecv float64
|
||||||
|
SyncookiesFailed float64
|
||||||
|
EmbryonicRsts float64
|
||||||
|
PruneCalled float64
|
||||||
|
RcvPruned float64
|
||||||
|
OfoPruned float64
|
||||||
|
OutOfWindowIcmps float64
|
||||||
|
LockDroppedIcmps float64
|
||||||
|
ArpFilter float64
|
||||||
|
TW float64
|
||||||
|
TWRecycled float64
|
||||||
|
TWKilled float64
|
||||||
|
PAWSActive float64
|
||||||
|
PAWSEstab float64
|
||||||
|
DelayedACKs float64
|
||||||
|
DelayedACKLocked float64
|
||||||
|
DelayedACKLost float64
|
||||||
|
ListenOverflows float64
|
||||||
|
ListenDrops float64
|
||||||
|
TCPHPHits float64
|
||||||
|
TCPPureAcks float64
|
||||||
|
TCPHPAcks float64
|
||||||
|
TCPRenoRecovery float64
|
||||||
|
TCPSackRecovery float64
|
||||||
|
TCPSACKReneging float64
|
||||||
|
TCPSACKReorder float64
|
||||||
|
TCPRenoReorder float64
|
||||||
|
TCPTSReorder float64
|
||||||
|
TCPFullUndo float64
|
||||||
|
TCPPartialUndo float64
|
||||||
|
TCPDSACKUndo float64
|
||||||
|
TCPLossUndo float64
|
||||||
|
TCPLostRetransmit float64
|
||||||
|
TCPRenoFailures float64
|
||||||
|
TCPSackFailures float64
|
||||||
|
TCPLossFailures float64
|
||||||
|
TCPFastRetrans float64
|
||||||
|
TCPSlowStartRetrans float64
|
||||||
|
TCPTimeouts float64
|
||||||
|
TCPLossProbes float64
|
||||||
|
TCPLossProbeRecovery float64
|
||||||
|
TCPRenoRecoveryFail float64
|
||||||
|
TCPSackRecoveryFail float64
|
||||||
|
TCPRcvCollapsed float64
|
||||||
|
TCPDSACKOldSent float64
|
||||||
|
TCPDSACKOfoSent float64
|
||||||
|
TCPDSACKRecv float64
|
||||||
|
TCPDSACKOfoRecv float64
|
||||||
|
TCPAbortOnData float64
|
||||||
|
TCPAbortOnClose float64
|
||||||
|
TCPAbortOnMemory float64
|
||||||
|
TCPAbortOnTimeout float64
|
||||||
|
TCPAbortOnLinger float64
|
||||||
|
TCPAbortFailed float64
|
||||||
|
TCPMemoryPressures float64
|
||||||
|
TCPMemoryPressuresChrono float64
|
||||||
|
TCPSACKDiscard float64
|
||||||
|
TCPDSACKIgnoredOld float64
|
||||||
|
TCPDSACKIgnoredNoUndo float64
|
||||||
|
TCPSpuriousRTOs float64
|
||||||
|
TCPMD5NotFound float64
|
||||||
|
TCPMD5Unexpected float64
|
||||||
|
TCPMD5Failure float64
|
||||||
|
TCPSackShifted float64
|
||||||
|
TCPSackMerged float64
|
||||||
|
TCPSackShiftFallback float64
|
||||||
|
TCPBacklogDrop float64
|
||||||
|
PFMemallocDrop float64
|
||||||
|
TCPMinTTLDrop float64
|
||||||
|
TCPDeferAcceptDrop float64
|
||||||
|
IPReversePathFilter float64
|
||||||
|
TCPTimeWaitOverflow float64
|
||||||
|
TCPReqQFullDoCookies float64
|
||||||
|
TCPReqQFullDrop float64
|
||||||
|
TCPRetransFail float64
|
||||||
|
TCPRcvCoalesce float64
|
||||||
|
TCPOFOQueue float64
|
||||||
|
TCPOFODrop float64
|
||||||
|
TCPOFOMerge float64
|
||||||
|
TCPChallengeACK float64
|
||||||
|
TCPSYNChallenge float64
|
||||||
|
TCPFastOpenActive float64
|
||||||
|
TCPFastOpenActiveFail float64
|
||||||
|
TCPFastOpenPassive float64
|
||||||
|
TCPFastOpenPassiveFail float64
|
||||||
|
TCPFastOpenListenOverflow float64
|
||||||
|
TCPFastOpenCookieReqd float64
|
||||||
|
TCPFastOpenBlackhole float64
|
||||||
|
TCPSpuriousRtxHostQueues float64
|
||||||
|
BusyPollRxPackets float64
|
||||||
|
TCPAutoCorking float64
|
||||||
|
TCPFromZeroWindowAdv float64
|
||||||
|
TCPToZeroWindowAdv float64
|
||||||
|
TCPWantZeroWindowAdv float64
|
||||||
|
TCPSynRetrans float64
|
||||||
|
TCPOrigDataSent float64
|
||||||
|
TCPHystartTrainDetect float64
|
||||||
|
TCPHystartTrainCwnd float64
|
||||||
|
TCPHystartDelayDetect float64
|
||||||
|
TCPHystartDelayCwnd float64
|
||||||
|
TCPACKSkippedSynRecv float64
|
||||||
|
TCPACKSkippedPAWS float64
|
||||||
|
TCPACKSkippedSeq float64
|
||||||
|
TCPACKSkippedFinWait2 float64
|
||||||
|
TCPACKSkippedTimeWait float64
|
||||||
|
TCPACKSkippedChallenge float64
|
||||||
|
TCPWinProbe float64
|
||||||
|
TCPKeepAlive float64
|
||||||
|
TCPMTUPFail float64
|
||||||
|
TCPMTUPSuccess float64
|
||||||
|
TCPWqueueTooBig float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type IpExt struct { // nolint:revive
|
||||||
|
InNoRoutes float64
|
||||||
|
InTruncatedPkts float64
|
||||||
|
InMcastPkts float64
|
||||||
|
OutMcastPkts float64
|
||||||
|
InBcastPkts float64
|
||||||
|
OutBcastPkts float64
|
||||||
|
InOctets float64
|
||||||
|
OutOctets float64
|
||||||
|
InMcastOctets float64
|
||||||
|
OutMcastOctets float64
|
||||||
|
InBcastOctets float64
|
||||||
|
OutBcastOctets float64
|
||||||
|
InCsumErrors float64
|
||||||
|
InNoECTPkts float64
|
||||||
|
InECT1Pkts float64
|
||||||
|
InECT0Pkts float64
|
||||||
|
InCEPkts float64
|
||||||
|
ReasmOverlaps float64
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p Proc) Netstat() (ProcNetstat, error) {
|
||||||
|
filename := p.path("net/netstat")
|
||||||
|
data, err := util.ReadFileNoStat(filename)
|
||||||
|
if err != nil {
|
||||||
|
return ProcNetstat{PID: p.PID}, err
|
||||||
|
}
|
||||||
|
procNetstat, err := parseNetstat(bytes.NewReader(data), filename)
|
||||||
|
procNetstat.PID = p.PID
|
||||||
|
return procNetstat, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// parseNetstat parses the metrics from proc/<pid>/net/netstat file
|
||||||
|
// and returns a ProcNetstat structure.
|
||||||
|
func parseNetstat(r io.Reader, fileName string) (ProcNetstat, error) {
|
||||||
|
var (
|
||||||
|
scanner = bufio.NewScanner(r)
|
||||||
|
procNetstat = ProcNetstat{}
|
||||||
|
)
|
||||||
|
|
||||||
|
for scanner.Scan() {
|
||||||
|
nameParts := strings.Split(scanner.Text(), " ")
|
||||||
|
scanner.Scan()
|
||||||
|
valueParts := strings.Split(scanner.Text(), " ")
|
||||||
|
// Remove trailing :.
|
||||||
|
protocol := strings.TrimSuffix(nameParts[0], ":")
|
||||||
|
if len(nameParts) != len(valueParts) {
|
||||||
|
return procNetstat, fmt.Errorf("mismatch field count mismatch in %s: %s",
|
||||||
|
fileName, protocol)
|
||||||
|
}
|
||||||
|
for i := 1; i < len(nameParts); i++ {
|
||||||
|
value, err := strconv.ParseFloat(valueParts[i], 64)
|
||||||
|
if err != nil {
|
||||||
|
return procNetstat, err
|
||||||
|
}
|
||||||
|
key := nameParts[i]
|
||||||
|
|
||||||
|
switch protocol {
|
||||||
|
case "TcpExt":
|
||||||
|
switch key {
|
||||||
|
case "SyncookiesSent":
|
||||||
|
procNetstat.TcpExt.SyncookiesSent = value
|
||||||
|
case "SyncookiesRecv":
|
||||||
|
procNetstat.TcpExt.SyncookiesRecv = value
|
||||||
|
case "SyncookiesFailed":
|
||||||
|
procNetstat.TcpExt.SyncookiesFailed = value
|
||||||
|
case "EmbryonicRsts":
|
||||||
|
procNetstat.TcpExt.EmbryonicRsts = value
|
||||||
|
case "PruneCalled":
|
||||||
|
procNetstat.TcpExt.PruneCalled = value
|
||||||
|
case "RcvPruned":
|
||||||
|
procNetstat.TcpExt.RcvPruned = value
|
||||||
|
case "OfoPruned":
|
||||||
|
procNetstat.TcpExt.OfoPruned = value
|
||||||
|
case "OutOfWindowIcmps":
|
||||||
|
procNetstat.TcpExt.OutOfWindowIcmps = value
|
||||||
|
case "LockDroppedIcmps":
|
||||||
|
procNetstat.TcpExt.LockDroppedIcmps = value
|
||||||
|
case "ArpFilter":
|
||||||
|
procNetstat.TcpExt.ArpFilter = value
|
||||||
|
case "TW":
|
||||||
|
procNetstat.TcpExt.TW = value
|
||||||
|
case "TWRecycled":
|
||||||
|
procNetstat.TcpExt.TWRecycled = value
|
||||||
|
case "TWKilled":
|
||||||
|
procNetstat.TcpExt.TWKilled = value
|
||||||
|
case "PAWSActive":
|
||||||
|
procNetstat.TcpExt.PAWSActive = value
|
||||||
|
case "PAWSEstab":
|
||||||
|
procNetstat.TcpExt.PAWSEstab = value
|
||||||
|
case "DelayedACKs":
|
||||||
|
procNetstat.TcpExt.DelayedACKs = value
|
||||||
|
case "DelayedACKLocked":
|
||||||
|
procNetstat.TcpExt.DelayedACKLocked = value
|
||||||
|
case "DelayedACKLost":
|
||||||
|
procNetstat.TcpExt.DelayedACKLost = value
|
||||||
|
case "ListenOverflows":
|
||||||
|
procNetstat.TcpExt.ListenOverflows = value
|
||||||
|
case "ListenDrops":
|
||||||
|
procNetstat.TcpExt.ListenDrops = value
|
||||||
|
case "TCPHPHits":
|
||||||
|
procNetstat.TcpExt.TCPHPHits = value
|
||||||
|
case "TCPPureAcks":
|
||||||
|
procNetstat.TcpExt.TCPPureAcks = value
|
||||||
|
case "TCPHPAcks":
|
||||||
|
procNetstat.TcpExt.TCPHPAcks = value
|
||||||
|
case "TCPRenoRecovery":
|
||||||
|
procNetstat.TcpExt.TCPRenoRecovery = value
|
||||||
|
case "TCPSackRecovery":
|
||||||
|
procNetstat.TcpExt.TCPSackRecovery = value
|
||||||
|
case "TCPSACKReneging":
|
||||||
|
procNetstat.TcpExt.TCPSACKReneging = value
|
||||||
|
case "TCPSACKReorder":
|
||||||
|
procNetstat.TcpExt.TCPSACKReorder = value
|
||||||
|
case "TCPRenoReorder":
|
||||||
|
procNetstat.TcpExt.TCPRenoReorder = value
|
||||||
|
case "TCPTSReorder":
|
||||||
|
procNetstat.TcpExt.TCPTSReorder = value
|
||||||
|
case "TCPFullUndo":
|
||||||
|
procNetstat.TcpExt.TCPFullUndo = value
|
||||||
|
case "TCPPartialUndo":
|
||||||
|
procNetstat.TcpExt.TCPPartialUndo = value
|
||||||
|
case "TCPDSACKUndo":
|
||||||
|
procNetstat.TcpExt.TCPDSACKUndo = value
|
||||||
|
case "TCPLossUndo":
|
||||||
|
procNetstat.TcpExt.TCPLossUndo = value
|
||||||
|
case "TCPLostRetransmit":
|
||||||
|
procNetstat.TcpExt.TCPLostRetransmit = value
|
||||||
|
case "TCPRenoFailures":
|
||||||
|
procNetstat.TcpExt.TCPRenoFailures = value
|
||||||
|
case "TCPSackFailures":
|
||||||
|
procNetstat.TcpExt.TCPSackFailures = value
|
||||||
|
case "TCPLossFailures":
|
||||||
|
procNetstat.TcpExt.TCPLossFailures = value
|
||||||
|
case "TCPFastRetrans":
|
||||||
|
procNetstat.TcpExt.TCPFastRetrans = value
|
||||||
|
case "TCPSlowStartRetrans":
|
||||||
|
procNetstat.TcpExt.TCPSlowStartRetrans = value
|
||||||
|
case "TCPTimeouts":
|
||||||
|
procNetstat.TcpExt.TCPTimeouts = value
|
||||||
|
case "TCPLossProbes":
|
||||||
|
procNetstat.TcpExt.TCPLossProbes = value
|
||||||
|
case "TCPLossProbeRecovery":
|
||||||
|
procNetstat.TcpExt.TCPLossProbeRecovery = value
|
||||||
|
case "TCPRenoRecoveryFail":
|
||||||
|
procNetstat.TcpExt.TCPRenoRecoveryFail = value
|
||||||
|
case "TCPSackRecoveryFail":
|
||||||
|
procNetstat.TcpExt.TCPSackRecoveryFail = value
|
||||||
|
case "TCPRcvCollapsed":
|
||||||
|
procNetstat.TcpExt.TCPRcvCollapsed = value
|
||||||
|
case "TCPDSACKOldSent":
|
||||||
|
procNetstat.TcpExt.TCPDSACKOldSent = value
|
||||||
|
case "TCPDSACKOfoSent":
|
||||||
|
procNetstat.TcpExt.TCPDSACKOfoSent = value
|
||||||
|
case "TCPDSACKRecv":
|
||||||
|
procNetstat.TcpExt.TCPDSACKRecv = value
|
||||||
|
case "TCPDSACKOfoRecv":
|
||||||
|
procNetstat.TcpExt.TCPDSACKOfoRecv = value
|
||||||
|
case "TCPAbortOnData":
|
||||||
|
procNetstat.TcpExt.TCPAbortOnData = value
|
||||||
|
case "TCPAbortOnClose":
|
||||||
|
procNetstat.TcpExt.TCPAbortOnClose = value
|
||||||
|
case "TCPDeferAcceptDrop":
|
||||||
|
procNetstat.TcpExt.TCPDeferAcceptDrop = value
|
||||||
|
case "IPReversePathFilter":
|
||||||
|
procNetstat.TcpExt.IPReversePathFilter = value
|
||||||
|
case "TCPTimeWaitOverflow":
|
||||||
|
procNetstat.TcpExt.TCPTimeWaitOverflow = value
|
||||||
|
case "TCPReqQFullDoCookies":
|
||||||
|
procNetstat.TcpExt.TCPReqQFullDoCookies = value
|
||||||
|
case "TCPReqQFullDrop":
|
||||||
|
procNetstat.TcpExt.TCPReqQFullDrop = value
|
||||||
|
case "TCPRetransFail":
|
||||||
|
procNetstat.TcpExt.TCPRetransFail = value
|
||||||
|
case "TCPRcvCoalesce":
|
||||||
|
procNetstat.TcpExt.TCPRcvCoalesce = value
|
||||||
|
case "TCPOFOQueue":
|
||||||
|
procNetstat.TcpExt.TCPOFOQueue = value
|
||||||
|
case "TCPOFODrop":
|
||||||
|
procNetstat.TcpExt.TCPOFODrop = value
|
||||||
|
case "TCPOFOMerge":
|
||||||
|
procNetstat.TcpExt.TCPOFOMerge = value
|
||||||
|
case "TCPChallengeACK":
|
||||||
|
procNetstat.TcpExt.TCPChallengeACK = value
|
||||||
|
case "TCPSYNChallenge":
|
||||||
|
procNetstat.TcpExt.TCPSYNChallenge = value
|
||||||
|
case "TCPFastOpenActive":
|
||||||
|
procNetstat.TcpExt.TCPFastOpenActive = value
|
||||||
|
case "TCPFastOpenActiveFail":
|
||||||
|
procNetstat.TcpExt.TCPFastOpenActiveFail = value
|
||||||
|
case "TCPFastOpenPassive":
|
||||||
|
procNetstat.TcpExt.TCPFastOpenPassive = value
|
||||||
|
case "TCPFastOpenPassiveFail":
|
||||||
|
procNetstat.TcpExt.TCPFastOpenPassiveFail = value
|
||||||
|
case "TCPFastOpenListenOverflow":
|
||||||
|
procNetstat.TcpExt.TCPFastOpenListenOverflow = value
|
||||||
|
case "TCPFastOpenCookieReqd":
|
||||||
|
procNetstat.TcpExt.TCPFastOpenCookieReqd = value
|
||||||
|
case "TCPFastOpenBlackhole":
|
||||||
|
procNetstat.TcpExt.TCPFastOpenBlackhole = value
|
||||||
|
case "TCPSpuriousRtxHostQueues":
|
||||||
|
procNetstat.TcpExt.TCPSpuriousRtxHostQueues = value
|
||||||
|
case "BusyPollRxPackets":
|
||||||
|
procNetstat.TcpExt.BusyPollRxPackets = value
|
||||||
|
case "TCPAutoCorking":
|
||||||
|
procNetstat.TcpExt.TCPAutoCorking = value
|
||||||
|
case "TCPFromZeroWindowAdv":
|
||||||
|
procNetstat.TcpExt.TCPFromZeroWindowAdv = value
|
||||||
|
case "TCPToZeroWindowAdv":
|
||||||
|
procNetstat.TcpExt.TCPToZeroWindowAdv = value
|
||||||
|
case "TCPWantZeroWindowAdv":
|
||||||
|
procNetstat.TcpExt.TCPWantZeroWindowAdv = value
|
||||||
|
case "TCPSynRetrans":
|
||||||
|
procNetstat.TcpExt.TCPSynRetrans = value
|
||||||
|
case "TCPOrigDataSent":
|
||||||
|
procNetstat.TcpExt.TCPOrigDataSent = value
|
||||||
|
case "TCPHystartTrainDetect":
|
||||||
|
procNetstat.TcpExt.TCPHystartTrainDetect = value
|
||||||
|
case "TCPHystartTrainCwnd":
|
||||||
|
procNetstat.TcpExt.TCPHystartTrainCwnd = value
|
||||||
|
case "TCPHystartDelayDetect":
|
||||||
|
procNetstat.TcpExt.TCPHystartDelayDetect = value
|
||||||
|
case "TCPHystartDelayCwnd":
|
||||||
|
procNetstat.TcpExt.TCPHystartDelayCwnd = value
|
||||||
|
case "TCPACKSkippedSynRecv":
|
||||||
|
procNetstat.TcpExt.TCPACKSkippedSynRecv = value
|
||||||
|
case "TCPACKSkippedPAWS":
|
||||||
|
procNetstat.TcpExt.TCPACKSkippedPAWS = value
|
||||||
|
case "TCPACKSkippedSeq":
|
||||||
|
procNetstat.TcpExt.TCPACKSkippedSeq = value
|
||||||
|
case "TCPACKSkippedFinWait2":
|
||||||
|
procNetstat.TcpExt.TCPACKSkippedFinWait2 = value
|
||||||
|
case "TCPACKSkippedTimeWait":
|
||||||
|
procNetstat.TcpExt.TCPACKSkippedTimeWait = value
|
||||||
|
case "TCPACKSkippedChallenge":
|
||||||
|
procNetstat.TcpExt.TCPACKSkippedChallenge = value
|
||||||
|
case "TCPWinProbe":
|
||||||
|
procNetstat.TcpExt.TCPWinProbe = value
|
||||||
|
case "TCPKeepAlive":
|
||||||
|
procNetstat.TcpExt.TCPKeepAlive = value
|
||||||
|
case "TCPMTUPFail":
|
||||||
|
procNetstat.TcpExt.TCPMTUPFail = value
|
||||||
|
case "TCPMTUPSuccess":
|
||||||
|
procNetstat.TcpExt.TCPMTUPSuccess = value
|
||||||
|
case "TCPWqueueTooBig":
|
||||||
|
procNetstat.TcpExt.TCPWqueueTooBig = value
|
||||||
|
}
|
||||||
|
case "IpExt":
|
||||||
|
switch key {
|
||||||
|
case "InNoRoutes":
|
||||||
|
procNetstat.IpExt.InNoRoutes = value
|
||||||
|
case "InTruncatedPkts":
|
||||||
|
procNetstat.IpExt.InTruncatedPkts = value
|
||||||
|
case "InMcastPkts":
|
||||||
|
procNetstat.IpExt.InMcastPkts = value
|
||||||
|
case "OutMcastPkts":
|
||||||
|
procNetstat.IpExt.OutMcastPkts = value
|
||||||
|
case "InBcastPkts":
|
||||||
|
procNetstat.IpExt.InBcastPkts = value
|
||||||
|
case "OutBcastPkts":
|
||||||
|
procNetstat.IpExt.OutBcastPkts = value
|
||||||
|
case "InOctets":
|
||||||
|
procNetstat.IpExt.InOctets = value
|
||||||
|
case "OutOctets":
|
||||||
|
procNetstat.IpExt.OutOctets = value
|
||||||
|
case "InMcastOctets":
|
||||||
|
procNetstat.IpExt.InMcastOctets = value
|
||||||
|
case "OutMcastOctets":
|
||||||
|
procNetstat.IpExt.OutMcastOctets = value
|
||||||
|
case "InBcastOctets":
|
||||||
|
procNetstat.IpExt.InBcastOctets = value
|
||||||
|
case "OutBcastOctets":
|
||||||
|
procNetstat.IpExt.OutBcastOctets = value
|
||||||
|
case "InCsumErrors":
|
||||||
|
procNetstat.IpExt.InCsumErrors = value
|
||||||
|
case "InNoECTPkts":
|
||||||
|
procNetstat.IpExt.InNoECTPkts = value
|
||||||
|
case "InECT1Pkts":
|
||||||
|
procNetstat.IpExt.InECT1Pkts = value
|
||||||
|
case "InECT0Pkts":
|
||||||
|
procNetstat.IpExt.InECT0Pkts = value
|
||||||
|
case "InCEPkts":
|
||||||
|
procNetstat.IpExt.InCEPkts = value
|
||||||
|
case "ReasmOverlaps":
|
||||||
|
procNetstat.IpExt.ReasmOverlaps = value
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return procNetstat, scanner.Err()
|
||||||
|
}
|
14
vendor/github.com/prometheus/procfs/proc_psi.go
generated
vendored
14
vendor/github.com/prometheus/procfs/proc_psi.go
generated
vendored
@ -35,9 +35,10 @@ import (
|
|||||||
|
|
||||||
const lineFormat = "avg10=%f avg60=%f avg300=%f total=%d"
|
const lineFormat = "avg10=%f avg60=%f avg300=%f total=%d"
|
||||||
|
|
||||||
// PSILine is a single line of values as returned by /proc/pressure/*
|
// PSILine is a single line of values as returned by `/proc/pressure/*`.
|
||||||
// The Avg entries are averages over n seconds, as a percentage
|
//
|
||||||
// The Total line is in microseconds
|
// The Avg entries are averages over n seconds, as a percentage.
|
||||||
|
// The Total line is in microseconds.
|
||||||
type PSILine struct {
|
type PSILine struct {
|
||||||
Avg10 float64
|
Avg10 float64
|
||||||
Avg60 float64
|
Avg60 float64
|
||||||
@ -46,8 +47,9 @@ type PSILine struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// PSIStats represent pressure stall information from /proc/pressure/*
|
// PSIStats represent pressure stall information from /proc/pressure/*
|
||||||
// Some indicates the share of time in which at least some tasks are stalled
|
//
|
||||||
// Full indicates the share of time in which all non-idle tasks are stalled simultaneously
|
// "Some" indicates the share of time in which at least some tasks are stalled.
|
||||||
|
// "Full" indicates the share of time in which all non-idle tasks are stalled simultaneously.
|
||||||
type PSIStats struct {
|
type PSIStats struct {
|
||||||
Some *PSILine
|
Some *PSILine
|
||||||
Full *PSILine
|
Full *PSILine
|
||||||
@ -65,7 +67,7 @@ func (fs FS) PSIStatsForResource(resource string) (PSIStats, error) {
|
|||||||
return parsePSIStats(resource, bytes.NewReader(data))
|
return parsePSIStats(resource, bytes.NewReader(data))
|
||||||
}
|
}
|
||||||
|
|
||||||
// parsePSIStats parses the specified file for pressure stall information
|
// parsePSIStats parses the specified file for pressure stall information.
|
||||||
func parsePSIStats(resource string, r io.Reader) (PSIStats, error) {
|
func parsePSIStats(resource string, r io.Reader) (PSIStats, error) {
|
||||||
psiStats := PSIStats{}
|
psiStats := PSIStats{}
|
||||||
|
|
||||||
|
23
vendor/github.com/prometheus/procfs/proc_smaps.go
generated
vendored
23
vendor/github.com/prometheus/procfs/proc_smaps.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build !windows
|
||||||
// +build !windows
|
// +build !windows
|
||||||
|
|
||||||
package procfs
|
package procfs
|
||||||
@ -28,30 +29,30 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
// match the header line before each mapped zone in /proc/pid/smaps
|
// match the header line before each mapped zone in `/proc/pid/smaps`.
|
||||||
procSMapsHeaderLine = regexp.MustCompile(`^[a-f0-9].*$`)
|
procSMapsHeaderLine = regexp.MustCompile(`^[a-f0-9].*$`)
|
||||||
)
|
)
|
||||||
|
|
||||||
type ProcSMapsRollup struct {
|
type ProcSMapsRollup struct {
|
||||||
// Amount of the mapping that is currently resident in RAM
|
// Amount of the mapping that is currently resident in RAM.
|
||||||
Rss uint64
|
Rss uint64
|
||||||
// Process's proportional share of this mapping
|
// Process's proportional share of this mapping.
|
||||||
Pss uint64
|
Pss uint64
|
||||||
// Size in bytes of clean shared pages
|
// Size in bytes of clean shared pages.
|
||||||
SharedClean uint64
|
SharedClean uint64
|
||||||
// Size in bytes of dirty shared pages
|
// Size in bytes of dirty shared pages.
|
||||||
SharedDirty uint64
|
SharedDirty uint64
|
||||||
// Size in bytes of clean private pages
|
// Size in bytes of clean private pages.
|
||||||
PrivateClean uint64
|
PrivateClean uint64
|
||||||
// Size in bytes of dirty private pages
|
// Size in bytes of dirty private pages.
|
||||||
PrivateDirty uint64
|
PrivateDirty uint64
|
||||||
// Amount of memory currently marked as referenced or accessed
|
// Amount of memory currently marked as referenced or accessed.
|
||||||
Referenced uint64
|
Referenced uint64
|
||||||
// Amount of memory that does not belong to any file
|
// Amount of memory that does not belong to any file.
|
||||||
Anonymous uint64
|
Anonymous uint64
|
||||||
// Amount would-be-anonymous memory currently on swap
|
// Amount would-be-anonymous memory currently on swap.
|
||||||
Swap uint64
|
Swap uint64
|
||||||
// Process's proportional memory on swap
|
// Process's proportional memory on swap.
|
||||||
SwapPss uint64
|
SwapPss uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
|
353
vendor/github.com/prometheus/procfs/proc_snmp.go
generated
vendored
Normal file
353
vendor/github.com/prometheus/procfs/proc_snmp.go
generated
vendored
Normal file
@ -0,0 +1,353 @@
|
|||||||
|
// Copyright 2022 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package procfs
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/prometheus/procfs/internal/util"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ProcSnmp models the content of /proc/<pid>/net/snmp.
|
||||||
|
type ProcSnmp struct {
|
||||||
|
// The process ID.
|
||||||
|
PID int
|
||||||
|
Ip
|
||||||
|
Icmp
|
||||||
|
IcmpMsg
|
||||||
|
Tcp
|
||||||
|
Udp
|
||||||
|
UdpLite
|
||||||
|
}
|
||||||
|
|
||||||
|
type Ip struct { // nolint:revive
|
||||||
|
Forwarding float64
|
||||||
|
DefaultTTL float64
|
||||||
|
InReceives float64
|
||||||
|
InHdrErrors float64
|
||||||
|
InAddrErrors float64
|
||||||
|
ForwDatagrams float64
|
||||||
|
InUnknownProtos float64
|
||||||
|
InDiscards float64
|
||||||
|
InDelivers float64
|
||||||
|
OutRequests float64
|
||||||
|
OutDiscards float64
|
||||||
|
OutNoRoutes float64
|
||||||
|
ReasmTimeout float64
|
||||||
|
ReasmReqds float64
|
||||||
|
ReasmOKs float64
|
||||||
|
ReasmFails float64
|
||||||
|
FragOKs float64
|
||||||
|
FragFails float64
|
||||||
|
FragCreates float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type Icmp struct {
|
||||||
|
InMsgs float64
|
||||||
|
InErrors float64
|
||||||
|
InCsumErrors float64
|
||||||
|
InDestUnreachs float64
|
||||||
|
InTimeExcds float64
|
||||||
|
InParmProbs float64
|
||||||
|
InSrcQuenchs float64
|
||||||
|
InRedirects float64
|
||||||
|
InEchos float64
|
||||||
|
InEchoReps float64
|
||||||
|
InTimestamps float64
|
||||||
|
InTimestampReps float64
|
||||||
|
InAddrMasks float64
|
||||||
|
InAddrMaskReps float64
|
||||||
|
OutMsgs float64
|
||||||
|
OutErrors float64
|
||||||
|
OutDestUnreachs float64
|
||||||
|
OutTimeExcds float64
|
||||||
|
OutParmProbs float64
|
||||||
|
OutSrcQuenchs float64
|
||||||
|
OutRedirects float64
|
||||||
|
OutEchos float64
|
||||||
|
OutEchoReps float64
|
||||||
|
OutTimestamps float64
|
||||||
|
OutTimestampReps float64
|
||||||
|
OutAddrMasks float64
|
||||||
|
OutAddrMaskReps float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type IcmpMsg struct {
|
||||||
|
InType3 float64
|
||||||
|
OutType3 float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type Tcp struct { // nolint:revive
|
||||||
|
RtoAlgorithm float64
|
||||||
|
RtoMin float64
|
||||||
|
RtoMax float64
|
||||||
|
MaxConn float64
|
||||||
|
ActiveOpens float64
|
||||||
|
PassiveOpens float64
|
||||||
|
AttemptFails float64
|
||||||
|
EstabResets float64
|
||||||
|
CurrEstab float64
|
||||||
|
InSegs float64
|
||||||
|
OutSegs float64
|
||||||
|
RetransSegs float64
|
||||||
|
InErrs float64
|
||||||
|
OutRsts float64
|
||||||
|
InCsumErrors float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type Udp struct { // nolint:revive
|
||||||
|
InDatagrams float64
|
||||||
|
NoPorts float64
|
||||||
|
InErrors float64
|
||||||
|
OutDatagrams float64
|
||||||
|
RcvbufErrors float64
|
||||||
|
SndbufErrors float64
|
||||||
|
InCsumErrors float64
|
||||||
|
IgnoredMulti float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type UdpLite struct { // nolint:revive
|
||||||
|
InDatagrams float64
|
||||||
|
NoPorts float64
|
||||||
|
InErrors float64
|
||||||
|
OutDatagrams float64
|
||||||
|
RcvbufErrors float64
|
||||||
|
SndbufErrors float64
|
||||||
|
InCsumErrors float64
|
||||||
|
IgnoredMulti float64
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p Proc) Snmp() (ProcSnmp, error) {
|
||||||
|
filename := p.path("net/snmp")
|
||||||
|
data, err := util.ReadFileNoStat(filename)
|
||||||
|
if err != nil {
|
||||||
|
return ProcSnmp{PID: p.PID}, err
|
||||||
|
}
|
||||||
|
procSnmp, err := parseSnmp(bytes.NewReader(data), filename)
|
||||||
|
procSnmp.PID = p.PID
|
||||||
|
return procSnmp, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// parseSnmp parses the metrics from proc/<pid>/net/snmp file
|
||||||
|
// and returns a map contains those metrics (e.g. {"Ip": {"Forwarding": 2}}).
|
||||||
|
func parseSnmp(r io.Reader, fileName string) (ProcSnmp, error) {
|
||||||
|
var (
|
||||||
|
scanner = bufio.NewScanner(r)
|
||||||
|
procSnmp = ProcSnmp{}
|
||||||
|
)
|
||||||
|
|
||||||
|
for scanner.Scan() {
|
||||||
|
nameParts := strings.Split(scanner.Text(), " ")
|
||||||
|
scanner.Scan()
|
||||||
|
valueParts := strings.Split(scanner.Text(), " ")
|
||||||
|
// Remove trailing :.
|
||||||
|
protocol := strings.TrimSuffix(nameParts[0], ":")
|
||||||
|
if len(nameParts) != len(valueParts) {
|
||||||
|
return procSnmp, fmt.Errorf("mismatch field count mismatch in %s: %s",
|
||||||
|
fileName, protocol)
|
||||||
|
}
|
||||||
|
for i := 1; i < len(nameParts); i++ {
|
||||||
|
value, err := strconv.ParseFloat(valueParts[i], 64)
|
||||||
|
if err != nil {
|
||||||
|
return procSnmp, err
|
||||||
|
}
|
||||||
|
key := nameParts[i]
|
||||||
|
|
||||||
|
switch protocol {
|
||||||
|
case "Ip":
|
||||||
|
switch key {
|
||||||
|
case "Forwarding":
|
||||||
|
procSnmp.Ip.Forwarding = value
|
||||||
|
case "DefaultTTL":
|
||||||
|
procSnmp.Ip.DefaultTTL = value
|
||||||
|
case "InReceives":
|
||||||
|
procSnmp.Ip.InReceives = value
|
||||||
|
case "InHdrErrors":
|
||||||
|
procSnmp.Ip.InHdrErrors = value
|
||||||
|
case "InAddrErrors":
|
||||||
|
procSnmp.Ip.InAddrErrors = value
|
||||||
|
case "ForwDatagrams":
|
||||||
|
procSnmp.Ip.ForwDatagrams = value
|
||||||
|
case "InUnknownProtos":
|
||||||
|
procSnmp.Ip.InUnknownProtos = value
|
||||||
|
case "InDiscards":
|
||||||
|
procSnmp.Ip.InDiscards = value
|
||||||
|
case "InDelivers":
|
||||||
|
procSnmp.Ip.InDelivers = value
|
||||||
|
case "OutRequests":
|
||||||
|
procSnmp.Ip.OutRequests = value
|
||||||
|
case "OutDiscards":
|
||||||
|
procSnmp.Ip.OutDiscards = value
|
||||||
|
case "OutNoRoutes":
|
||||||
|
procSnmp.Ip.OutNoRoutes = value
|
||||||
|
case "ReasmTimeout":
|
||||||
|
procSnmp.Ip.ReasmTimeout = value
|
||||||
|
case "ReasmReqds":
|
||||||
|
procSnmp.Ip.ReasmReqds = value
|
||||||
|
case "ReasmOKs":
|
||||||
|
procSnmp.Ip.ReasmOKs = value
|
||||||
|
case "ReasmFails":
|
||||||
|
procSnmp.Ip.ReasmFails = value
|
||||||
|
case "FragOKs":
|
||||||
|
procSnmp.Ip.FragOKs = value
|
||||||
|
case "FragFails":
|
||||||
|
procSnmp.Ip.FragFails = value
|
||||||
|
case "FragCreates":
|
||||||
|
procSnmp.Ip.FragCreates = value
|
||||||
|
}
|
||||||
|
case "Icmp":
|
||||||
|
switch key {
|
||||||
|
case "InMsgs":
|
||||||
|
procSnmp.Icmp.InMsgs = value
|
||||||
|
case "InErrors":
|
||||||
|
procSnmp.Icmp.InErrors = value
|
||||||
|
case "InCsumErrors":
|
||||||
|
procSnmp.Icmp.InCsumErrors = value
|
||||||
|
case "InDestUnreachs":
|
||||||
|
procSnmp.Icmp.InDestUnreachs = value
|
||||||
|
case "InTimeExcds":
|
||||||
|
procSnmp.Icmp.InTimeExcds = value
|
||||||
|
case "InParmProbs":
|
||||||
|
procSnmp.Icmp.InParmProbs = value
|
||||||
|
case "InSrcQuenchs":
|
||||||
|
procSnmp.Icmp.InSrcQuenchs = value
|
||||||
|
case "InRedirects":
|
||||||
|
procSnmp.Icmp.InRedirects = value
|
||||||
|
case "InEchos":
|
||||||
|
procSnmp.Icmp.InEchos = value
|
||||||
|
case "InEchoReps":
|
||||||
|
procSnmp.Icmp.InEchoReps = value
|
||||||
|
case "InTimestamps":
|
||||||
|
procSnmp.Icmp.InTimestamps = value
|
||||||
|
case "InTimestampReps":
|
||||||
|
procSnmp.Icmp.InTimestampReps = value
|
||||||
|
case "InAddrMasks":
|
||||||
|
procSnmp.Icmp.InAddrMasks = value
|
||||||
|
case "InAddrMaskReps":
|
||||||
|
procSnmp.Icmp.InAddrMaskReps = value
|
||||||
|
case "OutMsgs":
|
||||||
|
procSnmp.Icmp.OutMsgs = value
|
||||||
|
case "OutErrors":
|
||||||
|
procSnmp.Icmp.OutErrors = value
|
||||||
|
case "OutDestUnreachs":
|
||||||
|
procSnmp.Icmp.OutDestUnreachs = value
|
||||||
|
case "OutTimeExcds":
|
||||||
|
procSnmp.Icmp.OutTimeExcds = value
|
||||||
|
case "OutParmProbs":
|
||||||
|
procSnmp.Icmp.OutParmProbs = value
|
||||||
|
case "OutSrcQuenchs":
|
||||||
|
procSnmp.Icmp.OutSrcQuenchs = value
|
||||||
|
case "OutRedirects":
|
||||||
|
procSnmp.Icmp.OutRedirects = value
|
||||||
|
case "OutEchos":
|
||||||
|
procSnmp.Icmp.OutEchos = value
|
||||||
|
case "OutEchoReps":
|
||||||
|
procSnmp.Icmp.OutEchoReps = value
|
||||||
|
case "OutTimestamps":
|
||||||
|
procSnmp.Icmp.OutTimestamps = value
|
||||||
|
case "OutTimestampReps":
|
||||||
|
procSnmp.Icmp.OutTimestampReps = value
|
||||||
|
case "OutAddrMasks":
|
||||||
|
procSnmp.Icmp.OutAddrMasks = value
|
||||||
|
case "OutAddrMaskReps":
|
||||||
|
procSnmp.Icmp.OutAddrMaskReps = value
|
||||||
|
}
|
||||||
|
case "IcmpMsg":
|
||||||
|
switch key {
|
||||||
|
case "InType3":
|
||||||
|
procSnmp.IcmpMsg.InType3 = value
|
||||||
|
case "OutType3":
|
||||||
|
procSnmp.IcmpMsg.OutType3 = value
|
||||||
|
}
|
||||||
|
case "Tcp":
|
||||||
|
switch key {
|
||||||
|
case "RtoAlgorithm":
|
||||||
|
procSnmp.Tcp.RtoAlgorithm = value
|
||||||
|
case "RtoMin":
|
||||||
|
procSnmp.Tcp.RtoMin = value
|
||||||
|
case "RtoMax":
|
||||||
|
procSnmp.Tcp.RtoMax = value
|
||||||
|
case "MaxConn":
|
||||||
|
procSnmp.Tcp.MaxConn = value
|
||||||
|
case "ActiveOpens":
|
||||||
|
procSnmp.Tcp.ActiveOpens = value
|
||||||
|
case "PassiveOpens":
|
||||||
|
procSnmp.Tcp.PassiveOpens = value
|
||||||
|
case "AttemptFails":
|
||||||
|
procSnmp.Tcp.AttemptFails = value
|
||||||
|
case "EstabResets":
|
||||||
|
procSnmp.Tcp.EstabResets = value
|
||||||
|
case "CurrEstab":
|
||||||
|
procSnmp.Tcp.CurrEstab = value
|
||||||
|
case "InSegs":
|
||||||
|
procSnmp.Tcp.InSegs = value
|
||||||
|
case "OutSegs":
|
||||||
|
procSnmp.Tcp.OutSegs = value
|
||||||
|
case "RetransSegs":
|
||||||
|
procSnmp.Tcp.RetransSegs = value
|
||||||
|
case "InErrs":
|
||||||
|
procSnmp.Tcp.InErrs = value
|
||||||
|
case "OutRsts":
|
||||||
|
procSnmp.Tcp.OutRsts = value
|
||||||
|
case "InCsumErrors":
|
||||||
|
procSnmp.Tcp.InCsumErrors = value
|
||||||
|
}
|
||||||
|
case "Udp":
|
||||||
|
switch key {
|
||||||
|
case "InDatagrams":
|
||||||
|
procSnmp.Udp.InDatagrams = value
|
||||||
|
case "NoPorts":
|
||||||
|
procSnmp.Udp.NoPorts = value
|
||||||
|
case "InErrors":
|
||||||
|
procSnmp.Udp.InErrors = value
|
||||||
|
case "OutDatagrams":
|
||||||
|
procSnmp.Udp.OutDatagrams = value
|
||||||
|
case "RcvbufErrors":
|
||||||
|
procSnmp.Udp.RcvbufErrors = value
|
||||||
|
case "SndbufErrors":
|
||||||
|
procSnmp.Udp.SndbufErrors = value
|
||||||
|
case "InCsumErrors":
|
||||||
|
procSnmp.Udp.InCsumErrors = value
|
||||||
|
case "IgnoredMulti":
|
||||||
|
procSnmp.Udp.IgnoredMulti = value
|
||||||
|
}
|
||||||
|
case "UdpLite":
|
||||||
|
switch key {
|
||||||
|
case "InDatagrams":
|
||||||
|
procSnmp.UdpLite.InDatagrams = value
|
||||||
|
case "NoPorts":
|
||||||
|
procSnmp.UdpLite.NoPorts = value
|
||||||
|
case "InErrors":
|
||||||
|
procSnmp.UdpLite.InErrors = value
|
||||||
|
case "OutDatagrams":
|
||||||
|
procSnmp.UdpLite.OutDatagrams = value
|
||||||
|
case "RcvbufErrors":
|
||||||
|
procSnmp.UdpLite.RcvbufErrors = value
|
||||||
|
case "SndbufErrors":
|
||||||
|
procSnmp.UdpLite.SndbufErrors = value
|
||||||
|
case "InCsumErrors":
|
||||||
|
procSnmp.UdpLite.InCsumErrors = value
|
||||||
|
case "IgnoredMulti":
|
||||||
|
procSnmp.UdpLite.IgnoredMulti = value
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return procSnmp, scanner.Err()
|
||||||
|
}
|
381
vendor/github.com/prometheus/procfs/proc_snmp6.go
generated
vendored
Normal file
381
vendor/github.com/prometheus/procfs/proc_snmp6.go
generated
vendored
Normal file
@ -0,0 +1,381 @@
|
|||||||
|
// Copyright 2022 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package procfs
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"errors"
|
||||||
|
"io"
|
||||||
|
"os"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/prometheus/procfs/internal/util"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ProcSnmp6 models the content of /proc/<pid>/net/snmp6.
|
||||||
|
type ProcSnmp6 struct {
|
||||||
|
// The process ID.
|
||||||
|
PID int
|
||||||
|
Ip6
|
||||||
|
Icmp6
|
||||||
|
Udp6
|
||||||
|
UdpLite6
|
||||||
|
}
|
||||||
|
|
||||||
|
type Ip6 struct { // nolint:revive
|
||||||
|
InReceives float64
|
||||||
|
InHdrErrors float64
|
||||||
|
InTooBigErrors float64
|
||||||
|
InNoRoutes float64
|
||||||
|
InAddrErrors float64
|
||||||
|
InUnknownProtos float64
|
||||||
|
InTruncatedPkts float64
|
||||||
|
InDiscards float64
|
||||||
|
InDelivers float64
|
||||||
|
OutForwDatagrams float64
|
||||||
|
OutRequests float64
|
||||||
|
OutDiscards float64
|
||||||
|
OutNoRoutes float64
|
||||||
|
ReasmTimeout float64
|
||||||
|
ReasmReqds float64
|
||||||
|
ReasmOKs float64
|
||||||
|
ReasmFails float64
|
||||||
|
FragOKs float64
|
||||||
|
FragFails float64
|
||||||
|
FragCreates float64
|
||||||
|
InMcastPkts float64
|
||||||
|
OutMcastPkts float64
|
||||||
|
InOctets float64
|
||||||
|
OutOctets float64
|
||||||
|
InMcastOctets float64
|
||||||
|
OutMcastOctets float64
|
||||||
|
InBcastOctets float64
|
||||||
|
OutBcastOctets float64
|
||||||
|
InNoECTPkts float64
|
||||||
|
InECT1Pkts float64
|
||||||
|
InECT0Pkts float64
|
||||||
|
InCEPkts float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type Icmp6 struct {
|
||||||
|
InMsgs float64
|
||||||
|
InErrors float64
|
||||||
|
OutMsgs float64
|
||||||
|
OutErrors float64
|
||||||
|
InCsumErrors float64
|
||||||
|
InDestUnreachs float64
|
||||||
|
InPktTooBigs float64
|
||||||
|
InTimeExcds float64
|
||||||
|
InParmProblems float64
|
||||||
|
InEchos float64
|
||||||
|
InEchoReplies float64
|
||||||
|
InGroupMembQueries float64
|
||||||
|
InGroupMembResponses float64
|
||||||
|
InGroupMembReductions float64
|
||||||
|
InRouterSolicits float64
|
||||||
|
InRouterAdvertisements float64
|
||||||
|
InNeighborSolicits float64
|
||||||
|
InNeighborAdvertisements float64
|
||||||
|
InRedirects float64
|
||||||
|
InMLDv2Reports float64
|
||||||
|
OutDestUnreachs float64
|
||||||
|
OutPktTooBigs float64
|
||||||
|
OutTimeExcds float64
|
||||||
|
OutParmProblems float64
|
||||||
|
OutEchos float64
|
||||||
|
OutEchoReplies float64
|
||||||
|
OutGroupMembQueries float64
|
||||||
|
OutGroupMembResponses float64
|
||||||
|
OutGroupMembReductions float64
|
||||||
|
OutRouterSolicits float64
|
||||||
|
OutRouterAdvertisements float64
|
||||||
|
OutNeighborSolicits float64
|
||||||
|
OutNeighborAdvertisements float64
|
||||||
|
OutRedirects float64
|
||||||
|
OutMLDv2Reports float64
|
||||||
|
InType1 float64
|
||||||
|
InType134 float64
|
||||||
|
InType135 float64
|
||||||
|
InType136 float64
|
||||||
|
InType143 float64
|
||||||
|
OutType133 float64
|
||||||
|
OutType135 float64
|
||||||
|
OutType136 float64
|
||||||
|
OutType143 float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type Udp6 struct { // nolint:revive
|
||||||
|
InDatagrams float64
|
||||||
|
NoPorts float64
|
||||||
|
InErrors float64
|
||||||
|
OutDatagrams float64
|
||||||
|
RcvbufErrors float64
|
||||||
|
SndbufErrors float64
|
||||||
|
InCsumErrors float64
|
||||||
|
IgnoredMulti float64
|
||||||
|
}
|
||||||
|
|
||||||
|
type UdpLite6 struct { // nolint:revive
|
||||||
|
InDatagrams float64
|
||||||
|
NoPorts float64
|
||||||
|
InErrors float64
|
||||||
|
OutDatagrams float64
|
||||||
|
RcvbufErrors float64
|
||||||
|
SndbufErrors float64
|
||||||
|
InCsumErrors float64
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p Proc) Snmp6() (ProcSnmp6, error) {
|
||||||
|
filename := p.path("net/snmp6")
|
||||||
|
data, err := util.ReadFileNoStat(filename)
|
||||||
|
if err != nil {
|
||||||
|
// On systems with IPv6 disabled, this file won't exist.
|
||||||
|
// Do nothing.
|
||||||
|
if errors.Is(err, os.ErrNotExist) {
|
||||||
|
return ProcSnmp6{PID: p.PID}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return ProcSnmp6{PID: p.PID}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
procSnmp6, err := parseSNMP6Stats(bytes.NewReader(data))
|
||||||
|
procSnmp6.PID = p.PID
|
||||||
|
return procSnmp6, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// parseSnmp6 parses the metrics from proc/<pid>/net/snmp6 file
|
||||||
|
// and returns a map contains those metrics.
|
||||||
|
func parseSNMP6Stats(r io.Reader) (ProcSnmp6, error) {
|
||||||
|
var (
|
||||||
|
scanner = bufio.NewScanner(r)
|
||||||
|
procSnmp6 = ProcSnmp6{}
|
||||||
|
)
|
||||||
|
|
||||||
|
for scanner.Scan() {
|
||||||
|
stat := strings.Fields(scanner.Text())
|
||||||
|
if len(stat) < 2 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// Expect to have "6" in metric name, skip line otherwise
|
||||||
|
if sixIndex := strings.Index(stat[0], "6"); sixIndex != -1 {
|
||||||
|
protocol := stat[0][:sixIndex+1]
|
||||||
|
key := stat[0][sixIndex+1:]
|
||||||
|
value, err := strconv.ParseFloat(stat[1], 64)
|
||||||
|
if err != nil {
|
||||||
|
return procSnmp6, err
|
||||||
|
}
|
||||||
|
|
||||||
|
switch protocol {
|
||||||
|
case "Ip6":
|
||||||
|
switch key {
|
||||||
|
case "InReceives":
|
||||||
|
procSnmp6.Ip6.InReceives = value
|
||||||
|
case "InHdrErrors":
|
||||||
|
procSnmp6.Ip6.InHdrErrors = value
|
||||||
|
case "InTooBigErrors":
|
||||||
|
procSnmp6.Ip6.InTooBigErrors = value
|
||||||
|
case "InNoRoutes":
|
||||||
|
procSnmp6.Ip6.InNoRoutes = value
|
||||||
|
case "InAddrErrors":
|
||||||
|
procSnmp6.Ip6.InAddrErrors = value
|
||||||
|
case "InUnknownProtos":
|
||||||
|
procSnmp6.Ip6.InUnknownProtos = value
|
||||||
|
case "InTruncatedPkts":
|
||||||
|
procSnmp6.Ip6.InTruncatedPkts = value
|
||||||
|
case "InDiscards":
|
||||||
|
procSnmp6.Ip6.InDiscards = value
|
||||||
|
case "InDelivers":
|
||||||
|
procSnmp6.Ip6.InDelivers = value
|
||||||
|
case "OutForwDatagrams":
|
||||||
|
procSnmp6.Ip6.OutForwDatagrams = value
|
||||||
|
case "OutRequests":
|
||||||
|
procSnmp6.Ip6.OutRequests = value
|
||||||
|
case "OutDiscards":
|
||||||
|
procSnmp6.Ip6.OutDiscards = value
|
||||||
|
case "OutNoRoutes":
|
||||||
|
procSnmp6.Ip6.OutNoRoutes = value
|
||||||
|
case "ReasmTimeout":
|
||||||
|
procSnmp6.Ip6.ReasmTimeout = value
|
||||||
|
case "ReasmReqds":
|
||||||
|
procSnmp6.Ip6.ReasmReqds = value
|
||||||
|
case "ReasmOKs":
|
||||||
|
procSnmp6.Ip6.ReasmOKs = value
|
||||||
|
case "ReasmFails":
|
||||||
|
procSnmp6.Ip6.ReasmFails = value
|
||||||
|
case "FragOKs":
|
||||||
|
procSnmp6.Ip6.FragOKs = value
|
||||||
|
case "FragFails":
|
||||||
|
procSnmp6.Ip6.FragFails = value
|
||||||
|
case "FragCreates":
|
||||||
|
procSnmp6.Ip6.FragCreates = value
|
||||||
|
case "InMcastPkts":
|
||||||
|
procSnmp6.Ip6.InMcastPkts = value
|
||||||
|
case "OutMcastPkts":
|
||||||
|
procSnmp6.Ip6.OutMcastPkts = value
|
||||||
|
case "InOctets":
|
||||||
|
procSnmp6.Ip6.InOctets = value
|
||||||
|
case "OutOctets":
|
||||||
|
procSnmp6.Ip6.OutOctets = value
|
||||||
|
case "InMcastOctets":
|
||||||
|
procSnmp6.Ip6.InMcastOctets = value
|
||||||
|
case "OutMcastOctets":
|
||||||
|
procSnmp6.Ip6.OutMcastOctets = value
|
||||||
|
case "InBcastOctets":
|
||||||
|
procSnmp6.Ip6.InBcastOctets = value
|
||||||
|
case "OutBcastOctets":
|
||||||
|
procSnmp6.Ip6.OutBcastOctets = value
|
||||||
|
case "InNoECTPkts":
|
||||||
|
procSnmp6.Ip6.InNoECTPkts = value
|
||||||
|
case "InECT1Pkts":
|
||||||
|
procSnmp6.Ip6.InECT1Pkts = value
|
||||||
|
case "InECT0Pkts":
|
||||||
|
procSnmp6.Ip6.InECT0Pkts = value
|
||||||
|
case "InCEPkts":
|
||||||
|
procSnmp6.Ip6.InCEPkts = value
|
||||||
|
|
||||||
|
}
|
||||||
|
case "Icmp6":
|
||||||
|
switch key {
|
||||||
|
case "InMsgs":
|
||||||
|
procSnmp6.Icmp6.InMsgs = value
|
||||||
|
case "InErrors":
|
||||||
|
procSnmp6.Icmp6.InErrors = value
|
||||||
|
case "OutMsgs":
|
||||||
|
procSnmp6.Icmp6.OutMsgs = value
|
||||||
|
case "OutErrors":
|
||||||
|
procSnmp6.Icmp6.OutErrors = value
|
||||||
|
case "InCsumErrors":
|
||||||
|
procSnmp6.Icmp6.InCsumErrors = value
|
||||||
|
case "InDestUnreachs":
|
||||||
|
procSnmp6.Icmp6.InDestUnreachs = value
|
||||||
|
case "InPktTooBigs":
|
||||||
|
procSnmp6.Icmp6.InPktTooBigs = value
|
||||||
|
case "InTimeExcds":
|
||||||
|
procSnmp6.Icmp6.InTimeExcds = value
|
||||||
|
case "InParmProblems":
|
||||||
|
procSnmp6.Icmp6.InParmProblems = value
|
||||||
|
case "InEchos":
|
||||||
|
procSnmp6.Icmp6.InEchos = value
|
||||||
|
case "InEchoReplies":
|
||||||
|
procSnmp6.Icmp6.InEchoReplies = value
|
||||||
|
case "InGroupMembQueries":
|
||||||
|
procSnmp6.Icmp6.InGroupMembQueries = value
|
||||||
|
case "InGroupMembResponses":
|
||||||
|
procSnmp6.Icmp6.InGroupMembResponses = value
|
||||||
|
case "InGroupMembReductions":
|
||||||
|
procSnmp6.Icmp6.InGroupMembReductions = value
|
||||||
|
case "InRouterSolicits":
|
||||||
|
procSnmp6.Icmp6.InRouterSolicits = value
|
||||||
|
case "InRouterAdvertisements":
|
||||||
|
procSnmp6.Icmp6.InRouterAdvertisements = value
|
||||||
|
case "InNeighborSolicits":
|
||||||
|
procSnmp6.Icmp6.InNeighborSolicits = value
|
||||||
|
case "InNeighborAdvertisements":
|
||||||
|
procSnmp6.Icmp6.InNeighborAdvertisements = value
|
||||||
|
case "InRedirects":
|
||||||
|
procSnmp6.Icmp6.InRedirects = value
|
||||||
|
case "InMLDv2Reports":
|
||||||
|
procSnmp6.Icmp6.InMLDv2Reports = value
|
||||||
|
case "OutDestUnreachs":
|
||||||
|
procSnmp6.Icmp6.OutDestUnreachs = value
|
||||||
|
case "OutPktTooBigs":
|
||||||
|
procSnmp6.Icmp6.OutPktTooBigs = value
|
||||||
|
case "OutTimeExcds":
|
||||||
|
procSnmp6.Icmp6.OutTimeExcds = value
|
||||||
|
case "OutParmProblems":
|
||||||
|
procSnmp6.Icmp6.OutParmProblems = value
|
||||||
|
case "OutEchos":
|
||||||
|
procSnmp6.Icmp6.OutEchos = value
|
||||||
|
case "OutEchoReplies":
|
||||||
|
procSnmp6.Icmp6.OutEchoReplies = value
|
||||||
|
case "OutGroupMembQueries":
|
||||||
|
procSnmp6.Icmp6.OutGroupMembQueries = value
|
||||||
|
case "OutGroupMembResponses":
|
||||||
|
procSnmp6.Icmp6.OutGroupMembResponses = value
|
||||||
|
case "OutGroupMembReductions":
|
||||||
|
procSnmp6.Icmp6.OutGroupMembReductions = value
|
||||||
|
case "OutRouterSolicits":
|
||||||
|
procSnmp6.Icmp6.OutRouterSolicits = value
|
||||||
|
case "OutRouterAdvertisements":
|
||||||
|
procSnmp6.Icmp6.OutRouterAdvertisements = value
|
||||||
|
case "OutNeighborSolicits":
|
||||||
|
procSnmp6.Icmp6.OutNeighborSolicits = value
|
||||||
|
case "OutNeighborAdvertisements":
|
||||||
|
procSnmp6.Icmp6.OutNeighborAdvertisements = value
|
||||||
|
case "OutRedirects":
|
||||||
|
procSnmp6.Icmp6.OutRedirects = value
|
||||||
|
case "OutMLDv2Reports":
|
||||||
|
procSnmp6.Icmp6.OutMLDv2Reports = value
|
||||||
|
case "InType1":
|
||||||
|
procSnmp6.Icmp6.InType1 = value
|
||||||
|
case "InType134":
|
||||||
|
procSnmp6.Icmp6.InType134 = value
|
||||||
|
case "InType135":
|
||||||
|
procSnmp6.Icmp6.InType135 = value
|
||||||
|
case "InType136":
|
||||||
|
procSnmp6.Icmp6.InType136 = value
|
||||||
|
case "InType143":
|
||||||
|
procSnmp6.Icmp6.InType143 = value
|
||||||
|
case "OutType133":
|
||||||
|
procSnmp6.Icmp6.OutType133 = value
|
||||||
|
case "OutType135":
|
||||||
|
procSnmp6.Icmp6.OutType135 = value
|
||||||
|
case "OutType136":
|
||||||
|
procSnmp6.Icmp6.OutType136 = value
|
||||||
|
case "OutType143":
|
||||||
|
procSnmp6.Icmp6.OutType143 = value
|
||||||
|
}
|
||||||
|
case "Udp6":
|
||||||
|
switch key {
|
||||||
|
case "InDatagrams":
|
||||||
|
procSnmp6.Udp6.InDatagrams = value
|
||||||
|
case "NoPorts":
|
||||||
|
procSnmp6.Udp6.NoPorts = value
|
||||||
|
case "InErrors":
|
||||||
|
procSnmp6.Udp6.InErrors = value
|
||||||
|
case "OutDatagrams":
|
||||||
|
procSnmp6.Udp6.OutDatagrams = value
|
||||||
|
case "RcvbufErrors":
|
||||||
|
procSnmp6.Udp6.RcvbufErrors = value
|
||||||
|
case "SndbufErrors":
|
||||||
|
procSnmp6.Udp6.SndbufErrors = value
|
||||||
|
case "InCsumErrors":
|
||||||
|
procSnmp6.Udp6.InCsumErrors = value
|
||||||
|
case "IgnoredMulti":
|
||||||
|
procSnmp6.Udp6.IgnoredMulti = value
|
||||||
|
}
|
||||||
|
case "UdpLite6":
|
||||||
|
switch key {
|
||||||
|
case "InDatagrams":
|
||||||
|
procSnmp6.UdpLite6.InDatagrams = value
|
||||||
|
case "NoPorts":
|
||||||
|
procSnmp6.UdpLite6.NoPorts = value
|
||||||
|
case "InErrors":
|
||||||
|
procSnmp6.UdpLite6.InErrors = value
|
||||||
|
case "OutDatagrams":
|
||||||
|
procSnmp6.UdpLite6.OutDatagrams = value
|
||||||
|
case "RcvbufErrors":
|
||||||
|
procSnmp6.UdpLite6.RcvbufErrors = value
|
||||||
|
case "SndbufErrors":
|
||||||
|
procSnmp6.UdpLite6.SndbufErrors = value
|
||||||
|
case "InCsumErrors":
|
||||||
|
procSnmp6.UdpLite6.InCsumErrors = value
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return procSnmp6, scanner.Err()
|
||||||
|
}
|
11
vendor/github.com/prometheus/procfs/proc_stat.go
generated
vendored
11
vendor/github.com/prometheus/procfs/proc_stat.go
generated
vendored
@ -81,10 +81,10 @@ type ProcStat struct {
|
|||||||
STime uint
|
STime uint
|
||||||
// Amount of time that this process's waited-for children have been
|
// Amount of time that this process's waited-for children have been
|
||||||
// scheduled in user mode, measured in clock ticks.
|
// scheduled in user mode, measured in clock ticks.
|
||||||
CUTime uint
|
CUTime int
|
||||||
// Amount of time that this process's waited-for children have been
|
// Amount of time that this process's waited-for children have been
|
||||||
// scheduled in kernel mode, measured in clock ticks.
|
// scheduled in kernel mode, measured in clock ticks.
|
||||||
CSTime uint
|
CSTime int
|
||||||
// For processes running a real-time scheduling policy, this is the negated
|
// For processes running a real-time scheduling policy, this is the negated
|
||||||
// scheduling priority, minus one.
|
// scheduling priority, minus one.
|
||||||
Priority int
|
Priority int
|
||||||
@ -115,7 +115,7 @@ type ProcStat struct {
|
|||||||
|
|
||||||
// NewStat returns the current status information of the process.
|
// NewStat returns the current status information of the process.
|
||||||
//
|
//
|
||||||
// Deprecated: use p.Stat() instead
|
// Deprecated: Use p.Stat() instead.
|
||||||
func (p Proc) NewStat() (ProcStat, error) {
|
func (p Proc) NewStat() (ProcStat, error) {
|
||||||
return p.Stat()
|
return p.Stat()
|
||||||
}
|
}
|
||||||
@ -141,6 +141,11 @@ func (p Proc) Stat() (ProcStat, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
s.Comm = string(data[l+1 : r])
|
s.Comm = string(data[l+1 : r])
|
||||||
|
|
||||||
|
// Check the following resources for the details about the particular stat
|
||||||
|
// fields and their data types:
|
||||||
|
// * https://man7.org/linux/man-pages/man5/proc.5.html
|
||||||
|
// * https://man7.org/linux/man-pages/man3/scanf.3.html
|
||||||
_, err = fmt.Fscan(
|
_, err = fmt.Fscan(
|
||||||
bytes.NewBuffer(data[r+2:]),
|
bytes.NewBuffer(data[r+2:]),
|
||||||
&s.State,
|
&s.State,
|
||||||
|
32
vendor/github.com/prometheus/procfs/proc_status.go
generated
vendored
32
vendor/github.com/prometheus/procfs/proc_status.go
generated
vendored
@ -33,37 +33,37 @@ type ProcStatus struct {
|
|||||||
TGID int
|
TGID int
|
||||||
|
|
||||||
// Peak virtual memory size.
|
// Peak virtual memory size.
|
||||||
VmPeak uint64 // nolint:golint
|
VmPeak uint64 // nolint:revive
|
||||||
// Virtual memory size.
|
// Virtual memory size.
|
||||||
VmSize uint64 // nolint:golint
|
VmSize uint64 // nolint:revive
|
||||||
// Locked memory size.
|
// Locked memory size.
|
||||||
VmLck uint64 // nolint:golint
|
VmLck uint64 // nolint:revive
|
||||||
// Pinned memory size.
|
// Pinned memory size.
|
||||||
VmPin uint64 // nolint:golint
|
VmPin uint64 // nolint:revive
|
||||||
// Peak resident set size.
|
// Peak resident set size.
|
||||||
VmHWM uint64 // nolint:golint
|
VmHWM uint64 // nolint:revive
|
||||||
// Resident set size (sum of RssAnnon RssFile and RssShmem).
|
// Resident set size (sum of RssAnnon RssFile and RssShmem).
|
||||||
VmRSS uint64 // nolint:golint
|
VmRSS uint64 // nolint:revive
|
||||||
// Size of resident anonymous memory.
|
// Size of resident anonymous memory.
|
||||||
RssAnon uint64 // nolint:golint
|
RssAnon uint64 // nolint:revive
|
||||||
// Size of resident file mappings.
|
// Size of resident file mappings.
|
||||||
RssFile uint64 // nolint:golint
|
RssFile uint64 // nolint:revive
|
||||||
// Size of resident shared memory.
|
// Size of resident shared memory.
|
||||||
RssShmem uint64 // nolint:golint
|
RssShmem uint64 // nolint:revive
|
||||||
// Size of data segments.
|
// Size of data segments.
|
||||||
VmData uint64 // nolint:golint
|
VmData uint64 // nolint:revive
|
||||||
// Size of stack segments.
|
// Size of stack segments.
|
||||||
VmStk uint64 // nolint:golint
|
VmStk uint64 // nolint:revive
|
||||||
// Size of text segments.
|
// Size of text segments.
|
||||||
VmExe uint64 // nolint:golint
|
VmExe uint64 // nolint:revive
|
||||||
// Shared library code size.
|
// Shared library code size.
|
||||||
VmLib uint64 // nolint:golint
|
VmLib uint64 // nolint:revive
|
||||||
// Page table entries size.
|
// Page table entries size.
|
||||||
VmPTE uint64 // nolint:golint
|
VmPTE uint64 // nolint:revive
|
||||||
// Size of second-level page tables.
|
// Size of second-level page tables.
|
||||||
VmPMD uint64 // nolint:golint
|
VmPMD uint64 // nolint:revive
|
||||||
// Swapped-out virtual memory size by anonymous private.
|
// Swapped-out virtual memory size by anonymous private.
|
||||||
VmSwap uint64 // nolint:golint
|
VmSwap uint64 // nolint:revive
|
||||||
// Size of hugetlb memory portions
|
// Size of hugetlb memory portions
|
||||||
HugetlbPages uint64
|
HugetlbPages uint64
|
||||||
|
|
||||||
|
51
vendor/github.com/prometheus/procfs/proc_sys.go
generated
vendored
Normal file
51
vendor/github.com/prometheus/procfs/proc_sys.go
generated
vendored
Normal file
@ -0,0 +1,51 @@
|
|||||||
|
// Copyright 2022 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package procfs
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/prometheus/procfs/internal/util"
|
||||||
|
)
|
||||||
|
|
||||||
|
func sysctlToPath(sysctl string) string {
|
||||||
|
return strings.Replace(sysctl, ".", "/", -1)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fs FS) SysctlStrings(sysctl string) ([]string, error) {
|
||||||
|
value, err := util.SysReadFile(fs.proc.Path("sys", sysctlToPath(sysctl)))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return strings.Fields(value), nil
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fs FS) SysctlInts(sysctl string) ([]int, error) {
|
||||||
|
fields, err := fs.SysctlStrings(sysctl)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
values := make([]int, len(fields))
|
||||||
|
for i, f := range fields {
|
||||||
|
vp := util.NewValueParser(f)
|
||||||
|
values[i] = vp.Int()
|
||||||
|
if err := vp.Err(); err != nil {
|
||||||
|
return nil, fmt.Errorf("field %d in sysctl %s is not a valid int: %w", i, sysctl, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return values, nil
|
||||||
|
}
|
6
vendor/github.com/prometheus/procfs/schedstat.go
generated
vendored
6
vendor/github.com/prometheus/procfs/schedstat.go
generated
vendored
@ -40,7 +40,7 @@ type Schedstat struct {
|
|||||||
CPUs []*SchedstatCPU
|
CPUs []*SchedstatCPU
|
||||||
}
|
}
|
||||||
|
|
||||||
// SchedstatCPU contains the values from one "cpu<N>" line
|
// SchedstatCPU contains the values from one "cpu<N>" line.
|
||||||
type SchedstatCPU struct {
|
type SchedstatCPU struct {
|
||||||
CPUNum string
|
CPUNum string
|
||||||
|
|
||||||
@ -49,14 +49,14 @@ type SchedstatCPU struct {
|
|||||||
RunTimeslices uint64
|
RunTimeslices uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
// ProcSchedstat contains the values from /proc/<pid>/schedstat
|
// ProcSchedstat contains the values from `/proc/<pid>/schedstat`.
|
||||||
type ProcSchedstat struct {
|
type ProcSchedstat struct {
|
||||||
RunningNanoseconds uint64
|
RunningNanoseconds uint64
|
||||||
WaitingNanoseconds uint64
|
WaitingNanoseconds uint64
|
||||||
RunTimeslices uint64
|
RunTimeslices uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
// Schedstat reads data from /proc/schedstat
|
// Schedstat reads data from `/proc/schedstat`.
|
||||||
func (fs FS) Schedstat() (*Schedstat, error) {
|
func (fs FS) Schedstat() (*Schedstat, error) {
|
||||||
file, err := os.Open(fs.proc.Path("schedstat"))
|
file, err := os.Open(fs.proc.Path("schedstat"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
2
vendor/github.com/prometheus/procfs/slab.go
generated
vendored
2
vendor/github.com/prometheus/procfs/slab.go
generated
vendored
@ -137,7 +137,7 @@ func parseSlabInfo21(r *bytes.Reader) (SlabInfo, error) {
|
|||||||
return s, nil
|
return s, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// SlabInfo reads data from /proc/slabinfo
|
// SlabInfo reads data from `/proc/slabinfo`.
|
||||||
func (fs FS) SlabInfo() (SlabInfo, error) {
|
func (fs FS) SlabInfo() (SlabInfo, error) {
|
||||||
// TODO: Consider passing options to allow for parsing different
|
// TODO: Consider passing options to allow for parsing different
|
||||||
// slabinfo versions. However, slabinfo 2.1 has been stable since
|
// slabinfo versions. However, slabinfo 2.1 has been stable since
|
||||||
|
160
vendor/github.com/prometheus/procfs/softirqs.go
generated
vendored
Normal file
160
vendor/github.com/prometheus/procfs/softirqs.go
generated
vendored
Normal file
@ -0,0 +1,160 @@
|
|||||||
|
// Copyright 2022 The Prometheus Authors
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package procfs
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/prometheus/procfs/internal/util"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Softirqs represents the softirq statistics.
|
||||||
|
type Softirqs struct {
|
||||||
|
Hi []uint64
|
||||||
|
Timer []uint64
|
||||||
|
NetTx []uint64
|
||||||
|
NetRx []uint64
|
||||||
|
Block []uint64
|
||||||
|
IRQPoll []uint64
|
||||||
|
Tasklet []uint64
|
||||||
|
Sched []uint64
|
||||||
|
HRTimer []uint64
|
||||||
|
RCU []uint64
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fs FS) Softirqs() (Softirqs, error) {
|
||||||
|
fileName := fs.proc.Path("softirqs")
|
||||||
|
data, err := util.ReadFileNoStat(fileName)
|
||||||
|
if err != nil {
|
||||||
|
return Softirqs{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
reader := bytes.NewReader(data)
|
||||||
|
|
||||||
|
return parseSoftirqs(reader)
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseSoftirqs(r io.Reader) (Softirqs, error) {
|
||||||
|
var (
|
||||||
|
softirqs = Softirqs{}
|
||||||
|
scanner = bufio.NewScanner(r)
|
||||||
|
)
|
||||||
|
|
||||||
|
if !scanner.Scan() {
|
||||||
|
return Softirqs{}, fmt.Errorf("softirqs empty")
|
||||||
|
}
|
||||||
|
|
||||||
|
for scanner.Scan() {
|
||||||
|
parts := strings.Fields(scanner.Text())
|
||||||
|
var err error
|
||||||
|
|
||||||
|
// require at least one cpu
|
||||||
|
if len(parts) < 2 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
switch {
|
||||||
|
case parts[0] == "HI:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.Hi = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.Hi[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (HI%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "TIMER:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.Timer = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.Timer[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (TIMER%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "NET_TX:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.NetTx = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.NetTx[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (NET_TX%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "NET_RX:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.NetRx = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.NetRx[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (NET_RX%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "BLOCK:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.Block = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.Block[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (BLOCK%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "IRQ_POLL:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.IRQPoll = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.IRQPoll[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (IRQ_POLL%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "TASKLET:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.Tasklet = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.Tasklet[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (TASKLET%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "SCHED:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.Sched = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.Sched[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (SCHED%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "HRTIMER:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.HRTimer = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.HRTimer[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (HRTIMER%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case parts[0] == "RCU:":
|
||||||
|
perCPU := parts[1:]
|
||||||
|
softirqs.RCU = make([]uint64, len(perCPU))
|
||||||
|
for i, count := range perCPU {
|
||||||
|
if softirqs.RCU[i], err = strconv.ParseUint(count, 10, 64); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse %q (RCU%d): %w", count, i, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := scanner.Err(); err != nil {
|
||||||
|
return Softirqs{}, fmt.Errorf("couldn't parse softirqs: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return softirqs, scanner.Err()
|
||||||
|
}
|
10
vendor/github.com/prometheus/procfs/stat.go
generated
vendored
10
vendor/github.com/prometheus/procfs/stat.go
generated
vendored
@ -41,7 +41,7 @@ type CPUStat struct {
|
|||||||
|
|
||||||
// SoftIRQStat represent the softirq statistics as exported in the procfs stat file.
|
// SoftIRQStat represent the softirq statistics as exported in the procfs stat file.
|
||||||
// A nice introduction can be found at https://0xax.gitbooks.io/linux-insides/content/interrupts/interrupts-9.html
|
// A nice introduction can be found at https://0xax.gitbooks.io/linux-insides/content/interrupts/interrupts-9.html
|
||||||
// It is possible to get per-cpu stats by reading /proc/softirqs
|
// It is possible to get per-cpu stats by reading `/proc/softirqs`.
|
||||||
type SoftIRQStat struct {
|
type SoftIRQStat struct {
|
||||||
Hi uint64
|
Hi uint64
|
||||||
Timer uint64
|
Timer uint64
|
||||||
@ -145,7 +145,7 @@ func parseSoftIRQStat(line string) (SoftIRQStat, uint64, error) {
|
|||||||
// NewStat returns information about current cpu/process statistics.
|
// NewStat returns information about current cpu/process statistics.
|
||||||
// See https://www.kernel.org/doc/Documentation/filesystems/proc.txt
|
// See https://www.kernel.org/doc/Documentation/filesystems/proc.txt
|
||||||
//
|
//
|
||||||
// Deprecated: use fs.Stat() instead
|
// Deprecated: Use fs.Stat() instead.
|
||||||
func NewStat() (Stat, error) {
|
func NewStat() (Stat, error) {
|
||||||
fs, err := NewFS(fs.DefaultProcMountPoint)
|
fs, err := NewFS(fs.DefaultProcMountPoint)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -155,15 +155,15 @@ func NewStat() (Stat, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// NewStat returns information about current cpu/process statistics.
|
// NewStat returns information about current cpu/process statistics.
|
||||||
// See https://www.kernel.org/doc/Documentation/filesystems/proc.txt
|
// See: https://www.kernel.org/doc/Documentation/filesystems/proc.txt
|
||||||
//
|
//
|
||||||
// Deprecated: use fs.Stat() instead
|
// Deprecated: Use fs.Stat() instead.
|
||||||
func (fs FS) NewStat() (Stat, error) {
|
func (fs FS) NewStat() (Stat, error) {
|
||||||
return fs.Stat()
|
return fs.Stat()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Stat returns information about current cpu/process statistics.
|
// Stat returns information about current cpu/process statistics.
|
||||||
// See https://www.kernel.org/doc/Documentation/filesystems/proc.txt
|
// See: https://www.kernel.org/doc/Documentation/filesystems/proc.txt
|
||||||
func (fs FS) Stat() (Stat, error) {
|
func (fs FS) Stat() (Stat, error) {
|
||||||
fileName := fs.proc.Path("stat")
|
fileName := fs.proc.Path("stat")
|
||||||
data, err := util.ReadFileNoStat(fileName)
|
data, err := util.ReadFileNoStat(fileName)
|
||||||
|
6
vendor/github.com/prometheus/procfs/vm.go
generated
vendored
6
vendor/github.com/prometheus/procfs/vm.go
generated
vendored
@ -11,13 +11,13 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build !windows
|
||||||
// +build !windows
|
// +build !windows
|
||||||
|
|
||||||
package procfs
|
package procfs
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
@ -29,7 +29,7 @@ import (
|
|||||||
// https://www.kernel.org/doc/Documentation/sysctl/vm.txt
|
// https://www.kernel.org/doc/Documentation/sysctl/vm.txt
|
||||||
// Each setting is exposed as a single file.
|
// Each setting is exposed as a single file.
|
||||||
// Each file contains one line with a single numerical value, except lowmem_reserve_ratio which holds an array
|
// Each file contains one line with a single numerical value, except lowmem_reserve_ratio which holds an array
|
||||||
// and numa_zonelist_order (deprecated) which is a string
|
// and numa_zonelist_order (deprecated) which is a string.
|
||||||
type VM struct {
|
type VM struct {
|
||||||
AdminReserveKbytes *int64 // /proc/sys/vm/admin_reserve_kbytes
|
AdminReserveKbytes *int64 // /proc/sys/vm/admin_reserve_kbytes
|
||||||
BlockDump *int64 // /proc/sys/vm/block_dump
|
BlockDump *int64 // /proc/sys/vm/block_dump
|
||||||
@ -87,7 +87,7 @@ func (fs FS) VM() (*VM, error) {
|
|||||||
return nil, fmt.Errorf("%s is not a directory", path)
|
return nil, fmt.Errorf("%s is not a directory", path)
|
||||||
}
|
}
|
||||||
|
|
||||||
files, err := ioutil.ReadDir(path)
|
files, err := os.ReadDir(path)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
5
vendor/github.com/prometheus/procfs/zoneinfo.go
generated
vendored
5
vendor/github.com/prometheus/procfs/zoneinfo.go
generated
vendored
@ -11,6 +11,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
|
//go:build !windows
|
||||||
// +build !windows
|
// +build !windows
|
||||||
|
|
||||||
package procfs
|
package procfs
|
||||||
@ -18,7 +19,7 @@ package procfs
|
|||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
"os"
|
||||||
"regexp"
|
"regexp"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
@ -72,7 +73,7 @@ var nodeZoneRE = regexp.MustCompile(`(\d+), zone\s+(\w+)`)
|
|||||||
// structs containing the relevant info. More information available here:
|
// structs containing the relevant info. More information available here:
|
||||||
// https://www.kernel.org/doc/Documentation/sysctl/vm.txt
|
// https://www.kernel.org/doc/Documentation/sysctl/vm.txt
|
||||||
func (fs FS) Zoneinfo() ([]Zoneinfo, error) {
|
func (fs FS) Zoneinfo() ([]Zoneinfo, error) {
|
||||||
data, err := ioutil.ReadFile(fs.proc.Path("zoneinfo"))
|
data, err := os.ReadFile(fs.proc.Path("zoneinfo"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("error reading zoneinfo %q: %w", fs.proc.Path("zoneinfo"), err)
|
return nil, fmt.Errorf("error reading zoneinfo %q: %w", fs.proc.Path("zoneinfo"), err)
|
||||||
}
|
}
|
||||||
|
14
vendor/modules.txt
vendored
14
vendor/modules.txt
vendored
@ -488,24 +488,24 @@ github.com/pkg/xattr
|
|||||||
# github.com/pmezard/go-difflib v1.0.0
|
# github.com/pmezard/go-difflib v1.0.0
|
||||||
## explicit
|
## explicit
|
||||||
github.com/pmezard/go-difflib/difflib
|
github.com/pmezard/go-difflib/difflib
|
||||||
# github.com/prometheus/client_golang v1.12.2
|
# github.com/prometheus/client_golang v1.14.0
|
||||||
## explicit; go 1.13
|
## explicit; go 1.17
|
||||||
github.com/prometheus/client_golang/prometheus
|
github.com/prometheus/client_golang/prometheus
|
||||||
github.com/prometheus/client_golang/prometheus/collectors
|
github.com/prometheus/client_golang/prometheus/collectors
|
||||||
github.com/prometheus/client_golang/prometheus/internal
|
github.com/prometheus/client_golang/prometheus/internal
|
||||||
github.com/prometheus/client_golang/prometheus/promhttp
|
github.com/prometheus/client_golang/prometheus/promhttp
|
||||||
github.com/prometheus/client_golang/prometheus/testutil
|
github.com/prometheus/client_golang/prometheus/testutil
|
||||||
github.com/prometheus/client_golang/prometheus/testutil/promlint
|
github.com/prometheus/client_golang/prometheus/testutil/promlint
|
||||||
# github.com/prometheus/client_model v0.2.0
|
# github.com/prometheus/client_model v0.3.0
|
||||||
## explicit; go 1.9
|
## explicit; go 1.9
|
||||||
github.com/prometheus/client_model/go
|
github.com/prometheus/client_model/go
|
||||||
# github.com/prometheus/common v0.32.1
|
# github.com/prometheus/common v0.37.0
|
||||||
## explicit; go 1.13
|
## explicit; go 1.16
|
||||||
github.com/prometheus/common/expfmt
|
github.com/prometheus/common/expfmt
|
||||||
github.com/prometheus/common/internal/bitbucket.org/ww/goautoneg
|
github.com/prometheus/common/internal/bitbucket.org/ww/goautoneg
|
||||||
github.com/prometheus/common/model
|
github.com/prometheus/common/model
|
||||||
# github.com/prometheus/procfs v0.7.3
|
# github.com/prometheus/procfs v0.8.0
|
||||||
## explicit; go 1.13
|
## explicit; go 1.17
|
||||||
github.com/prometheus/procfs
|
github.com/prometheus/procfs
|
||||||
github.com/prometheus/procfs/internal/fs
|
github.com/prometheus/procfs/internal/fs
|
||||||
github.com/prometheus/procfs/internal/util
|
github.com/prometheus/procfs/internal/util
|
||||||
|
Loading…
Reference in New Issue
Block a user