mirror of
https://github.com/open-telemetry/opentelemetry-go.git
synced 2025-02-03 13:11:53 +02:00
Update proto v0.5.0 (#1230)
* Updating version of OTLP to 0.5.0 * updating trace transform to use latest protos * update otlp span test * forgot one test * finishing the rest of the implementation In this change: - MetricDescriptor is no longer a member of the metric - splitting Sum, Gauge, Histogram by Int/Float - SummaryDataPoints are no longer around, MinMaxSumCount is now a Histogram * update changelog * Update CHANGELOG.md Move changes to the Unreleased section. This is to account for the recent release. Co-authored-by: Tyler Yahn <MrAlias@users.noreply.github.com>
This commit is contained in:
parent
27c84d689d
commit
25ccf5a08a
@ -16,6 +16,7 @@ This project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.htm
|
||||
- `ErrorOption` has been changed to an interface to conform with project design standards which included adding a `NewErrorConfig` function.
|
||||
- `EmptySpanContext` is removed.
|
||||
- Move the `go.opentelemetry.io/otel/api/trace/tracetest` package into `go.opentelemetry.io/otel/oteltest`. (#1229)
|
||||
- OTLP Exporter supports OTLP v0.5.0. (#1230)
|
||||
|
||||
## [0.13.0] - 2020-10-08
|
||||
|
||||
|
@ -1 +1 @@
|
||||
Subproject commit e43e1abc40428a6ee98e3bfd79bec1dfa2ed18cd
|
||||
Subproject commit 313a868be259dce6c6516dd417d3ad5fd3321acf
|
@ -0,0 +1,577 @@
|
||||
// Code generated by protoc-gen-gogo. DO NOT EDIT.
|
||||
// source: opentelemetry/proto/collector/logs/v1/logs_service.proto
|
||||
|
||||
package v1
|
||||
|
||||
import (
|
||||
context "context"
|
||||
fmt "fmt"
|
||||
proto "github.com/gogo/protobuf/proto"
|
||||
v1 "go.opentelemetry.io/otel/exporters/otlp/internal/opentelemetry-proto-gen/logs/v1"
|
||||
grpc "google.golang.org/grpc"
|
||||
codes "google.golang.org/grpc/codes"
|
||||
status "google.golang.org/grpc/status"
|
||||
io "io"
|
||||
math "math"
|
||||
math_bits "math/bits"
|
||||
)
|
||||
|
||||
// Reference imports to suppress errors if they are not otherwise used.
|
||||
var _ = proto.Marshal
|
||||
var _ = fmt.Errorf
|
||||
var _ = math.Inf
|
||||
|
||||
// This is a compile-time assertion to ensure that this generated file
|
||||
// is compatible with the proto package it is being compiled against.
|
||||
// A compilation error at this line likely means your copy of the
|
||||
// proto package needs to be updated.
|
||||
const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package
|
||||
|
||||
type ExportLogsServiceRequest struct {
|
||||
// An array of ResourceLogs.
|
||||
// For data coming from a single resource this array will typically contain one
|
||||
// element. Intermediary nodes (such as OpenTelemetry Collector) that receive
|
||||
// data from multiple origins typically batch the data before forwarding further and
|
||||
// in that case this array will contain multiple elements.
|
||||
ResourceLogs []*v1.ResourceLogs `protobuf:"bytes,1,rep,name=resource_logs,json=resourceLogs,proto3" json:"resource_logs,omitempty"`
|
||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||
XXX_unrecognized []byte `json:"-"`
|
||||
XXX_sizecache int32 `json:"-"`
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceRequest) Reset() { *m = ExportLogsServiceRequest{} }
|
||||
func (m *ExportLogsServiceRequest) String() string { return proto.CompactTextString(m) }
|
||||
func (*ExportLogsServiceRequest) ProtoMessage() {}
|
||||
func (*ExportLogsServiceRequest) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_8e3bf87aaa43acd4, []int{0}
|
||||
}
|
||||
func (m *ExportLogsServiceRequest) XXX_Unmarshal(b []byte) error {
|
||||
return m.Unmarshal(b)
|
||||
}
|
||||
func (m *ExportLogsServiceRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
if deterministic {
|
||||
return xxx_messageInfo_ExportLogsServiceRequest.Marshal(b, m, deterministic)
|
||||
} else {
|
||||
b = b[:cap(b)]
|
||||
n, err := m.MarshalToSizedBuffer(b)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return b[:n], nil
|
||||
}
|
||||
}
|
||||
func (m *ExportLogsServiceRequest) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_ExportLogsServiceRequest.Merge(m, src)
|
||||
}
|
||||
func (m *ExportLogsServiceRequest) XXX_Size() int {
|
||||
return m.Size()
|
||||
}
|
||||
func (m *ExportLogsServiceRequest) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_ExportLogsServiceRequest.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
var xxx_messageInfo_ExportLogsServiceRequest proto.InternalMessageInfo
|
||||
|
||||
func (m *ExportLogsServiceRequest) GetResourceLogs() []*v1.ResourceLogs {
|
||||
if m != nil {
|
||||
return m.ResourceLogs
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
type ExportLogsServiceResponse struct {
|
||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||
XXX_unrecognized []byte `json:"-"`
|
||||
XXX_sizecache int32 `json:"-"`
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceResponse) Reset() { *m = ExportLogsServiceResponse{} }
|
||||
func (m *ExportLogsServiceResponse) String() string { return proto.CompactTextString(m) }
|
||||
func (*ExportLogsServiceResponse) ProtoMessage() {}
|
||||
func (*ExportLogsServiceResponse) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_8e3bf87aaa43acd4, []int{1}
|
||||
}
|
||||
func (m *ExportLogsServiceResponse) XXX_Unmarshal(b []byte) error {
|
||||
return m.Unmarshal(b)
|
||||
}
|
||||
func (m *ExportLogsServiceResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
if deterministic {
|
||||
return xxx_messageInfo_ExportLogsServiceResponse.Marshal(b, m, deterministic)
|
||||
} else {
|
||||
b = b[:cap(b)]
|
||||
n, err := m.MarshalToSizedBuffer(b)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return b[:n], nil
|
||||
}
|
||||
}
|
||||
func (m *ExportLogsServiceResponse) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_ExportLogsServiceResponse.Merge(m, src)
|
||||
}
|
||||
func (m *ExportLogsServiceResponse) XXX_Size() int {
|
||||
return m.Size()
|
||||
}
|
||||
func (m *ExportLogsServiceResponse) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_ExportLogsServiceResponse.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
var xxx_messageInfo_ExportLogsServiceResponse proto.InternalMessageInfo
|
||||
|
||||
func init() {
|
||||
proto.RegisterType((*ExportLogsServiceRequest)(nil), "opentelemetry.proto.collector.logs.v1.ExportLogsServiceRequest")
|
||||
proto.RegisterType((*ExportLogsServiceResponse)(nil), "opentelemetry.proto.collector.logs.v1.ExportLogsServiceResponse")
|
||||
}
|
||||
|
||||
func init() {
|
||||
proto.RegisterFile("opentelemetry/proto/collector/logs/v1/logs_service.proto", fileDescriptor_8e3bf87aaa43acd4)
|
||||
}
|
||||
|
||||
var fileDescriptor_8e3bf87aaa43acd4 = []byte{
|
||||
// 290 bytes of a gzipped FileDescriptorProto
|
||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x92, 0x41, 0x4a, 0x03, 0x31,
|
||||
0x14, 0x86, 0x0d, 0x42, 0x17, 0xa9, 0x82, 0xcc, 0xaa, 0x56, 0x18, 0x64, 0x40, 0xa9, 0x8b, 0x26,
|
||||
0xb4, 0x6e, 0xdc, 0x29, 0x05, 0x77, 0x22, 0x65, 0xdc, 0x75, 0x53, 0x74, 0x78, 0x0c, 0x23, 0x31,
|
||||
0x2f, 0xbe, 0xa4, 0x83, 0x1e, 0xc2, 0x23, 0xb8, 0xf5, 0x2c, 0x2e, 0x3d, 0x82, 0xcc, 0x49, 0x64,
|
||||
0x92, 0x22, 0x53, 0x1d, 0x61, 0x70, 0x15, 0xf2, 0xf2, 0x7f, 0xff, 0xff, 0x3f, 0x08, 0x3f, 0x43,
|
||||
0x03, 0xda, 0x81, 0x82, 0x07, 0x70, 0xf4, 0x2c, 0x0d, 0xa1, 0x43, 0x99, 0xa1, 0x52, 0x90, 0x39,
|
||||
0x24, 0xa9, 0x30, 0xb7, 0xb2, 0x9c, 0xf8, 0x73, 0x69, 0x81, 0xca, 0x22, 0x03, 0xe1, 0x45, 0xd1,
|
||||
0xd1, 0x06, 0x19, 0x86, 0xe2, 0x9b, 0x14, 0x35, 0x21, 0xca, 0xc9, 0xf0, 0xb8, 0x2d, 0xa0, 0x69,
|
||||
0x1b, 0xc8, 0xe4, 0x9e, 0x0f, 0x2e, 0x9f, 0x0c, 0x92, 0xbb, 0xc2, 0xdc, 0xde, 0x84, 0xa4, 0x14,
|
||||
0x1e, 0x57, 0x60, 0x5d, 0x74, 0xcd, 0x77, 0x09, 0x2c, 0xae, 0x28, 0x83, 0x65, 0x8d, 0x0c, 0xd8,
|
||||
0xe1, 0xf6, 0xa8, 0x3f, 0x3d, 0x11, 0x6d, 0x15, 0xd6, 0xc1, 0x22, 0x5d, 0x13, 0xb5, 0x5f, 0xba,
|
||||
0x43, 0x8d, 0x5b, 0x72, 0xc0, 0xf7, 0x5b, 0xb2, 0xac, 0x41, 0x6d, 0x61, 0xfa, 0xca, 0x78, 0xbf,
|
||||
0x31, 0x8f, 0x5e, 0x18, 0xef, 0x05, 0x75, 0x74, 0x2e, 0x3a, 0xed, 0x2c, 0xfe, 0x5a, 0x64, 0x78,
|
||||
0xf1, 0x7f, 0x83, 0xd0, 0x2e, 0xd9, 0x9a, 0xbd, 0xb1, 0xf7, 0x2a, 0x66, 0x1f, 0x55, 0xcc, 0x3e,
|
||||
0xab, 0x98, 0xf1, 0x51, 0x81, 0xdd, 0x4c, 0x67, 0x7b, 0x0d, 0xbf, 0x79, 0xad, 0x99, 0xb3, 0xc5,
|
||||
0x22, 0xff, 0x49, 0x17, 0x28, 0xd1, 0x81, 0x92, 0xe0, 0x2b, 0x00, 0x59, 0x89, 0x4e, 0x19, 0x59,
|
||||
0x68, 0x07, 0xa4, 0x6f, 0x95, 0xdc, 0x50, 0x8f, 0x7d, 0xd6, 0x38, 0x07, 0xfd, 0xfb, 0xcf, 0xdc,
|
||||
0xf5, 0xfc, 0xe3, 0xe9, 0x57, 0x00, 0x00, 0x00, 0xff, 0xff, 0x46, 0x1c, 0xa2, 0x18, 0x63, 0x02,
|
||||
0x00, 0x00,
|
||||
}
|
||||
|
||||
// Reference imports to suppress errors if they are not otherwise used.
|
||||
var _ context.Context
|
||||
var _ grpc.ClientConn
|
||||
|
||||
// This is a compile-time assertion to ensure that this generated file
|
||||
// is compatible with the grpc package it is being compiled against.
|
||||
const _ = grpc.SupportPackageIsVersion4
|
||||
|
||||
// LogsServiceClient is the client API for LogsService service.
|
||||
//
|
||||
// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.
|
||||
type LogsServiceClient interface {
|
||||
// For performance reasons, it is recommended to keep this RPC
|
||||
// alive for the entire life of the application.
|
||||
Export(ctx context.Context, in *ExportLogsServiceRequest, opts ...grpc.CallOption) (*ExportLogsServiceResponse, error)
|
||||
}
|
||||
|
||||
type logsServiceClient struct {
|
||||
cc *grpc.ClientConn
|
||||
}
|
||||
|
||||
func NewLogsServiceClient(cc *grpc.ClientConn) LogsServiceClient {
|
||||
return &logsServiceClient{cc}
|
||||
}
|
||||
|
||||
func (c *logsServiceClient) Export(ctx context.Context, in *ExportLogsServiceRequest, opts ...grpc.CallOption) (*ExportLogsServiceResponse, error) {
|
||||
out := new(ExportLogsServiceResponse)
|
||||
err := c.cc.Invoke(ctx, "/opentelemetry.proto.collector.logs.v1.LogsService/Export", in, out, opts...)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return out, nil
|
||||
}
|
||||
|
||||
// LogsServiceServer is the server API for LogsService service.
|
||||
type LogsServiceServer interface {
|
||||
// For performance reasons, it is recommended to keep this RPC
|
||||
// alive for the entire life of the application.
|
||||
Export(context.Context, *ExportLogsServiceRequest) (*ExportLogsServiceResponse, error)
|
||||
}
|
||||
|
||||
// UnimplementedLogsServiceServer can be embedded to have forward compatible implementations.
|
||||
type UnimplementedLogsServiceServer struct {
|
||||
}
|
||||
|
||||
func (*UnimplementedLogsServiceServer) Export(ctx context.Context, req *ExportLogsServiceRequest) (*ExportLogsServiceResponse, error) {
|
||||
return nil, status.Errorf(codes.Unimplemented, "method Export not implemented")
|
||||
}
|
||||
|
||||
func RegisterLogsServiceServer(s *grpc.Server, srv LogsServiceServer) {
|
||||
s.RegisterService(&_LogsService_serviceDesc, srv)
|
||||
}
|
||||
|
||||
func _LogsService_Export_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
|
||||
in := new(ExportLogsServiceRequest)
|
||||
if err := dec(in); err != nil {
|
||||
return nil, err
|
||||
}
|
||||
if interceptor == nil {
|
||||
return srv.(LogsServiceServer).Export(ctx, in)
|
||||
}
|
||||
info := &grpc.UnaryServerInfo{
|
||||
Server: srv,
|
||||
FullMethod: "/opentelemetry.proto.collector.logs.v1.LogsService/Export",
|
||||
}
|
||||
handler := func(ctx context.Context, req interface{}) (interface{}, error) {
|
||||
return srv.(LogsServiceServer).Export(ctx, req.(*ExportLogsServiceRequest))
|
||||
}
|
||||
return interceptor(ctx, in, info, handler)
|
||||
}
|
||||
|
||||
var _LogsService_serviceDesc = grpc.ServiceDesc{
|
||||
ServiceName: "opentelemetry.proto.collector.logs.v1.LogsService",
|
||||
HandlerType: (*LogsServiceServer)(nil),
|
||||
Methods: []grpc.MethodDesc{
|
||||
{
|
||||
MethodName: "Export",
|
||||
Handler: _LogsService_Export_Handler,
|
||||
},
|
||||
},
|
||||
Streams: []grpc.StreamDesc{},
|
||||
Metadata: "opentelemetry/proto/collector/logs/v1/logs_service.proto",
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceRequest) Marshal() (dAtA []byte, err error) {
|
||||
size := m.Size()
|
||||
dAtA = make([]byte, size)
|
||||
n, err := m.MarshalToSizedBuffer(dAtA[:size])
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return dAtA[:n], nil
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceRequest) MarshalTo(dAtA []byte) (int, error) {
|
||||
size := m.Size()
|
||||
return m.MarshalToSizedBuffer(dAtA[:size])
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceRequest) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
i := len(dAtA)
|
||||
_ = i
|
||||
var l int
|
||||
_ = l
|
||||
if m.XXX_unrecognized != nil {
|
||||
i -= len(m.XXX_unrecognized)
|
||||
copy(dAtA[i:], m.XXX_unrecognized)
|
||||
}
|
||||
if len(m.ResourceLogs) > 0 {
|
||||
for iNdEx := len(m.ResourceLogs) - 1; iNdEx >= 0; iNdEx-- {
|
||||
{
|
||||
size, err := m.ResourceLogs[iNdEx].MarshalToSizedBuffer(dAtA[:i])
|
||||
if err != nil {
|
||||
return 0, err
|
||||
}
|
||||
i -= size
|
||||
i = encodeVarintLogsService(dAtA, i, uint64(size))
|
||||
}
|
||||
i--
|
||||
dAtA[i] = 0xa
|
||||
}
|
||||
}
|
||||
return len(dAtA) - i, nil
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceResponse) Marshal() (dAtA []byte, err error) {
|
||||
size := m.Size()
|
||||
dAtA = make([]byte, size)
|
||||
n, err := m.MarshalToSizedBuffer(dAtA[:size])
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return dAtA[:n], nil
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceResponse) MarshalTo(dAtA []byte) (int, error) {
|
||||
size := m.Size()
|
||||
return m.MarshalToSizedBuffer(dAtA[:size])
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceResponse) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
i := len(dAtA)
|
||||
_ = i
|
||||
var l int
|
||||
_ = l
|
||||
if m.XXX_unrecognized != nil {
|
||||
i -= len(m.XXX_unrecognized)
|
||||
copy(dAtA[i:], m.XXX_unrecognized)
|
||||
}
|
||||
return len(dAtA) - i, nil
|
||||
}
|
||||
|
||||
func encodeVarintLogsService(dAtA []byte, offset int, v uint64) int {
|
||||
offset -= sovLogsService(v)
|
||||
base := offset
|
||||
for v >= 1<<7 {
|
||||
dAtA[offset] = uint8(v&0x7f | 0x80)
|
||||
v >>= 7
|
||||
offset++
|
||||
}
|
||||
dAtA[offset] = uint8(v)
|
||||
return base
|
||||
}
|
||||
func (m *ExportLogsServiceRequest) Size() (n int) {
|
||||
if m == nil {
|
||||
return 0
|
||||
}
|
||||
var l int
|
||||
_ = l
|
||||
if len(m.ResourceLogs) > 0 {
|
||||
for _, e := range m.ResourceLogs {
|
||||
l = e.Size()
|
||||
n += 1 + l + sovLogsService(uint64(l))
|
||||
}
|
||||
}
|
||||
if m.XXX_unrecognized != nil {
|
||||
n += len(m.XXX_unrecognized)
|
||||
}
|
||||
return n
|
||||
}
|
||||
|
||||
func (m *ExportLogsServiceResponse) Size() (n int) {
|
||||
if m == nil {
|
||||
return 0
|
||||
}
|
||||
var l int
|
||||
_ = l
|
||||
if m.XXX_unrecognized != nil {
|
||||
n += len(m.XXX_unrecognized)
|
||||
}
|
||||
return n
|
||||
}
|
||||
|
||||
func sovLogsService(x uint64) (n int) {
|
||||
return (math_bits.Len64(x|1) + 6) / 7
|
||||
}
|
||||
func sozLogsService(x uint64) (n int) {
|
||||
return sovLogsService(uint64((x << 1) ^ uint64((int64(x) >> 63))))
|
||||
}
|
||||
func (m *ExportLogsServiceRequest) Unmarshal(dAtA []byte) error {
|
||||
l := len(dAtA)
|
||||
iNdEx := 0
|
||||
for iNdEx < l {
|
||||
preIndex := iNdEx
|
||||
var wire uint64
|
||||
for shift := uint(0); ; shift += 7 {
|
||||
if shift >= 64 {
|
||||
return ErrIntOverflowLogsService
|
||||
}
|
||||
if iNdEx >= l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
b := dAtA[iNdEx]
|
||||
iNdEx++
|
||||
wire |= uint64(b&0x7F) << shift
|
||||
if b < 0x80 {
|
||||
break
|
||||
}
|
||||
}
|
||||
fieldNum := int32(wire >> 3)
|
||||
wireType := int(wire & 0x7)
|
||||
if wireType == 4 {
|
||||
return fmt.Errorf("proto: ExportLogsServiceRequest: wiretype end group for non-group")
|
||||
}
|
||||
if fieldNum <= 0 {
|
||||
return fmt.Errorf("proto: ExportLogsServiceRequest: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
}
|
||||
switch fieldNum {
|
||||
case 1:
|
||||
if wireType != 2 {
|
||||
return fmt.Errorf("proto: wrong wireType = %d for field ResourceLogs", wireType)
|
||||
}
|
||||
var msglen int
|
||||
for shift := uint(0); ; shift += 7 {
|
||||
if shift >= 64 {
|
||||
return ErrIntOverflowLogsService
|
||||
}
|
||||
if iNdEx >= l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
b := dAtA[iNdEx]
|
||||
iNdEx++
|
||||
msglen |= int(b&0x7F) << shift
|
||||
if b < 0x80 {
|
||||
break
|
||||
}
|
||||
}
|
||||
if msglen < 0 {
|
||||
return ErrInvalidLengthLogsService
|
||||
}
|
||||
postIndex := iNdEx + msglen
|
||||
if postIndex < 0 {
|
||||
return ErrInvalidLengthLogsService
|
||||
}
|
||||
if postIndex > l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
m.ResourceLogs = append(m.ResourceLogs, &v1.ResourceLogs{})
|
||||
if err := m.ResourceLogs[len(m.ResourceLogs)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
|
||||
return err
|
||||
}
|
||||
iNdEx = postIndex
|
||||
default:
|
||||
iNdEx = preIndex
|
||||
skippy, err := skipLogsService(dAtA[iNdEx:])
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if skippy < 0 {
|
||||
return ErrInvalidLengthLogsService
|
||||
}
|
||||
if (iNdEx + skippy) < 0 {
|
||||
return ErrInvalidLengthLogsService
|
||||
}
|
||||
if (iNdEx + skippy) > l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
m.XXX_unrecognized = append(m.XXX_unrecognized, dAtA[iNdEx:iNdEx+skippy]...)
|
||||
iNdEx += skippy
|
||||
}
|
||||
}
|
||||
|
||||
if iNdEx > l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
return nil
|
||||
}
|
||||
func (m *ExportLogsServiceResponse) Unmarshal(dAtA []byte) error {
|
||||
l := len(dAtA)
|
||||
iNdEx := 0
|
||||
for iNdEx < l {
|
||||
preIndex := iNdEx
|
||||
var wire uint64
|
||||
for shift := uint(0); ; shift += 7 {
|
||||
if shift >= 64 {
|
||||
return ErrIntOverflowLogsService
|
||||
}
|
||||
if iNdEx >= l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
b := dAtA[iNdEx]
|
||||
iNdEx++
|
||||
wire |= uint64(b&0x7F) << shift
|
||||
if b < 0x80 {
|
||||
break
|
||||
}
|
||||
}
|
||||
fieldNum := int32(wire >> 3)
|
||||
wireType := int(wire & 0x7)
|
||||
if wireType == 4 {
|
||||
return fmt.Errorf("proto: ExportLogsServiceResponse: wiretype end group for non-group")
|
||||
}
|
||||
if fieldNum <= 0 {
|
||||
return fmt.Errorf("proto: ExportLogsServiceResponse: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
}
|
||||
switch fieldNum {
|
||||
default:
|
||||
iNdEx = preIndex
|
||||
skippy, err := skipLogsService(dAtA[iNdEx:])
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if skippy < 0 {
|
||||
return ErrInvalidLengthLogsService
|
||||
}
|
||||
if (iNdEx + skippy) < 0 {
|
||||
return ErrInvalidLengthLogsService
|
||||
}
|
||||
if (iNdEx + skippy) > l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
m.XXX_unrecognized = append(m.XXX_unrecognized, dAtA[iNdEx:iNdEx+skippy]...)
|
||||
iNdEx += skippy
|
||||
}
|
||||
}
|
||||
|
||||
if iNdEx > l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
return nil
|
||||
}
|
||||
func skipLogsService(dAtA []byte) (n int, err error) {
|
||||
l := len(dAtA)
|
||||
iNdEx := 0
|
||||
depth := 0
|
||||
for iNdEx < l {
|
||||
var wire uint64
|
||||
for shift := uint(0); ; shift += 7 {
|
||||
if shift >= 64 {
|
||||
return 0, ErrIntOverflowLogsService
|
||||
}
|
||||
if iNdEx >= l {
|
||||
return 0, io.ErrUnexpectedEOF
|
||||
}
|
||||
b := dAtA[iNdEx]
|
||||
iNdEx++
|
||||
wire |= (uint64(b) & 0x7F) << shift
|
||||
if b < 0x80 {
|
||||
break
|
||||
}
|
||||
}
|
||||
wireType := int(wire & 0x7)
|
||||
switch wireType {
|
||||
case 0:
|
||||
for shift := uint(0); ; shift += 7 {
|
||||
if shift >= 64 {
|
||||
return 0, ErrIntOverflowLogsService
|
||||
}
|
||||
if iNdEx >= l {
|
||||
return 0, io.ErrUnexpectedEOF
|
||||
}
|
||||
iNdEx++
|
||||
if dAtA[iNdEx-1] < 0x80 {
|
||||
break
|
||||
}
|
||||
}
|
||||
case 1:
|
||||
iNdEx += 8
|
||||
case 2:
|
||||
var length int
|
||||
for shift := uint(0); ; shift += 7 {
|
||||
if shift >= 64 {
|
||||
return 0, ErrIntOverflowLogsService
|
||||
}
|
||||
if iNdEx >= l {
|
||||
return 0, io.ErrUnexpectedEOF
|
||||
}
|
||||
b := dAtA[iNdEx]
|
||||
iNdEx++
|
||||
length |= (int(b) & 0x7F) << shift
|
||||
if b < 0x80 {
|
||||
break
|
||||
}
|
||||
}
|
||||
if length < 0 {
|
||||
return 0, ErrInvalidLengthLogsService
|
||||
}
|
||||
iNdEx += length
|
||||
case 3:
|
||||
depth++
|
||||
case 4:
|
||||
if depth == 0 {
|
||||
return 0, ErrUnexpectedEndOfGroupLogsService
|
||||
}
|
||||
depth--
|
||||
case 5:
|
||||
iNdEx += 4
|
||||
default:
|
||||
return 0, fmt.Errorf("proto: illegal wireType %d", wireType)
|
||||
}
|
||||
if iNdEx < 0 {
|
||||
return 0, ErrInvalidLengthLogsService
|
||||
}
|
||||
if depth == 0 {
|
||||
return iNdEx, nil
|
||||
}
|
||||
}
|
||||
return 0, io.ErrUnexpectedEOF
|
||||
}
|
||||
|
||||
var (
|
||||
ErrInvalidLengthLogsService = fmt.Errorf("proto: negative length found during unmarshaling")
|
||||
ErrIntOverflowLogsService = fmt.Errorf("proto: integer overflow")
|
||||
ErrUnexpectedEndOfGroupLogsService = fmt.Errorf("proto: unexpected end of group")
|
||||
)
|
1447
exporters/otlp/internal/opentelemetry-proto-gen/logs/v1/logs.pb.go
Normal file
1447
exporters/otlp/internal/opentelemetry-proto-gen/logs/v1/logs.pb.go
Normal file
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
@ -35,39 +35,39 @@ const (
|
||||
Span_SPAN_KIND_UNSPECIFIED Span_SpanKind = 0
|
||||
// Indicates that the span represents an internal operation within an application,
|
||||
// as opposed to an operations happening at the boundaries. Default value.
|
||||
Span_INTERNAL Span_SpanKind = 1
|
||||
Span_SPAN_KIND_INTERNAL Span_SpanKind = 1
|
||||
// Indicates that the span covers server-side handling of an RPC or other
|
||||
// remote network request.
|
||||
Span_SERVER Span_SpanKind = 2
|
||||
Span_SPAN_KIND_SERVER Span_SpanKind = 2
|
||||
// Indicates that the span describes a request to some remote service.
|
||||
Span_CLIENT Span_SpanKind = 3
|
||||
Span_SPAN_KIND_CLIENT Span_SpanKind = 3
|
||||
// Indicates that the span describes a producer sending a message to a broker.
|
||||
// Unlike CLIENT and SERVER, there is often no direct critical path latency relationship
|
||||
// between producer and consumer spans. A PRODUCER span ends when the message was accepted
|
||||
// by the broker while the logical processing of the message might span a much longer time.
|
||||
Span_PRODUCER Span_SpanKind = 4
|
||||
Span_SPAN_KIND_PRODUCER Span_SpanKind = 4
|
||||
// Indicates that the span describes consumer receiving a message from a broker.
|
||||
// Like the PRODUCER kind, there is often no direct critical path latency relationship
|
||||
// between producer and consumer spans.
|
||||
Span_CONSUMER Span_SpanKind = 5
|
||||
Span_SPAN_KIND_CONSUMER Span_SpanKind = 5
|
||||
)
|
||||
|
||||
var Span_SpanKind_name = map[int32]string{
|
||||
0: "SPAN_KIND_UNSPECIFIED",
|
||||
1: "INTERNAL",
|
||||
2: "SERVER",
|
||||
3: "CLIENT",
|
||||
4: "PRODUCER",
|
||||
5: "CONSUMER",
|
||||
1: "SPAN_KIND_INTERNAL",
|
||||
2: "SPAN_KIND_SERVER",
|
||||
3: "SPAN_KIND_CLIENT",
|
||||
4: "SPAN_KIND_PRODUCER",
|
||||
5: "SPAN_KIND_CONSUMER",
|
||||
}
|
||||
|
||||
var Span_SpanKind_value = map[string]int32{
|
||||
"SPAN_KIND_UNSPECIFIED": 0,
|
||||
"INTERNAL": 1,
|
||||
"SERVER": 2,
|
||||
"CLIENT": 3,
|
||||
"PRODUCER": 4,
|
||||
"CONSUMER": 5,
|
||||
"SPAN_KIND_INTERNAL": 1,
|
||||
"SPAN_KIND_SERVER": 2,
|
||||
"SPAN_KIND_CLIENT": 3,
|
||||
"SPAN_KIND_PRODUCER": 4,
|
||||
"SPAN_KIND_CONSUMER": 5,
|
||||
}
|
||||
|
||||
func (x Span_SpanKind) String() string {
|
||||
@ -83,63 +83,63 @@ func (Span_SpanKind) EnumDescriptor() ([]byte, []int) {
|
||||
type Status_StatusCode int32
|
||||
|
||||
const (
|
||||
Status_Ok Status_StatusCode = 0
|
||||
Status_Cancelled Status_StatusCode = 1
|
||||
Status_UnknownError Status_StatusCode = 2
|
||||
Status_InvalidArgument Status_StatusCode = 3
|
||||
Status_DeadlineExceeded Status_StatusCode = 4
|
||||
Status_NotFound Status_StatusCode = 5
|
||||
Status_AlreadyExists Status_StatusCode = 6
|
||||
Status_PermissionDenied Status_StatusCode = 7
|
||||
Status_ResourceExhausted Status_StatusCode = 8
|
||||
Status_FailedPrecondition Status_StatusCode = 9
|
||||
Status_Aborted Status_StatusCode = 10
|
||||
Status_OutOfRange Status_StatusCode = 11
|
||||
Status_Unimplemented Status_StatusCode = 12
|
||||
Status_InternalError Status_StatusCode = 13
|
||||
Status_Unavailable Status_StatusCode = 14
|
||||
Status_DataLoss Status_StatusCode = 15
|
||||
Status_Unauthenticated Status_StatusCode = 16
|
||||
Status_STATUS_CODE_OK Status_StatusCode = 0
|
||||
Status_STATUS_CODE_CANCELLED Status_StatusCode = 1
|
||||
Status_STATUS_CODE_UNKNOWN_ERROR Status_StatusCode = 2
|
||||
Status_STATUS_CODE_INVALID_ARGUMENT Status_StatusCode = 3
|
||||
Status_STATUS_CODE_DEADLINE_EXCEEDED Status_StatusCode = 4
|
||||
Status_STATUS_CODE_NOT_FOUND Status_StatusCode = 5
|
||||
Status_STATUS_CODE_ALREADY_EXISTS Status_StatusCode = 6
|
||||
Status_STATUS_CODE_PERMISSION_DENIED Status_StatusCode = 7
|
||||
Status_STATUS_CODE_RESOURCE_EXHAUSTED Status_StatusCode = 8
|
||||
Status_STATUS_CODE_FAILED_PRECONDITION Status_StatusCode = 9
|
||||
Status_STATUS_CODE_ABORTED Status_StatusCode = 10
|
||||
Status_STATUS_CODE_OUT_OF_RANGE Status_StatusCode = 11
|
||||
Status_STATUS_CODE_UNIMPLEMENTED Status_StatusCode = 12
|
||||
Status_STATUS_CODE_INTERNAL_ERROR Status_StatusCode = 13
|
||||
Status_STATUS_CODE_UNAVAILABLE Status_StatusCode = 14
|
||||
Status_STATUS_CODE_DATA_LOSS Status_StatusCode = 15
|
||||
Status_STATUS_CODE_UNAUTHENTICATED Status_StatusCode = 16
|
||||
)
|
||||
|
||||
var Status_StatusCode_name = map[int32]string{
|
||||
0: "Ok",
|
||||
1: "Cancelled",
|
||||
2: "UnknownError",
|
||||
3: "InvalidArgument",
|
||||
4: "DeadlineExceeded",
|
||||
5: "NotFound",
|
||||
6: "AlreadyExists",
|
||||
7: "PermissionDenied",
|
||||
8: "ResourceExhausted",
|
||||
9: "FailedPrecondition",
|
||||
10: "Aborted",
|
||||
11: "OutOfRange",
|
||||
12: "Unimplemented",
|
||||
13: "InternalError",
|
||||
14: "Unavailable",
|
||||
15: "DataLoss",
|
||||
16: "Unauthenticated",
|
||||
0: "STATUS_CODE_OK",
|
||||
1: "STATUS_CODE_CANCELLED",
|
||||
2: "STATUS_CODE_UNKNOWN_ERROR",
|
||||
3: "STATUS_CODE_INVALID_ARGUMENT",
|
||||
4: "STATUS_CODE_DEADLINE_EXCEEDED",
|
||||
5: "STATUS_CODE_NOT_FOUND",
|
||||
6: "STATUS_CODE_ALREADY_EXISTS",
|
||||
7: "STATUS_CODE_PERMISSION_DENIED",
|
||||
8: "STATUS_CODE_RESOURCE_EXHAUSTED",
|
||||
9: "STATUS_CODE_FAILED_PRECONDITION",
|
||||
10: "STATUS_CODE_ABORTED",
|
||||
11: "STATUS_CODE_OUT_OF_RANGE",
|
||||
12: "STATUS_CODE_UNIMPLEMENTED",
|
||||
13: "STATUS_CODE_INTERNAL_ERROR",
|
||||
14: "STATUS_CODE_UNAVAILABLE",
|
||||
15: "STATUS_CODE_DATA_LOSS",
|
||||
16: "STATUS_CODE_UNAUTHENTICATED",
|
||||
}
|
||||
|
||||
var Status_StatusCode_value = map[string]int32{
|
||||
"Ok": 0,
|
||||
"Cancelled": 1,
|
||||
"UnknownError": 2,
|
||||
"InvalidArgument": 3,
|
||||
"DeadlineExceeded": 4,
|
||||
"NotFound": 5,
|
||||
"AlreadyExists": 6,
|
||||
"PermissionDenied": 7,
|
||||
"ResourceExhausted": 8,
|
||||
"FailedPrecondition": 9,
|
||||
"Aborted": 10,
|
||||
"OutOfRange": 11,
|
||||
"Unimplemented": 12,
|
||||
"InternalError": 13,
|
||||
"Unavailable": 14,
|
||||
"DataLoss": 15,
|
||||
"Unauthenticated": 16,
|
||||
"STATUS_CODE_OK": 0,
|
||||
"STATUS_CODE_CANCELLED": 1,
|
||||
"STATUS_CODE_UNKNOWN_ERROR": 2,
|
||||
"STATUS_CODE_INVALID_ARGUMENT": 3,
|
||||
"STATUS_CODE_DEADLINE_EXCEEDED": 4,
|
||||
"STATUS_CODE_NOT_FOUND": 5,
|
||||
"STATUS_CODE_ALREADY_EXISTS": 6,
|
||||
"STATUS_CODE_PERMISSION_DENIED": 7,
|
||||
"STATUS_CODE_RESOURCE_EXHAUSTED": 8,
|
||||
"STATUS_CODE_FAILED_PRECONDITION": 9,
|
||||
"STATUS_CODE_ABORTED": 10,
|
||||
"STATUS_CODE_OUT_OF_RANGE": 11,
|
||||
"STATUS_CODE_UNIMPLEMENTED": 12,
|
||||
"STATUS_CODE_INTERNAL_ERROR": 13,
|
||||
"STATUS_CODE_UNAVAILABLE": 14,
|
||||
"STATUS_CODE_DATA_LOSS": 15,
|
||||
"STATUS_CODE_UNAUTHENTICATED": 16,
|
||||
}
|
||||
|
||||
func (x Status_StatusCode) String() string {
|
||||
@ -725,7 +725,7 @@ func (m *Status) GetCode() Status_StatusCode {
|
||||
if m != nil {
|
||||
return m.Code
|
||||
}
|
||||
return Status_Ok
|
||||
return Status_STATUS_CODE_OK
|
||||
}
|
||||
|
||||
func (m *Status) GetMessage() string {
|
||||
@ -751,72 +751,75 @@ func init() {
|
||||
}
|
||||
|
||||
var fileDescriptor_5c407ac9c675a601 = []byte{
|
||||
// 1031 bytes of a gzipped FileDescriptorProto
|
||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x56, 0xcb, 0x6e, 0x23, 0x45,
|
||||
0x14, 0x9d, 0xf6, 0x3b, 0xd7, 0x8f, 0x74, 0x6a, 0x5e, 0x3d, 0x19, 0xc8, 0x58, 0xd6, 0x48, 0x18,
|
||||
0x46, 0x63, 0x93, 0x20, 0xa4, 0x41, 0x02, 0x81, 0xc7, 0xee, 0x20, 0x2b, 0xc1, 0x31, 0xe5, 0x78,
|
||||
0x16, 0x6c, 0xac, 0x8a, 0xab, 0xf0, 0x94, 0xd2, 0xae, 0xb2, 0xaa, 0xcb, 0xc6, 0x59, 0xf0, 0x0b,
|
||||
0xfc, 0x06, 0x5f, 0xc1, 0x0e, 0x09, 0x96, 0x6c, 0xd8, 0x22, 0x94, 0x1d, 0x7f, 0x81, 0xaa, 0xba,
|
||||
0x3b, 0x13, 0x47, 0x89, 0x33, 0x9b, 0x6c, 0xec, 0xaa, 0x7b, 0xcf, 0xb9, 0xe7, 0x9e, 0xba, 0xd5,
|
||||
0x6e, 0x43, 0x5d, 0xce, 0x98, 0xd0, 0x2c, 0x60, 0x53, 0xa6, 0xd5, 0x59, 0x73, 0xa6, 0xa4, 0x96,
|
||||
0x4d, 0xad, 0xc8, 0x98, 0x35, 0x17, 0xbb, 0xd1, 0xa2, 0x61, 0x83, 0xe8, 0x83, 0x15, 0x64, 0x14,
|
||||
0x6c, 0x44, 0x80, 0xc5, 0xee, 0xf6, 0x27, 0xd7, 0xd5, 0x19, 0xcb, 0xe9, 0x54, 0x0a, 0x53, 0x28,
|
||||
0x5a, 0x45, 0xa4, 0xed, 0xc6, 0x75, 0x58, 0xc5, 0x42, 0x39, 0x57, 0x91, 0x6c, 0xb2, 0x8e, 0xf0,
|
||||
0xb5, 0xbf, 0x1d, 0x28, 0xe3, 0x38, 0x34, 0x98, 0x11, 0x11, 0x22, 0x1f, 0x0a, 0x09, 0xc6, 0x73,
|
||||
0xaa, 0x4e, 0xbd, 0xb8, 0xf7, 0x71, 0xe3, 0xba, 0xf6, 0x2e, 0x0a, 0x2d, 0x76, 0x1b, 0x49, 0x05,
|
||||
0x7c, 0x41, 0x45, 0x3f, 0xc3, 0x87, 0x5c, 0x84, 0x5a, 0xcd, 0xa7, 0x4c, 0x68, 0xa2, 0xb9, 0x14,
|
||||
0xa3, 0x80, 0x9f, 0x28, 0xa2, 0xce, 0x46, 0xa1, 0xd1, 0xf1, 0x52, 0xd5, 0x74, 0xbd, 0xb8, 0xf7,
|
||||
0x45, 0x63, 0x9d, 0xf5, 0x46, 0x77, 0xb5, 0xc4, 0x61, 0x54, 0xc1, 0x36, 0x8a, 0x9f, 0xf2, 0x9b,
|
||||
0x93, 0xb5, 0x3f, 0x1c, 0x78, 0xba, 0x86, 0x8c, 0x04, 0x3c, 0xbe, 0xa1, 0xbd, 0xd8, 0xf4, 0xe7,
|
||||
0xd7, 0x36, 0x16, 0x9f, 0xf5, 0x8d, 0x9d, 0xe1, 0x47, 0xd7, 0x37, 0x85, 0x5e, 0x41, 0xf6, 0xb2,
|
||||
0xed, 0xda, 0x7a, 0xdb, 0xa6, 0x47, 0x1c, 0x11, 0x6a, 0xff, 0x6d, 0x40, 0xc6, 0xec, 0xd1, 0x13,
|
||||
0x28, 0x58, 0xc0, 0x88, 0x53, 0xdb, 0x63, 0x09, 0xe7, 0xed, 0xbe, 0x4b, 0xd1, 0x63, 0xc8, 0x1b,
|
||||
0xb0, 0xc9, 0xa4, 0x6c, 0x26, 0x67, 0xb6, 0x5d, 0x8a, 0x9e, 0x41, 0x31, 0xe2, 0x84, 0x9a, 0x68,
|
||||
0xe6, 0xa5, 0xab, 0x4e, 0x7d, 0x03, 0x83, 0x0d, 0x0d, 0x4c, 0x04, 0x3d, 0x87, 0xca, 0x8c, 0x28,
|
||||
0x26, 0xf4, 0x28, 0x29, 0x90, 0xb1, 0x05, 0x4a, 0x51, 0x74, 0x10, 0x95, 0x41, 0x90, 0x11, 0x64,
|
||||
0xca, 0xbc, 0xac, 0xe5, 0xdb, 0x35, 0xfa, 0x1a, 0x32, 0xa7, 0x5c, 0x50, 0x2f, 0x57, 0x75, 0xea,
|
||||
0x95, 0xbd, 0x17, 0xb7, 0x1b, 0xb2, 0x1f, 0x07, 0x5c, 0x50, 0x6c, 0x89, 0xa8, 0x09, 0x0f, 0x42,
|
||||
0x4d, 0x94, 0x1e, 0x69, 0x3e, 0x65, 0xa3, 0xb9, 0xe0, 0xcb, 0x91, 0x20, 0x42, 0x7a, 0xf9, 0xaa,
|
||||
0x53, 0xcf, 0xe1, 0x2d, 0x9b, 0x3b, 0xe6, 0x53, 0x36, 0x14, 0x7c, 0xd9, 0x23, 0x42, 0xa2, 0x17,
|
||||
0x80, 0x98, 0xa0, 0x57, 0xe1, 0x05, 0x0b, 0xdf, 0x64, 0x82, 0xae, 0x80, 0xbf, 0x05, 0x20, 0x5a,
|
||||
0x2b, 0x7e, 0x32, 0xd7, 0x2c, 0xf4, 0x36, 0xec, 0xa9, 0x7f, 0x74, 0xcb, 0x4c, 0x0f, 0xd8, 0xd9,
|
||||
0x1b, 0x12, 0xcc, 0x19, 0xbe, 0x44, 0x45, 0xaf, 0xc0, 0xa3, 0x4a, 0xce, 0x66, 0x8c, 0x8e, 0xde,
|
||||
0x45, 0x47, 0x63, 0x39, 0x17, 0xda, 0x83, 0xaa, 0x53, 0x2f, 0xe3, 0x47, 0x71, 0xbe, 0x75, 0x91,
|
||||
0x6e, 0x9b, 0x2c, 0xfa, 0x06, 0x72, 0x6c, 0xc1, 0x84, 0x0e, 0xbd, 0xa2, 0x95, 0xaf, 0xbf, 0xc7,
|
||||
0x19, 0xf9, 0x86, 0x80, 0x63, 0x1e, 0xfa, 0x14, 0x1e, 0x24, 0xda, 0x51, 0x24, 0xd6, 0x2d, 0x59,
|
||||
0x5d, 0x14, 0xe7, 0x2c, 0x27, 0xd6, 0xfc, 0x0a, 0xb2, 0x01, 0x17, 0xa7, 0xa1, 0x57, 0x5e, 0xe3,
|
||||
0x78, 0x55, 0xf2, 0x90, 0x8b, 0x53, 0x1c, 0xb1, 0x50, 0x03, 0xee, 0x27, 0x82, 0x36, 0x10, 0xeb,
|
||||
0x55, 0xac, 0xde, 0x56, 0x9c, 0x32, 0x84, 0x58, 0xee, 0x4b, 0xc8, 0x99, 0x9b, 0x35, 0x0f, 0xbd,
|
||||
0x4d, 0xfb, 0xd4, 0x3c, 0xbf, 0x45, 0xcf, 0x62, 0x71, 0xcc, 0xd9, 0xfe, 0xdd, 0x81, 0xac, 0x6d,
|
||||
0xde, 0x5c, 0xc3, 0x2b, 0x63, 0x75, 0xec, 0x58, 0x4b, 0xfa, 0xf2, 0x4c, 0x93, 0x6b, 0x98, 0xba,
|
||||
0x74, 0x0d, 0x57, 0xe7, 0x9c, 0xbe, 0x9b, 0x39, 0x67, 0xd6, 0xcd, 0x79, 0xfb, 0x1f, 0x07, 0x32,
|
||||
0xe6, 0x4c, 0xee, 0xe6, 0x09, 0x5d, 0x35, 0x98, 0xb9, 0x1b, 0x83, 0xd9, 0x75, 0x06, 0x6b, 0x13,
|
||||
0x28, 0x24, 0xcf, 0x2e, 0x7a, 0x02, 0x0f, 0x07, 0xfd, 0x56, 0x6f, 0x74, 0xd0, 0xed, 0x75, 0x46,
|
||||
0xc3, 0xde, 0xa0, 0xef, 0xb7, 0xbb, 0xfb, 0x5d, 0xbf, 0xe3, 0xde, 0x43, 0x25, 0x28, 0x74, 0x7b,
|
||||
0xc7, 0x3e, 0xee, 0xb5, 0x0e, 0x5d, 0x07, 0x01, 0xe4, 0x06, 0x3e, 0x7e, 0xe3, 0x63, 0x37, 0x65,
|
||||
0xd6, 0xed, 0xc3, 0xae, 0xdf, 0x3b, 0x76, 0xd3, 0x06, 0xd5, 0xc7, 0x47, 0x9d, 0x61, 0xdb, 0xc7,
|
||||
0x6e, 0xc6, 0xec, 0xda, 0x47, 0xbd, 0xc1, 0xf0, 0x3b, 0x1f, 0xbb, 0xd9, 0xda, 0xaf, 0x69, 0xc8,
|
||||
0x45, 0x77, 0x04, 0xb5, 0x21, 0x33, 0x96, 0x34, 0x7a, 0x05, 0x55, 0xf6, 0x9a, 0xef, 0x73, 0xaf,
|
||||
0xe2, 0xaf, 0xb6, 0xa4, 0x0c, 0x5b, 0x32, 0xf2, 0x20, 0x3f, 0x65, 0x61, 0x48, 0x26, 0xc9, 0x9d,
|
||||
0x49, 0xb6, 0xb5, 0xdf, 0x52, 0x00, 0xef, 0xe0, 0x28, 0x07, 0xa9, 0xa3, 0x53, 0xf7, 0x1e, 0x2a,
|
||||
0xc3, 0x46, 0x9b, 0x88, 0x31, 0x0b, 0x02, 0x46, 0x5d, 0x07, 0xb9, 0x50, 0x1a, 0x8a, 0x53, 0x21,
|
||||
0x7f, 0x12, 0xbe, 0x52, 0x52, 0xb9, 0x29, 0x74, 0x1f, 0x36, 0xbb, 0x62, 0x41, 0x02, 0x4e, 0x5b,
|
||||
0x6a, 0x62, 0x7f, 0xe6, 0xdd, 0x34, 0x7a, 0x00, 0x6e, 0x87, 0x11, 0x1a, 0x70, 0xc1, 0xfc, 0xe5,
|
||||
0x98, 0x31, 0xca, 0x68, 0x64, 0xad, 0x27, 0xf5, 0xbe, 0x9c, 0x0b, 0xea, 0x66, 0xd1, 0x16, 0x94,
|
||||
0x5b, 0x81, 0x62, 0x84, 0x9e, 0xf9, 0x4b, 0x1e, 0xea, 0xd0, 0xcd, 0x19, 0x5a, 0x9f, 0xa9, 0x29,
|
||||
0x0f, 0x43, 0x2e, 0x45, 0x87, 0x09, 0xce, 0xa8, 0x9b, 0x47, 0x0f, 0x61, 0x2b, 0x79, 0x9d, 0xfa,
|
||||
0xcb, 0xb7, 0x64, 0x1e, 0x6a, 0x46, 0xdd, 0x02, 0x7a, 0x04, 0x68, 0x9f, 0xf0, 0x80, 0xd1, 0xbe,
|
||||
0x62, 0x63, 0x29, 0x28, 0x37, 0x6f, 0x17, 0x77, 0x03, 0x15, 0x21, 0xdf, 0x3a, 0x91, 0xca, 0x80,
|
||||
0x00, 0x55, 0x00, 0x8e, 0xe6, 0xfa, 0xe8, 0x47, 0x4c, 0xc4, 0x84, 0xb9, 0x45, 0x23, 0x3a, 0x14,
|
||||
0x7c, 0x3a, 0x33, 0xc7, 0x26, 0x0c, 0xa4, 0x64, 0x42, 0x5d, 0xa1, 0x99, 0x12, 0x24, 0x88, 0x3c,
|
||||
0x95, 0xd1, 0x26, 0x14, 0x87, 0x82, 0x2c, 0x08, 0x0f, 0xc8, 0x49, 0xc0, 0xdc, 0x8a, 0xe9, 0xbc,
|
||||
0x43, 0x34, 0x39, 0x94, 0x61, 0xe8, 0x6e, 0x1a, 0xcb, 0x43, 0x41, 0xe6, 0xfa, 0x2d, 0x13, 0x9a,
|
||||
0x8f, 0x89, 0x29, 0xe3, 0xbe, 0xfe, 0xc5, 0xf9, 0xf3, 0x7c, 0xc7, 0xf9, 0xeb, 0x7c, 0xc7, 0xf9,
|
||||
0xf7, 0x7c, 0xc7, 0x81, 0x67, 0x5c, 0xae, 0x9d, 0xd0, 0x6b, 0x38, 0x36, 0xab, 0xbe, 0x09, 0xf6,
|
||||
0x9d, 0x1f, 0xbe, 0x9f, 0x5c, 0x85, 0x73, 0xd9, 0x94, 0x9a, 0x05, 0x4d, 0xb6, 0x9c, 0x19, 0x33,
|
||||
0x2a, 0x6c, 0x4a, 0x1d, 0xcc, 0x9a, 0x3c, 0x6e, 0xb5, 0xb9, 0x82, 0x7e, 0x69, 0x8b, 0xbf, 0x9c,
|
||||
0x30, 0x71, 0xf1, 0x77, 0xea, 0x24, 0x67, 0x63, 0x9f, 0xfd, 0x1f, 0x00, 0x00, 0xff, 0xff, 0x47,
|
||||
0x96, 0xd8, 0x1b, 0x75, 0x09, 0x00, 0x00,
|
||||
// 1088 bytes of a gzipped FileDescriptorProto
|
||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x56, 0x4f, 0x6f, 0xe3, 0xc4,
|
||||
0x1b, 0x5e, 0xb7, 0x4e, 0xda, 0x7d, 0xdb, 0x66, 0xbd, 0xb3, 0xfb, 0xdb, 0xba, 0xff, 0xf3, 0x0b,
|
||||
0x2b, 0x11, 0x58, 0x6d, 0x42, 0x8b, 0x90, 0x16, 0x09, 0x04, 0xae, 0x3d, 0xed, 0x5a, 0x75, 0xed,
|
||||
0x30, 0xb6, 0xcb, 0xc2, 0xc5, 0x72, 0x9b, 0x51, 0x65, 0x35, 0x19, 0x47, 0xb6, 0x53, 0xb5, 0x07,
|
||||
0xbe, 0x02, 0x37, 0x0e, 0x48, 0x7c, 0x1c, 0x24, 0x38, 0x72, 0xe1, 0x8a, 0x50, 0xbf, 0x08, 0x68,
|
||||
0xc6, 0x4e, 0x1b, 0x87, 0x36, 0xdd, 0x4b, 0x2f, 0xc9, 0xcc, 0xfb, 0x3e, 0xcf, 0xfb, 0xbc, 0xff,
|
||||
0x46, 0x09, 0x34, 0xe3, 0x01, 0x65, 0x19, 0xed, 0xd1, 0x3e, 0xcd, 0x92, 0xcb, 0xf6, 0x20, 0x89,
|
||||
0xb3, 0xb8, 0x9d, 0x25, 0xe1, 0x09, 0x6d, 0x9f, 0x6f, 0xe7, 0x87, 0x96, 0x30, 0xa2, 0xf5, 0x12,
|
||||
0x32, 0x37, 0xb6, 0x72, 0xc0, 0xf9, 0xf6, 0xea, 0xc7, 0xb7, 0xc5, 0x39, 0x89, 0xfb, 0xfd, 0x98,
|
||||
0xf1, 0x40, 0xf9, 0x29, 0x27, 0xad, 0xb6, 0x6e, 0xc3, 0x26, 0x34, 0x8d, 0x87, 0x49, 0x2e, 0x3b,
|
||||
0x3a, 0xe7, 0xf8, 0xc6, 0x9f, 0x12, 0x2c, 0x91, 0xc2, 0xe4, 0x0e, 0x42, 0x96, 0x22, 0x0c, 0xf3,
|
||||
0x23, 0x8c, 0x2a, 0xd5, 0xa5, 0xe6, 0xc2, 0xce, 0x47, 0xad, 0xdb, 0xd2, 0xbb, 0x0e, 0x74, 0xbe,
|
||||
0xdd, 0x1a, 0x45, 0x20, 0xd7, 0x54, 0xf4, 0x03, 0x6c, 0x44, 0x2c, 0xcd, 0x92, 0x61, 0x9f, 0xb2,
|
||||
0x2c, 0xcc, 0xa2, 0x98, 0x05, 0xbd, 0xe8, 0x38, 0x09, 0x93, 0xcb, 0x20, 0xe5, 0x3a, 0xea, 0x4c,
|
||||
0x7d, 0xb6, 0xb9, 0xb0, 0xf3, 0x79, 0x6b, 0x5a, 0xe9, 0x2d, 0xb3, 0x1c, 0xc2, 0xca, 0x23, 0x88,
|
||||
0x44, 0xc9, 0x5a, 0x74, 0xb7, 0xb3, 0xf1, 0x9b, 0x04, 0x6b, 0x53, 0xc8, 0x88, 0xc1, 0xf2, 0x1d,
|
||||
0xe9, 0x15, 0x45, 0x7f, 0x76, 0x6b, 0x62, 0x45, 0xaf, 0xef, 0xcc, 0x8c, 0xbc, 0xb8, 0x3d, 0x29,
|
||||
0xf4, 0x06, 0x2a, 0xe3, 0x65, 0x37, 0xa6, 0x97, 0xcd, 0x73, 0x24, 0x39, 0xa1, 0xf1, 0x0b, 0x80,
|
||||
0xcc, 0xef, 0x68, 0x05, 0xe6, 0x05, 0x20, 0x88, 0xba, 0x22, 0xc7, 0x45, 0x32, 0x27, 0xee, 0x66,
|
||||
0x17, 0x2d, 0xc3, 0x1c, 0x07, 0x73, 0xcf, 0x8c, 0xf0, 0x54, 0xf9, 0xd5, 0xec, 0xa2, 0x2d, 0x58,
|
||||
0xc8, 0x39, 0x69, 0x16, 0x66, 0x54, 0x9d, 0xad, 0x4b, 0xcd, 0xc7, 0x04, 0x84, 0xc9, 0xe5, 0x16,
|
||||
0xf4, 0x12, 0x6a, 0x83, 0x30, 0xa1, 0x2c, 0x0b, 0x46, 0x01, 0x64, 0x11, 0x60, 0x31, 0xb7, 0xba,
|
||||
0x79, 0x18, 0x04, 0x32, 0x0b, 0xfb, 0x54, 0xad, 0x08, 0xbe, 0x38, 0xa3, 0xaf, 0x40, 0x3e, 0x8b,
|
||||
0x58, 0x57, 0xad, 0xd6, 0xa5, 0x66, 0x6d, 0xe7, 0xd5, 0xfd, 0x05, 0x89, 0x8f, 0x83, 0x88, 0x75,
|
||||
0x89, 0x20, 0xa2, 0x36, 0x3c, 0x4f, 0xb3, 0x30, 0xc9, 0x82, 0x2c, 0xea, 0xd3, 0x60, 0xc8, 0xa2,
|
||||
0x8b, 0x80, 0x85, 0x2c, 0x56, 0xe7, 0xea, 0x52, 0xb3, 0x4a, 0x9e, 0x0a, 0x9f, 0x17, 0xf5, 0xa9,
|
||||
0xcf, 0xa2, 0x0b, 0x3b, 0x64, 0x31, 0x7a, 0x05, 0x88, 0xb2, 0xee, 0x24, 0x7c, 0x5e, 0xc0, 0x9f,
|
||||
0x50, 0xd6, 0x2d, 0x81, 0xf7, 0x01, 0xc2, 0x2c, 0x4b, 0xa2, 0xe3, 0x61, 0x46, 0x53, 0xf5, 0xb1,
|
||||
0xe8, 0xfa, 0x87, 0xf7, 0xcc, 0xf4, 0x80, 0x5e, 0x1e, 0x85, 0xbd, 0x21, 0x25, 0x63, 0x54, 0xf4,
|
||||
0x06, 0xd4, 0x6e, 0x12, 0x0f, 0x06, 0xb4, 0x1b, 0xdc, 0x58, 0x83, 0x93, 0x78, 0xc8, 0x32, 0x15,
|
||||
0xea, 0x52, 0x73, 0x89, 0xbc, 0x28, 0xfc, 0xda, 0xb5, 0x5b, 0xe7, 0x5e, 0xf4, 0x35, 0x54, 0xe9,
|
||||
0x39, 0x65, 0x59, 0xaa, 0x2e, 0x08, 0xf9, 0xe6, 0x7b, 0xf4, 0x08, 0x73, 0x02, 0x29, 0x78, 0xe8,
|
||||
0x13, 0x78, 0x3e, 0xd2, 0xce, 0x2d, 0x85, 0xee, 0xa2, 0xd0, 0x45, 0x85, 0x4f, 0x70, 0x0a, 0xcd,
|
||||
0x2f, 0xa1, 0xd2, 0x8b, 0xd8, 0x59, 0xaa, 0x2e, 0x4d, 0xa9, 0xb8, 0x2c, 0x69, 0x45, 0xec, 0x8c,
|
||||
0xe4, 0x2c, 0xd4, 0x82, 0x67, 0x23, 0x41, 0x61, 0x28, 0xf4, 0x6a, 0x42, 0xef, 0x69, 0xe1, 0xe2,
|
||||
0x84, 0x42, 0xee, 0x0b, 0xa8, 0xf2, 0xcd, 0x1a, 0xa6, 0xea, 0x13, 0xf1, 0x6a, 0x5e, 0xde, 0xa3,
|
||||
0x27, 0xb0, 0xa4, 0xe0, 0xac, 0xfe, 0x2a, 0x41, 0x45, 0x24, 0xcf, 0xd7, 0x70, 0x62, 0xac, 0x92,
|
||||
0x18, 0xeb, 0x62, 0x36, 0x3e, 0xd3, 0xd1, 0x1a, 0xce, 0x8c, 0xad, 0x61, 0x79, 0xce, 0xb3, 0x0f,
|
||||
0x33, 0x67, 0x79, 0xda, 0x9c, 0x57, 0xff, 0x92, 0x40, 0xe6, 0x3d, 0x79, 0x98, 0x17, 0x5a, 0x2e,
|
||||
0x50, 0x7e, 0x98, 0x02, 0x2b, 0xd3, 0x0a, 0x6c, 0xfc, 0x2c, 0xc1, 0xfc, 0xe8, 0xf1, 0xa2, 0x15,
|
||||
0xf8, 0x9f, 0xdb, 0xd1, 0xec, 0xe0, 0xc0, 0xb4, 0x8d, 0xc0, 0xb7, 0xdd, 0x0e, 0xd6, 0xcd, 0x3d,
|
||||
0x13, 0x1b, 0xca, 0x23, 0xf4, 0x02, 0xd0, 0x8d, 0xcb, 0xb4, 0x3d, 0x4c, 0x6c, 0xcd, 0x52, 0x24,
|
||||
0xf4, 0x1c, 0x94, 0x1b, 0xbb, 0x8b, 0xc9, 0x11, 0x26, 0xca, 0x4c, 0xd9, 0xaa, 0x5b, 0x26, 0xb6,
|
||||
0x3d, 0x65, 0xb6, 0x1c, 0xa3, 0x43, 0x1c, 0xc3, 0xd7, 0x31, 0x51, 0xe4, 0xb2, 0x5d, 0x77, 0x6c,
|
||||
0xd7, 0x3f, 0xc4, 0x44, 0xa9, 0x34, 0xfe, 0x91, 0xa1, 0x9a, 0xaf, 0x15, 0xd2, 0x41, 0x3e, 0x89,
|
||||
0xbb, 0xf9, 0xaf, 0x56, 0x6d, 0xa7, 0xfd, 0x3e, 0xab, 0x58, 0x7c, 0xe9, 0x71, 0x97, 0x12, 0x41,
|
||||
0x46, 0x2a, 0xcc, 0xf5, 0x69, 0x9a, 0x86, 0xa7, 0xa3, 0x35, 0x1b, 0x5d, 0x1b, 0x3f, 0xc9, 0x00,
|
||||
0x37, 0x70, 0x84, 0xa0, 0xe6, 0x7a, 0x9a, 0xe7, 0xbb, 0x81, 0xee, 0x18, 0x38, 0x70, 0x0e, 0x94,
|
||||
0x47, 0xa2, 0x37, 0x63, 0x36, 0x5d, 0xb3, 0x75, 0x6c, 0x59, 0xd8, 0x50, 0x24, 0xb4, 0x01, 0x2b,
|
||||
0xe3, 0x2e, 0xdf, 0x3e, 0xb0, 0x9d, 0x6f, 0xed, 0x00, 0x13, 0xe2, 0xf0, 0x66, 0xd4, 0x61, 0x7d,
|
||||
0xdc, 0x6d, 0xda, 0x47, 0x9a, 0x65, 0x1a, 0x81, 0x46, 0xf6, 0xfd, 0xc3, 0xbc, 0x31, 0xff, 0x87,
|
||||
0x8d, 0x71, 0x84, 0x81, 0x35, 0xc3, 0x32, 0x6d, 0x1c, 0xe0, 0x77, 0x3a, 0xc6, 0x06, 0x36, 0x14,
|
||||
0x79, 0x52, 0xde, 0x76, 0xbc, 0x60, 0xcf, 0xf1, 0x6d, 0x43, 0xa9, 0xa0, 0x4d, 0x58, 0x1d, 0x77,
|
||||
0x69, 0x16, 0xc1, 0x9a, 0xf1, 0x5d, 0x80, 0xdf, 0x99, 0xae, 0xe7, 0x2a, 0xd5, 0xc9, 0xe8, 0x1d,
|
||||
0x4c, 0x0e, 0x4d, 0xd7, 0x35, 0x1d, 0x3b, 0x30, 0xb0, 0xcd, 0xa7, 0x3b, 0x87, 0x1a, 0xb0, 0x39,
|
||||
0x0e, 0x21, 0xd8, 0x75, 0x7c, 0xa2, 0xf3, 0x04, 0xde, 0x6a, 0xbe, 0xeb, 0x61, 0x43, 0x99, 0x47,
|
||||
0x1f, 0xc0, 0xd6, 0x38, 0x66, 0x4f, 0x33, 0x2d, 0xcc, 0xc7, 0x88, 0x75, 0xc7, 0x36, 0x4c, 0xcf,
|
||||
0x74, 0x6c, 0xe5, 0x31, 0x5a, 0x86, 0x67, 0xa5, 0x5c, 0x76, 0x1d, 0xc2, 0xd9, 0x80, 0xd6, 0x41,
|
||||
0x2d, 0xb5, 0xd4, 0xf7, 0x02, 0x67, 0x2f, 0x20, 0x9a, 0xbd, 0x8f, 0x95, 0x85, 0xff, 0x76, 0xd0,
|
||||
0x3c, 0xec, 0x58, 0x98, 0x77, 0x07, 0x1b, 0xca, 0xe2, 0x64, 0x85, 0xa3, 0xf5, 0x2b, 0x3a, 0xbc,
|
||||
0x84, 0xd6, 0x60, 0xb9, 0x4c, 0xd7, 0x8e, 0x34, 0xd3, 0xd2, 0x76, 0x2d, 0xac, 0xd4, 0x26, 0x3b,
|
||||
0x67, 0x68, 0x9e, 0x16, 0x58, 0x8e, 0xeb, 0x2a, 0x4f, 0xd0, 0x16, 0xac, 0x4d, 0xf0, 0x7c, 0xef,
|
||||
0x2d, 0xb6, 0x3d, 0x53, 0xd7, 0xb8, 0xb0, 0xb2, 0xfb, 0xa3, 0xf4, 0xfb, 0xd5, 0xa6, 0xf4, 0xc7,
|
||||
0xd5, 0xa6, 0xf4, 0xf7, 0xd5, 0xa6, 0x04, 0x5b, 0x51, 0x3c, 0x75, 0xf3, 0x76, 0xc1, 0xe3, 0xa7,
|
||||
0x0e, 0x37, 0x76, 0xa4, 0xef, 0xbf, 0x39, 0x9d, 0x84, 0x47, 0x71, 0x3b, 0xce, 0x68, 0xaf, 0x4d,
|
||||
0x2f, 0x06, 0x71, 0x92, 0xd1, 0x24, 0x6d, 0xc7, 0x59, 0x6f, 0xd0, 0x8e, 0x58, 0x46, 0x13, 0x16,
|
||||
0xf6, 0xda, 0x25, 0xf4, 0x6b, 0x11, 0xfc, 0xf5, 0x29, 0x65, 0xd7, 0xff, 0x2c, 0x8f, 0xab, 0xc2,
|
||||
0xf6, 0xe9, 0xbf, 0x01, 0x00, 0x00, 0xff, 0xff, 0xcd, 0xf7, 0xd2, 0xe8, 0x80, 0x0a, 0x00, 0x00,
|
||||
}
|
||||
|
||||
func (m *ResourceSpans) Marshal() (dAtA []byte, err error) {
|
||||
|
@ -193,23 +193,26 @@ func sink(ctx context.Context, in <-chan result) ([]*metricpb.ResourceMetrics, e
|
||||
rb.InstrumentationLibraryBatches[res.InstrumentationLibrary] = mb
|
||||
}
|
||||
|
||||
mID := res.Metric.GetMetricDescriptor().String()
|
||||
mID := res.Metric.GetName()
|
||||
m, ok := mb[mID]
|
||||
if !ok {
|
||||
mb[mID] = res.Metric
|
||||
continue
|
||||
}
|
||||
if len(res.Metric.Int64DataPoints) > 0 {
|
||||
m.Int64DataPoints = append(m.Int64DataPoints, res.Metric.Int64DataPoints...)
|
||||
}
|
||||
if len(res.Metric.DoubleDataPoints) > 0 {
|
||||
m.DoubleDataPoints = append(m.DoubleDataPoints, res.Metric.DoubleDataPoints...)
|
||||
}
|
||||
if len(res.Metric.HistogramDataPoints) > 0 {
|
||||
m.HistogramDataPoints = append(m.HistogramDataPoints, res.Metric.HistogramDataPoints...)
|
||||
}
|
||||
if len(res.Metric.SummaryDataPoints) > 0 {
|
||||
m.SummaryDataPoints = append(m.SummaryDataPoints, res.Metric.SummaryDataPoints...)
|
||||
switch res.Metric.Data.(type) {
|
||||
case *metricpb.Metric_IntGauge:
|
||||
m.GetIntGauge().DataPoints = append(m.GetIntGauge().DataPoints, res.Metric.GetIntGauge().DataPoints...)
|
||||
case *metricpb.Metric_IntHistogram:
|
||||
m.GetIntHistogram().DataPoints = append(m.GetIntHistogram().DataPoints, res.Metric.GetIntHistogram().DataPoints...)
|
||||
case *metricpb.Metric_IntSum:
|
||||
m.GetIntSum().DataPoints = append(m.GetIntSum().DataPoints, res.Metric.GetIntSum().DataPoints...)
|
||||
case *metricpb.Metric_DoubleGauge:
|
||||
m.GetDoubleGauge().DataPoints = append(m.GetDoubleGauge().DataPoints, res.Metric.GetDoubleGauge().DataPoints...)
|
||||
case *metricpb.Metric_DoubleHistogram:
|
||||
m.GetDoubleHistogram().DataPoints = append(m.GetDoubleHistogram().DataPoints, res.Metric.GetDoubleHistogram().DataPoints...)
|
||||
case *metricpb.Metric_DoubleSum:
|
||||
m.GetDoubleSum().DataPoints = append(m.GetDoubleSum().DataPoints, res.Metric.GetDoubleSum().DataPoints...)
|
||||
default:
|
||||
}
|
||||
}
|
||||
|
||||
@ -284,13 +287,57 @@ func Record(r export.Record) (*metricpb.Metric, error) {
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return scalar(r, value, time.Time{}, tm)
|
||||
return gauge(r, value, time.Time{}, tm)
|
||||
|
||||
default:
|
||||
return nil, fmt.Errorf("%w: %T", ErrUnimplementedAgg, agg)
|
||||
}
|
||||
}
|
||||
|
||||
func gauge(record export.Record, num metric.Number, start, end time.Time) (*metricpb.Metric, error) {
|
||||
desc := record.Descriptor()
|
||||
labels := record.Labels()
|
||||
|
||||
m := &metricpb.Metric{
|
||||
Name: desc.Name(),
|
||||
Description: desc.Description(),
|
||||
Unit: string(desc.Unit()),
|
||||
}
|
||||
|
||||
switch n := desc.NumberKind(); n {
|
||||
case metric.Int64NumberKind:
|
||||
m.Data = &metricpb.Metric_IntGauge{
|
||||
IntGauge: &metricpb.IntGauge{
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: num.CoerceToInt64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(start),
|
||||
TimeUnixNano: toNanos(end),
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
case metric.Float64NumberKind:
|
||||
m.Data = &metricpb.Metric_DoubleGauge{
|
||||
DoubleGauge: &metricpb.DoubleGauge{
|
||||
DataPoints: []*metricpb.DoubleDataPoint{
|
||||
{
|
||||
Value: num.CoerceToFloat64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(start),
|
||||
TimeUnixNano: toNanos(end),
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
default:
|
||||
return nil, fmt.Errorf("%w: %v", ErrUnknownValueType, n)
|
||||
}
|
||||
|
||||
return m, nil
|
||||
}
|
||||
|
||||
// scalar transforms a Sum or LastValue Aggregator into an OTLP Metric.
|
||||
// For LastValue (Gauge), use start==time.Time{}.
|
||||
func scalar(record export.Record, num metric.Number, start, end time.Time) (*metricpb.Metric, error) {
|
||||
@ -298,32 +345,36 @@ func scalar(record export.Record, num metric.Number, start, end time.Time) (*met
|
||||
labels := record.Labels()
|
||||
|
||||
m := &metricpb.Metric{
|
||||
MetricDescriptor: &metricpb.MetricDescriptor{
|
||||
Name: desc.Name(),
|
||||
Description: desc.Description(),
|
||||
Unit: string(desc.Unit()),
|
||||
},
|
||||
Name: desc.Name(),
|
||||
Description: desc.Description(),
|
||||
Unit: string(desc.Unit()),
|
||||
}
|
||||
|
||||
switch n := desc.NumberKind(); n {
|
||||
case metric.Int64NumberKind:
|
||||
m.MetricDescriptor.Type = metricpb.MetricDescriptor_INT64
|
||||
m.Int64DataPoints = []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: num.CoerceToInt64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(start),
|
||||
TimeUnixNano: toNanos(end),
|
||||
m.Data = &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: num.CoerceToInt64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(start),
|
||||
TimeUnixNano: toNanos(end),
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
case metric.Float64NumberKind:
|
||||
m.MetricDescriptor.Type = metricpb.MetricDescriptor_DOUBLE
|
||||
m.DoubleDataPoints = []*metricpb.DoubleDataPoint{
|
||||
{
|
||||
Value: num.CoerceToFloat64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(start),
|
||||
TimeUnixNano: toNanos(end),
|
||||
m.Data = &metricpb.Metric_DoubleSum{
|
||||
DoubleSum: &metricpb.DoubleSum{
|
||||
DataPoints: []*metricpb.DoubleDataPoint{
|
||||
{
|
||||
Value: num.CoerceToFloat64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(start),
|
||||
TimeUnixNano: toNanos(end),
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
default:
|
||||
@ -360,34 +411,52 @@ func minMaxSumCount(record export.Record, a aggregation.MinMaxSumCount) (*metric
|
||||
return nil, err
|
||||
}
|
||||
|
||||
numKind := desc.NumberKind()
|
||||
return &metricpb.Metric{
|
||||
MetricDescriptor: &metricpb.MetricDescriptor{
|
||||
Name: desc.Name(),
|
||||
Description: desc.Description(),
|
||||
Unit: string(desc.Unit()),
|
||||
Type: metricpb.MetricDescriptor_SUMMARY,
|
||||
},
|
||||
SummaryDataPoints: []*metricpb.SummaryDataPoint{
|
||||
{
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
Count: uint64(count),
|
||||
Sum: sum.CoerceToFloat64(numKind),
|
||||
PercentileValues: []*metricpb.SummaryDataPoint_ValueAtPercentile{
|
||||
m := &metricpb.Metric{
|
||||
Name: desc.Name(),
|
||||
Description: desc.Description(),
|
||||
Unit: string(desc.Unit()),
|
||||
}
|
||||
|
||||
buckets := []uint64{min.AsRaw(), max.AsRaw()}
|
||||
bounds := []float64{0.0, 100.0}
|
||||
|
||||
switch n := desc.NumberKind(); n {
|
||||
case metric.Int64NumberKind:
|
||||
m.Data = &metricpb.Metric_IntHistogram{
|
||||
IntHistogram: &metricpb.IntHistogram{
|
||||
DataPoints: []*metricpb.IntHistogramDataPoint{
|
||||
{
|
||||
Percentile: 0.0,
|
||||
Value: min.CoerceToFloat64(numKind),
|
||||
},
|
||||
{
|
||||
Percentile: 100.0,
|
||||
Value: max.CoerceToFloat64(numKind),
|
||||
Sum: sum.CoerceToInt64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(record.StartTime()),
|
||||
TimeUnixNano: toNanos(record.EndTime()),
|
||||
Count: uint64(count),
|
||||
BucketCounts: buckets,
|
||||
ExplicitBounds: bounds,
|
||||
},
|
||||
},
|
||||
StartTimeUnixNano: toNanos(record.StartTime()),
|
||||
TimeUnixNano: toNanos(record.EndTime()),
|
||||
},
|
||||
},
|
||||
}, nil
|
||||
}
|
||||
case metric.Float64NumberKind:
|
||||
m.Data = &metricpb.Metric_DoubleHistogram{
|
||||
DoubleHistogram: &metricpb.DoubleHistogram{
|
||||
DataPoints: []*metricpb.DoubleHistogramDataPoint{
|
||||
{
|
||||
Sum: sum.CoerceToFloat64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(record.StartTime()),
|
||||
TimeUnixNano: toNanos(record.EndTime()),
|
||||
Count: uint64(count),
|
||||
BucketCounts: buckets,
|
||||
ExplicitBounds: bounds,
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
default:
|
||||
return nil, fmt.Errorf("%w: %v", ErrUnknownValueType, n)
|
||||
}
|
||||
return m, nil
|
||||
}
|
||||
|
||||
func histogramValues(a aggregation.Histogram) (boundaries []float64, counts []float64, err error) {
|
||||
@ -422,33 +491,53 @@ func histogram(record export.Record, a aggregation.Histogram) (*metricpb.Metric,
|
||||
return nil, err
|
||||
}
|
||||
|
||||
buckets := make([]*metricpb.HistogramDataPoint_Bucket, len(counts))
|
||||
buckets := make([]uint64, len(counts))
|
||||
for i := 0; i < len(counts); i++ {
|
||||
buckets[i] = &metricpb.HistogramDataPoint_Bucket{
|
||||
Count: uint64(counts[i]),
|
||||
buckets[i] = uint64(counts[i])
|
||||
}
|
||||
m := &metricpb.Metric{
|
||||
Name: desc.Name(),
|
||||
Description: desc.Description(),
|
||||
Unit: string(desc.Unit()),
|
||||
}
|
||||
switch n := desc.NumberKind(); n {
|
||||
case metric.Int64NumberKind:
|
||||
m.Data = &metricpb.Metric_IntHistogram{
|
||||
IntHistogram: &metricpb.IntHistogram{
|
||||
DataPoints: []*metricpb.IntHistogramDataPoint{
|
||||
{
|
||||
Sum: sum.CoerceToInt64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(record.StartTime()),
|
||||
TimeUnixNano: toNanos(record.EndTime()),
|
||||
Count: uint64(count),
|
||||
BucketCounts: buckets,
|
||||
ExplicitBounds: boundaries,
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
case metric.Float64NumberKind:
|
||||
m.Data = &metricpb.Metric_DoubleHistogram{
|
||||
DoubleHistogram: &metricpb.DoubleHistogram{
|
||||
DataPoints: []*metricpb.DoubleHistogramDataPoint{
|
||||
{
|
||||
Sum: sum.CoerceToFloat64(n),
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(record.StartTime()),
|
||||
TimeUnixNano: toNanos(record.EndTime()),
|
||||
Count: uint64(count),
|
||||
BucketCounts: buckets,
|
||||
ExplicitBounds: boundaries,
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
default:
|
||||
return nil, fmt.Errorf("%w: %v", ErrUnknownValueType, n)
|
||||
}
|
||||
|
||||
numKind := desc.NumberKind()
|
||||
return &metricpb.Metric{
|
||||
MetricDescriptor: &metricpb.MetricDescriptor{
|
||||
Name: desc.Name(),
|
||||
Description: desc.Description(),
|
||||
Unit: string(desc.Unit()),
|
||||
Type: metricpb.MetricDescriptor_HISTOGRAM,
|
||||
},
|
||||
HistogramDataPoints: []*metricpb.HistogramDataPoint{
|
||||
{
|
||||
Labels: stringKeyValues(labels.Iter()),
|
||||
StartTimeUnixNano: toNanos(record.StartTime()),
|
||||
TimeUnixNano: toNanos(record.EndTime()),
|
||||
Count: uint64(count),
|
||||
Sum: sum.CoerceToFloat64(numKind),
|
||||
Buckets: buckets,
|
||||
ExplicitBounds: boundaries,
|
||||
},
|
||||
},
|
||||
}, nil
|
||||
return m, nil
|
||||
}
|
||||
|
||||
// stringKeyValues transforms a label iterator into an OTLP StringKeyValues.
|
||||
|
@ -39,7 +39,6 @@ import (
|
||||
"go.opentelemetry.io/otel/sdk/metric/aggregator/sum"
|
||||
sumAgg "go.opentelemetry.io/otel/sdk/metric/aggregator/sum"
|
||||
"go.opentelemetry.io/otel/sdk/resource"
|
||||
"go.opentelemetry.io/otel/unit"
|
||||
)
|
||||
|
||||
var (
|
||||
@ -117,65 +116,6 @@ func TestMinMaxSumCountValue(t *testing.T) {
|
||||
}
|
||||
}
|
||||
|
||||
func TestMinMaxSumCountMetricDescriptor(t *testing.T) {
|
||||
tests := []struct {
|
||||
name string
|
||||
metricKind metric.Kind
|
||||
description string
|
||||
unit unit.Unit
|
||||
numberKind metric.NumberKind
|
||||
labels []label.KeyValue
|
||||
expected *metricpb.MetricDescriptor
|
||||
}{
|
||||
{
|
||||
"mmsc-test-a",
|
||||
metric.ValueRecorderKind,
|
||||
"test-a-description",
|
||||
unit.Dimensionless,
|
||||
metric.Int64NumberKind,
|
||||
[]label.KeyValue{},
|
||||
&metricpb.MetricDescriptor{
|
||||
Name: "mmsc-test-a",
|
||||
Description: "test-a-description",
|
||||
Unit: "1",
|
||||
Type: metricpb.MetricDescriptor_SUMMARY,
|
||||
},
|
||||
},
|
||||
{
|
||||
"mmsc-test-b",
|
||||
metric.CounterKind, // This shouldn't change anything.
|
||||
"test-b-description",
|
||||
unit.Bytes,
|
||||
metric.Float64NumberKind, // This shouldn't change anything.
|
||||
[]label.KeyValue{label.String("A", "1")},
|
||||
&metricpb.MetricDescriptor{
|
||||
Name: "mmsc-test-b",
|
||||
Description: "test-b-description",
|
||||
Unit: "By",
|
||||
Type: metricpb.MetricDescriptor_SUMMARY,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
ctx := context.Background()
|
||||
mmsc, ckpt := metrictest.Unslice2(minmaxsumcount.New(2, &metric.Descriptor{}))
|
||||
if !assert.NoError(t, mmsc.Update(ctx, 1, &metric.Descriptor{})) {
|
||||
return
|
||||
}
|
||||
require.NoError(t, mmsc.SynchronizedMove(ckpt, &metric.Descriptor{}))
|
||||
for _, test := range tests {
|
||||
desc := metric.NewDescriptor(test.name, test.metricKind, test.numberKind,
|
||||
metric.WithDescription(test.description),
|
||||
metric.WithUnit(test.unit))
|
||||
labels := label.NewSet(test.labels...)
|
||||
record := export.NewRecord(&desc, &labels, nil, ckpt.Aggregation(), intervalStart, intervalEnd)
|
||||
got, err := minMaxSumCount(record, ckpt.(aggregation.MinMaxSumCount))
|
||||
if assert.NoError(t, err) {
|
||||
assert.Equal(t, test.expected, got.MetricDescriptor)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestMinMaxSumCountDatapoints(t *testing.T) {
|
||||
desc := metric.NewDescriptor("", metric.ValueRecorderKind, metric.Int64NumberKind)
|
||||
labels := label.NewSet()
|
||||
@ -184,20 +124,12 @@ func TestMinMaxSumCountDatapoints(t *testing.T) {
|
||||
assert.NoError(t, mmsc.Update(context.Background(), 1, &desc))
|
||||
assert.NoError(t, mmsc.Update(context.Background(), 10, &desc))
|
||||
require.NoError(t, mmsc.SynchronizedMove(ckpt, &desc))
|
||||
expected := []*metricpb.SummaryDataPoint{
|
||||
expected := []*metricpb.IntHistogramDataPoint{
|
||||
{
|
||||
Count: 2,
|
||||
Sum: 11,
|
||||
PercentileValues: []*metricpb.SummaryDataPoint_ValueAtPercentile{
|
||||
{
|
||||
Percentile: 0.0,
|
||||
Value: 1,
|
||||
},
|
||||
{
|
||||
Percentile: 100.0,
|
||||
Value: 10,
|
||||
},
|
||||
},
|
||||
Count: 2,
|
||||
Sum: 11,
|
||||
ExplicitBounds: []float64{0.0, 100.0},
|
||||
BucketCounts: []uint64{1, 10},
|
||||
StartTimeUnixNano: uint64(intervalStart.UnixNano()),
|
||||
TimeUnixNano: uint64(intervalEnd.UnixNano()),
|
||||
},
|
||||
@ -205,10 +137,11 @@ func TestMinMaxSumCountDatapoints(t *testing.T) {
|
||||
record := export.NewRecord(&desc, &labels, nil, ckpt.Aggregation(), intervalStart, intervalEnd)
|
||||
m, err := minMaxSumCount(record, ckpt.(aggregation.MinMaxSumCount))
|
||||
if assert.NoError(t, err) {
|
||||
assert.Equal(t, []*metricpb.Int64DataPoint(nil), m.Int64DataPoints)
|
||||
assert.Equal(t, []*metricpb.DoubleDataPoint(nil), m.DoubleDataPoints)
|
||||
assert.Equal(t, []*metricpb.HistogramDataPoint(nil), m.HistogramDataPoints)
|
||||
assert.Equal(t, expected, m.SummaryDataPoints)
|
||||
assert.Nil(t, m.GetIntGauge())
|
||||
assert.Equal(t, expected, m.GetIntHistogram().DataPoints)
|
||||
assert.Nil(t, m.GetIntSum())
|
||||
assert.Nil(t, m.GetDoubleGauge())
|
||||
assert.Nil(t, m.GetDoubleHistogram())
|
||||
}
|
||||
}
|
||||
|
||||
@ -222,62 +155,7 @@ func TestMinMaxSumCountPropagatesErrors(t *testing.T) {
|
||||
assert.Equal(t, aggregation.ErrNoData, err)
|
||||
}
|
||||
|
||||
func TestSumMetricDescriptor(t *testing.T) {
|
||||
tests := []struct {
|
||||
name string
|
||||
metricKind metric.Kind
|
||||
description string
|
||||
unit unit.Unit
|
||||
numberKind metric.NumberKind
|
||||
labels []label.KeyValue
|
||||
expected *metricpb.MetricDescriptor
|
||||
}{
|
||||
{
|
||||
"sum-test-a",
|
||||
metric.CounterKind,
|
||||
"test-a-description",
|
||||
unit.Dimensionless,
|
||||
metric.Int64NumberKind,
|
||||
[]label.KeyValue{},
|
||||
&metricpb.MetricDescriptor{
|
||||
Name: "sum-test-a",
|
||||
Description: "test-a-description",
|
||||
Unit: "1",
|
||||
Type: metricpb.MetricDescriptor_INT64,
|
||||
},
|
||||
},
|
||||
{
|
||||
"sum-test-b",
|
||||
metric.ValueObserverKind, // This shouldn't change anything.
|
||||
"test-b-description",
|
||||
unit.Milliseconds,
|
||||
metric.Float64NumberKind,
|
||||
[]label.KeyValue{label.String("A", "1")},
|
||||
&metricpb.MetricDescriptor{
|
||||
Name: "sum-test-b",
|
||||
Description: "test-b-description",
|
||||
Unit: "ms",
|
||||
Type: metricpb.MetricDescriptor_DOUBLE,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
for _, test := range tests {
|
||||
desc := metric.NewDescriptor(test.name, test.metricKind, test.numberKind,
|
||||
metric.WithDescription(test.description),
|
||||
metric.WithUnit(test.unit),
|
||||
)
|
||||
labels := label.NewSet(test.labels...)
|
||||
emptyAgg := &sumAgg.New(1)[0]
|
||||
record := export.NewRecord(&desc, &labels, nil, emptyAgg, intervalStart, intervalEnd)
|
||||
got, err := scalar(record, 0, time.Time{}, time.Time{})
|
||||
if assert.NoError(t, err) {
|
||||
assert.Equal(t, test.expected, got.MetricDescriptor)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestSumInt64DataPoints(t *testing.T) {
|
||||
func TestSumIntDataPoints(t *testing.T) {
|
||||
desc := metric.NewDescriptor("", metric.ValueRecorderKind, metric.Int64NumberKind)
|
||||
labels := label.NewSet()
|
||||
s, ckpt := metrictest.Unslice2(sumAgg.New(2))
|
||||
@ -290,18 +168,19 @@ func TestSumInt64DataPoints(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
|
||||
if m, err := scalar(record, value, record.StartTime(), record.EndTime()); assert.NoError(t, err) {
|
||||
assert.Equal(t, []*metricpb.Int64DataPoint{{
|
||||
assert.Nil(t, m.GetIntGauge())
|
||||
assert.Nil(t, m.GetIntHistogram())
|
||||
assert.Equal(t, []*metricpb.IntDataPoint{{
|
||||
Value: 1,
|
||||
StartTimeUnixNano: uint64(intervalStart.UnixNano()),
|
||||
TimeUnixNano: uint64(intervalEnd.UnixNano()),
|
||||
}}, m.Int64DataPoints)
|
||||
assert.Equal(t, []*metricpb.DoubleDataPoint(nil), m.DoubleDataPoints)
|
||||
assert.Equal(t, []*metricpb.HistogramDataPoint(nil), m.HistogramDataPoints)
|
||||
assert.Equal(t, []*metricpb.SummaryDataPoint(nil), m.SummaryDataPoints)
|
||||
}}, m.GetIntSum().DataPoints)
|
||||
assert.Nil(t, m.GetDoubleGauge())
|
||||
assert.Nil(t, m.GetDoubleHistogram())
|
||||
}
|
||||
}
|
||||
|
||||
func TestSumFloat64DataPoints(t *testing.T) {
|
||||
func TestSumFloatDataPoints(t *testing.T) {
|
||||
desc := metric.NewDescriptor("", metric.ValueRecorderKind, metric.Float64NumberKind)
|
||||
labels := label.NewSet()
|
||||
s, ckpt := metrictest.Unslice2(sumAgg.New(2))
|
||||
@ -314,18 +193,20 @@ func TestSumFloat64DataPoints(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
|
||||
if m, err := scalar(record, value, record.StartTime(), record.EndTime()); assert.NoError(t, err) {
|
||||
assert.Equal(t, []*metricpb.Int64DataPoint(nil), m.Int64DataPoints)
|
||||
assert.Nil(t, m.GetIntGauge())
|
||||
assert.Nil(t, m.GetIntHistogram())
|
||||
assert.Nil(t, m.GetIntSum())
|
||||
assert.Nil(t, m.GetDoubleGauge())
|
||||
assert.Nil(t, m.GetDoubleHistogram())
|
||||
assert.Equal(t, []*metricpb.DoubleDataPoint{{
|
||||
Value: 1,
|
||||
StartTimeUnixNano: uint64(intervalStart.UnixNano()),
|
||||
TimeUnixNano: uint64(intervalEnd.UnixNano()),
|
||||
}}, m.DoubleDataPoints)
|
||||
assert.Equal(t, []*metricpb.HistogramDataPoint(nil), m.HistogramDataPoints)
|
||||
assert.Equal(t, []*metricpb.SummaryDataPoint(nil), m.SummaryDataPoints)
|
||||
}}, m.GetDoubleSum().DataPoints)
|
||||
}
|
||||
}
|
||||
|
||||
func TestLastValueInt64DataPoints(t *testing.T) {
|
||||
func TestLastValueIntDataPoints(t *testing.T) {
|
||||
desc := metric.NewDescriptor("", metric.ValueRecorderKind, metric.Int64NumberKind)
|
||||
labels := label.NewSet()
|
||||
s, ckpt := metrictest.Unslice2(lvAgg.New(2))
|
||||
@ -337,15 +218,17 @@ func TestLastValueInt64DataPoints(t *testing.T) {
|
||||
value, timestamp, err := sum.LastValue()
|
||||
require.NoError(t, err)
|
||||
|
||||
if m, err := scalar(record, value, time.Time{}, timestamp); assert.NoError(t, err) {
|
||||
assert.Equal(t, []*metricpb.Int64DataPoint{{
|
||||
if m, err := gauge(record, value, time.Time{}, timestamp); assert.NoError(t, err) {
|
||||
assert.Equal(t, []*metricpb.IntDataPoint{{
|
||||
Value: 100,
|
||||
StartTimeUnixNano: 0,
|
||||
TimeUnixNano: uint64(timestamp.UnixNano()),
|
||||
}}, m.Int64DataPoints)
|
||||
assert.Equal(t, []*metricpb.DoubleDataPoint(nil), m.DoubleDataPoints)
|
||||
assert.Equal(t, []*metricpb.HistogramDataPoint(nil), m.HistogramDataPoints)
|
||||
assert.Equal(t, []*metricpb.SummaryDataPoint(nil), m.SummaryDataPoints)
|
||||
}}, m.GetIntGauge().DataPoints)
|
||||
assert.Nil(t, m.GetIntHistogram())
|
||||
assert.Nil(t, m.GetIntSum())
|
||||
assert.Nil(t, m.GetDoubleGauge())
|
||||
assert.Nil(t, m.GetDoubleHistogram())
|
||||
assert.Nil(t, m.GetDoubleSum())
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -129,9 +129,9 @@ func status(status codes.Code, message string) *tracepb.Status {
|
||||
var c tracepb.Status_StatusCode
|
||||
switch status {
|
||||
case codes.Error:
|
||||
c = tracepb.Status_UnknownError
|
||||
c = tracepb.Status_STATUS_CODE_UNKNOWN_ERROR
|
||||
default:
|
||||
c = tracepb.Status_Ok
|
||||
c = tracepb.Status_STATUS_CODE_OK
|
||||
}
|
||||
return &tracepb.Status{
|
||||
Code: c,
|
||||
@ -196,15 +196,15 @@ func spanEvents(es []export.Event) []*tracepb.Span_Event {
|
||||
func spanKind(kind otel.SpanKind) tracepb.Span_SpanKind {
|
||||
switch kind {
|
||||
case otel.SpanKindInternal:
|
||||
return tracepb.Span_INTERNAL
|
||||
return tracepb.Span_SPAN_KIND_INTERNAL
|
||||
case otel.SpanKindClient:
|
||||
return tracepb.Span_CLIENT
|
||||
return tracepb.Span_SPAN_KIND_CLIENT
|
||||
case otel.SpanKindServer:
|
||||
return tracepb.Span_SERVER
|
||||
return tracepb.Span_SPAN_KIND_SERVER
|
||||
case otel.SpanKindProducer:
|
||||
return tracepb.Span_PRODUCER
|
||||
return tracepb.Span_SPAN_KIND_PRODUCER
|
||||
case otel.SpanKindConsumer:
|
||||
return tracepb.Span_CONSUMER
|
||||
return tracepb.Span_SPAN_KIND_CONSUMER
|
||||
default:
|
||||
return tracepb.Span_SPAN_KIND_UNSPECIFIED
|
||||
}
|
||||
|
@ -41,23 +41,23 @@ func TestSpanKind(t *testing.T) {
|
||||
}{
|
||||
{
|
||||
otel.SpanKindInternal,
|
||||
tracepb.Span_INTERNAL,
|
||||
tracepb.Span_SPAN_KIND_INTERNAL,
|
||||
},
|
||||
{
|
||||
otel.SpanKindClient,
|
||||
tracepb.Span_CLIENT,
|
||||
tracepb.Span_SPAN_KIND_CLIENT,
|
||||
},
|
||||
{
|
||||
otel.SpanKindServer,
|
||||
tracepb.Span_SERVER,
|
||||
tracepb.Span_SPAN_KIND_SERVER,
|
||||
},
|
||||
{
|
||||
otel.SpanKindProducer,
|
||||
tracepb.Span_PRODUCER,
|
||||
tracepb.Span_SPAN_KIND_PRODUCER,
|
||||
},
|
||||
{
|
||||
otel.SpanKindConsumer,
|
||||
tracepb.Span_CONSUMER,
|
||||
tracepb.Span_SPAN_KIND_CONSUMER,
|
||||
},
|
||||
{
|
||||
otel.SpanKind(-1),
|
||||
@ -162,17 +162,17 @@ func TestStatus(t *testing.T) {
|
||||
{
|
||||
codes.Ok,
|
||||
"test Ok",
|
||||
tracepb.Status_Ok,
|
||||
tracepb.Status_STATUS_CODE_OK,
|
||||
},
|
||||
{
|
||||
codes.Unset,
|
||||
"test Unset",
|
||||
tracepb.Status_Ok,
|
||||
tracepb.Status_STATUS_CODE_OK,
|
||||
},
|
||||
{
|
||||
codes.Error,
|
||||
"test Error",
|
||||
tracepb.Status_UnknownError,
|
||||
tracepb.Status_STATUS_CODE_UNKNOWN_ERROR,
|
||||
},
|
||||
} {
|
||||
expected := &tracepb.Status{Code: test.otlpStatus, Message: test.message}
|
||||
@ -267,7 +267,7 @@ func TestSpanData(t *testing.T) {
|
||||
SpanId: []byte{0xFF, 0xFE, 0xFD, 0xFC, 0xFB, 0xFA, 0xF9, 0xF8},
|
||||
ParentSpanId: []byte{0xEF, 0xEE, 0xED, 0xEC, 0xEB, 0xEA, 0xE9, 0xE8},
|
||||
Name: spanData.Name,
|
||||
Kind: tracepb.Span_SERVER,
|
||||
Kind: tracepb.Span_SPAN_KIND_SERVER,
|
||||
StartTimeUnixNano: uint64(startTime.UnixNano()),
|
||||
EndTimeUnixNano: uint64(endTime.UnixNano()),
|
||||
Status: status(spanData.StatusCode, spanData.StatusMessage),
|
||||
|
@ -26,7 +26,6 @@ import (
|
||||
"github.com/stretchr/testify/require"
|
||||
|
||||
commonpb "go.opentelemetry.io/otel/exporters/otlp/internal/opentelemetry-proto-gen/common/v1"
|
||||
metricpb "go.opentelemetry.io/otel/exporters/otlp/internal/opentelemetry-proto-gen/metrics/v1"
|
||||
"go.opentelemetry.io/otel/label"
|
||||
|
||||
"go.opentelemetry.io/otel/api/metric"
|
||||
@ -238,37 +237,58 @@ func newExporterEndToEndTest(t *testing.T, additionalOpts []otlp.ExporterOption)
|
||||
assert.Len(t, metrics, len(instruments), "not enough metrics exported")
|
||||
seen := make(map[string]struct{}, len(instruments))
|
||||
for _, m := range metrics {
|
||||
desc := m.GetMetricDescriptor()
|
||||
data, ok := instruments[desc.Name]
|
||||
data, ok := instruments[m.Name]
|
||||
if !ok {
|
||||
assert.Failf(t, "unknown metrics", desc.Name)
|
||||
assert.Failf(t, "unknown metrics", m.Name)
|
||||
continue
|
||||
}
|
||||
seen[desc.Name] = struct{}{}
|
||||
seen[m.Name] = struct{}{}
|
||||
|
||||
switch data.iKind {
|
||||
case metric.CounterKind, metric.ValueObserverKind:
|
||||
case metric.CounterKind:
|
||||
switch data.nKind {
|
||||
case metricapi.Int64NumberKind:
|
||||
assert.Equal(t, metricpb.MetricDescriptor_INT64.String(), desc.GetType().String())
|
||||
if dp := m.GetInt64DataPoints(); assert.Len(t, dp, 1) {
|
||||
assert.Equal(t, data.val, dp[0].Value, "invalid value for %q", desc.Name)
|
||||
if dp := m.GetIntSum().DataPoints; assert.Len(t, dp, 1) {
|
||||
assert.Equal(t, data.val, dp[0].Value, "invalid value for %q", m.Name)
|
||||
}
|
||||
case metricapi.Float64NumberKind:
|
||||
assert.Equal(t, metricpb.MetricDescriptor_DOUBLE.String(), desc.GetType().String())
|
||||
if dp := m.GetDoubleDataPoints(); assert.Len(t, dp, 1) {
|
||||
assert.Equal(t, float64(data.val), dp[0].Value, "invalid value for %q", desc.Name)
|
||||
if dp := m.GetDoubleSum().DataPoints; assert.Len(t, dp, 1) {
|
||||
assert.Equal(t, float64(data.val), dp[0].Value, "invalid value for %q", m.Name)
|
||||
}
|
||||
default:
|
||||
assert.Failf(t, "invalid number kind", data.nKind.String())
|
||||
}
|
||||
case metric.ValueObserverKind:
|
||||
switch data.nKind {
|
||||
case metricapi.Int64NumberKind:
|
||||
if dp := m.GetIntGauge().DataPoints; assert.Len(t, dp, 1) {
|
||||
assert.Equal(t, data.val, dp[0].Value, "invalid value for %q", m.Name)
|
||||
}
|
||||
case metricapi.Float64NumberKind:
|
||||
if dp := m.GetDoubleGauge().DataPoints; assert.Len(t, dp, 1) {
|
||||
assert.Equal(t, float64(data.val), dp[0].Value, "invalid value for %q", m.Name)
|
||||
}
|
||||
default:
|
||||
assert.Failf(t, "invalid number kind", data.nKind.String())
|
||||
}
|
||||
case metric.ValueRecorderKind:
|
||||
assert.Equal(t, metricpb.MetricDescriptor_SUMMARY.String(), desc.GetType().String())
|
||||
m.GetSummaryDataPoints()
|
||||
if dp := m.GetSummaryDataPoints(); assert.Len(t, dp, 1) {
|
||||
count := dp[0].Count
|
||||
assert.Equal(t, uint64(1), count, "invalid count for %q", desc.Name)
|
||||
assert.Equal(t, float64(data.val*int64(count)), dp[0].Sum, "invalid sum for %q (value %d)", desc.Name, data.val)
|
||||
switch data.nKind {
|
||||
case metricapi.Int64NumberKind:
|
||||
assert.NotNil(t, m.GetIntHistogram())
|
||||
if dp := m.GetIntHistogram().DataPoints; assert.Len(t, dp, 1) {
|
||||
count := dp[0].Count
|
||||
assert.Equal(t, uint64(1), count, "invalid count for %q", m.Name)
|
||||
assert.Equal(t, int64(data.val*int64(count)), dp[0].Sum, "invalid sum for %q (value %d)", m.Name, data.val)
|
||||
}
|
||||
case metricapi.Float64NumberKind:
|
||||
assert.NotNil(t, m.GetDoubleHistogram())
|
||||
if dp := m.GetDoubleHistogram().DataPoints; assert.Len(t, dp, 1) {
|
||||
count := dp[0].Count
|
||||
assert.Equal(t, uint64(1), count, "invalid count for %q", m.Name)
|
||||
assert.Equal(t, float64(data.val*int64(count)), dp[0].Sum, "invalid sum for %q (value %d)", m.Name, data.val)
|
||||
}
|
||||
default:
|
||||
assert.Failf(t, "invalid number kind", data.nKind.String())
|
||||
}
|
||||
default:
|
||||
assert.Failf(t, "invalid metrics kind", data.iKind.String())
|
||||
|
@ -109,11 +109,6 @@ var (
|
||||
|
||||
testHistogramBoundaries = []float64{2.0, 4.0, 8.0}
|
||||
|
||||
md = &metricpb.MetricDescriptor{
|
||||
Name: "int64-count",
|
||||
Type: metricpb.MetricDescriptor_INT64,
|
||||
}
|
||||
|
||||
cpu1Labels = []*commonpb.StringKeyValue{
|
||||
{
|
||||
Key: "CPU",
|
||||
@ -189,19 +184,25 @@ func TestNoGroupingExport(t *testing.T) {
|
||||
{
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: md,
|
||||
Int64DataPoints: []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu2Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Name: "int64-count",
|
||||
Data: &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu2Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -229,50 +230,48 @@ func TestValuerecorderMetricGroupingExport(t *testing.T) {
|
||||
{
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: &metricpb.MetricDescriptor{
|
||||
Name: "valuerecorder",
|
||||
Type: metricpb.MetricDescriptor_HISTOGRAM,
|
||||
},
|
||||
HistogramDataPoints: []*metricpb.HistogramDataPoint{
|
||||
{
|
||||
Labels: []*commonpb.StringKeyValue{
|
||||
Name: "valuerecorder",
|
||||
Data: &metricpb.Metric_IntHistogram{
|
||||
IntHistogram: &metricpb.IntHistogram{
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntHistogramDataPoint{
|
||||
{
|
||||
Key: "CPU",
|
||||
Value: "1",
|
||||
Labels: []*commonpb.StringKeyValue{
|
||||
{
|
||||
Key: "CPU",
|
||||
Value: "1",
|
||||
},
|
||||
{
|
||||
Key: "host",
|
||||
Value: "test.com",
|
||||
},
|
||||
},
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Count: 2,
|
||||
Sum: 11,
|
||||
ExplicitBounds: testHistogramBoundaries,
|
||||
BucketCounts: []uint64{1, 0, 0, 1},
|
||||
},
|
||||
{
|
||||
Key: "host",
|
||||
Value: "test.com",
|
||||
Labels: []*commonpb.StringKeyValue{
|
||||
{
|
||||
Key: "CPU",
|
||||
Value: "1",
|
||||
},
|
||||
{
|
||||
Key: "host",
|
||||
Value: "test.com",
|
||||
},
|
||||
},
|
||||
Count: 2,
|
||||
Sum: 11,
|
||||
ExplicitBounds: testHistogramBoundaries,
|
||||
BucketCounts: []uint64{1, 0, 0, 1},
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Count: 2,
|
||||
Sum: 11,
|
||||
ExplicitBounds: testHistogramBoundaries,
|
||||
Buckets: []*metricpb.HistogramDataPoint_Bucket{
|
||||
{Count: 1}, {Count: 0}, {Count: 0}, {Count: 1},
|
||||
},
|
||||
},
|
||||
{
|
||||
Labels: []*commonpb.StringKeyValue{
|
||||
{
|
||||
Key: "CPU",
|
||||
Value: "1",
|
||||
},
|
||||
{
|
||||
Key: "host",
|
||||
Value: "test.com",
|
||||
},
|
||||
},
|
||||
Count: 2,
|
||||
Sum: 11,
|
||||
ExplicitBounds: testHistogramBoundaries,
|
||||
Buckets: []*metricpb.HistogramDataPoint_Bucket{
|
||||
{Count: 1}, {Count: 0}, {Count: 0}, {Count: 1},
|
||||
},
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -282,9 +281,6 @@ func TestValuerecorderMetricGroupingExport(t *testing.T) {
|
||||
},
|
||||
}
|
||||
runMetricExportTests(t, []record{r, r}, expected)
|
||||
//changing the number kind should make no difference.
|
||||
r.nKind = metric.Float64NumberKind
|
||||
runMetricExportTests(t, []record{r, r}, expected)
|
||||
}
|
||||
|
||||
func TestCountInt64MetricGroupingExport(t *testing.T) {
|
||||
@ -306,19 +302,25 @@ func TestCountInt64MetricGroupingExport(t *testing.T) {
|
||||
{
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: md,
|
||||
Int64DataPoints: []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Name: "int64-count",
|
||||
Data: &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -349,40 +351,43 @@ func TestCountFloat64MetricGroupingExport(t *testing.T) {
|
||||
{
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: &metricpb.MetricDescriptor{
|
||||
Name: "float64-count",
|
||||
Type: metricpb.MetricDescriptor_DOUBLE,
|
||||
},
|
||||
DoubleDataPoints: []*metricpb.DoubleDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: []*commonpb.StringKeyValue{
|
||||
Name: "float64-count",
|
||||
Data: &metricpb.Metric_DoubleSum{
|
||||
DoubleSum: &metricpb.DoubleSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.DoubleDataPoint{
|
||||
{
|
||||
Key: "CPU",
|
||||
Value: "1",
|
||||
Value: 11,
|
||||
Labels: []*commonpb.StringKeyValue{
|
||||
{
|
||||
Key: "CPU",
|
||||
Value: "1",
|
||||
},
|
||||
{
|
||||
Key: "host",
|
||||
Value: "test.com",
|
||||
},
|
||||
},
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Key: "host",
|
||||
Value: "test.com",
|
||||
Value: 11,
|
||||
Labels: []*commonpb.StringKeyValue{
|
||||
{
|
||||
Key: "CPU",
|
||||
Value: "1",
|
||||
},
|
||||
{
|
||||
Key: "host",
|
||||
Value: "test.com",
|
||||
},
|
||||
},
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: []*commonpb.StringKeyValue{
|
||||
{
|
||||
Key: "CPU",
|
||||
Value: "1",
|
||||
},
|
||||
{
|
||||
Key: "host",
|
||||
Value: "test.com",
|
||||
},
|
||||
},
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -438,25 +443,31 @@ func TestResourceMetricGroupingExport(t *testing.T) {
|
||||
{
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: md,
|
||||
Int64DataPoints: []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu2Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Name: "int64-count",
|
||||
Data: &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu2Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -470,13 +481,19 @@ func TestResourceMetricGroupingExport(t *testing.T) {
|
||||
{
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: md,
|
||||
Int64DataPoints: []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Name: "int64-count",
|
||||
Data: &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -563,25 +580,31 @@ func TestResourceInstLibMetricGroupingExport(t *testing.T) {
|
||||
},
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: md,
|
||||
Int64DataPoints: []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu2Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Name: "int64-count",
|
||||
Data: &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu2Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -594,13 +617,19 @@ func TestResourceInstLibMetricGroupingExport(t *testing.T) {
|
||||
},
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: md,
|
||||
Int64DataPoints: []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Name: "int64-count",
|
||||
Data: &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -612,13 +641,19 @@ func TestResourceInstLibMetricGroupingExport(t *testing.T) {
|
||||
},
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: md,
|
||||
Int64DataPoints: []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Name: "int64-count",
|
||||
Data: &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -636,13 +671,19 @@ func TestResourceInstLibMetricGroupingExport(t *testing.T) {
|
||||
},
|
||||
Metrics: []*metricpb.Metric{
|
||||
{
|
||||
MetricDescriptor: md,
|
||||
Int64DataPoints: []*metricpb.Int64DataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
Name: "int64-count",
|
||||
Data: &metricpb.Metric_IntSum{
|
||||
IntSum: &metricpb.IntSum{
|
||||
IsMonotonic: true,
|
||||
AggregationTemporality: metricpb.AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE,
|
||||
DataPoints: []*metricpb.IntDataPoint{
|
||||
{
|
||||
Value: 11,
|
||||
Labels: cpu1Labels,
|
||||
StartTimeUnixNano: startTime(),
|
||||
TimeUnixNano: pointTime(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -739,14 +780,25 @@ func runMetricExportTest(t *testing.T, exp *Exporter, rs []record, expected []me
|
||||
continue
|
||||
}
|
||||
for i, expected := range ilm.GetMetrics() {
|
||||
assert.Equal(t, expected.GetMetricDescriptor(), g[i].GetMetricDescriptor())
|
||||
// Compare each list directly because there is no order
|
||||
// guarantee with the concurrent processing design of the exporter
|
||||
// and ElementsMatch does not apply to contained slices.
|
||||
assert.ElementsMatch(t, expected.GetInt64DataPoints(), g[i].GetInt64DataPoints())
|
||||
assert.ElementsMatch(t, expected.GetDoubleDataPoints(), g[i].GetDoubleDataPoints())
|
||||
assert.ElementsMatch(t, expected.GetHistogramDataPoints(), g[i].GetHistogramDataPoints())
|
||||
assert.ElementsMatch(t, expected.GetSummaryDataPoints(), g[i].GetSummaryDataPoints())
|
||||
assert.Equal(t, expected.Name, g[i].Name)
|
||||
assert.Equal(t, expected.Unit, g[i].Unit)
|
||||
assert.Equal(t, expected.Description, g[i].Description)
|
||||
switch g[i].Data.(type) {
|
||||
case *metricpb.Metric_IntGauge:
|
||||
assert.ElementsMatch(t, expected.GetIntGauge().DataPoints, g[i].GetIntGauge().DataPoints)
|
||||
case *metricpb.Metric_IntHistogram:
|
||||
assert.ElementsMatch(t, expected.GetIntHistogram().DataPoints, g[i].GetIntHistogram().DataPoints)
|
||||
case *metricpb.Metric_IntSum:
|
||||
assert.ElementsMatch(t, expected.GetIntSum().DataPoints, g[i].GetIntSum().DataPoints)
|
||||
case *metricpb.Metric_DoubleGauge:
|
||||
assert.ElementsMatch(t, expected.GetDoubleGauge().DataPoints, g[i].GetDoubleGauge().DataPoints)
|
||||
case *metricpb.Metric_DoubleHistogram:
|
||||
assert.ElementsMatch(t, expected.GetDoubleHistogram().DataPoints, g[i].GetDoubleHistogram().DataPoints)
|
||||
case *metricpb.Metric_DoubleSum:
|
||||
assert.ElementsMatch(t, expected.GetDoubleSum().DataPoints, g[i].GetDoubleSum().DataPoints)
|
||||
default:
|
||||
assert.Failf(t, "unknown data type", g[i].Name)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -196,7 +196,7 @@ func TestExportSpans(t *testing.T) {
|
||||
TraceId: []byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1},
|
||||
SpanId: []byte{0, 0, 0, 0, 0, 0, 0, 1},
|
||||
Name: "parent process",
|
||||
Kind: tracepb.Span_SERVER,
|
||||
Kind: tracepb.Span_SPAN_KIND_SERVER,
|
||||
StartTimeUnixNano: uint64(startTime.UnixNano()),
|
||||
EndTimeUnixNano: uint64(endTime.UnixNano()),
|
||||
Attributes: []*commonpb.KeyValue{
|
||||
@ -218,7 +218,7 @@ func TestExportSpans(t *testing.T) {
|
||||
},
|
||||
},
|
||||
Status: &tracepb.Status{
|
||||
Code: tracepb.Status_Ok,
|
||||
Code: tracepb.Status_STATUS_CODE_OK,
|
||||
Message: "Ok",
|
||||
},
|
||||
},
|
||||
@ -227,7 +227,7 @@ func TestExportSpans(t *testing.T) {
|
||||
SpanId: []byte{0, 0, 0, 0, 0, 0, 0, 2},
|
||||
ParentSpanId: []byte{0, 0, 0, 0, 0, 0, 0, 1},
|
||||
Name: "internal process",
|
||||
Kind: tracepb.Span_INTERNAL,
|
||||
Kind: tracepb.Span_SPAN_KIND_INTERNAL,
|
||||
StartTimeUnixNano: uint64(startTime.UnixNano()),
|
||||
EndTimeUnixNano: uint64(endTime.UnixNano()),
|
||||
Attributes: []*commonpb.KeyValue{
|
||||
@ -249,7 +249,7 @@ func TestExportSpans(t *testing.T) {
|
||||
},
|
||||
},
|
||||
Status: &tracepb.Status{
|
||||
Code: tracepb.Status_Ok,
|
||||
Code: tracepb.Status_STATUS_CODE_OK,
|
||||
Message: "Ok",
|
||||
},
|
||||
},
|
||||
@ -265,7 +265,7 @@ func TestExportSpans(t *testing.T) {
|
||||
TraceId: []byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2},
|
||||
SpanId: []byte{0, 0, 0, 0, 0, 0, 0, 1},
|
||||
Name: "secondary parent process",
|
||||
Kind: tracepb.Span_SERVER,
|
||||
Kind: tracepb.Span_SPAN_KIND_SERVER,
|
||||
StartTimeUnixNano: uint64(startTime.UnixNano()),
|
||||
EndTimeUnixNano: uint64(endTime.UnixNano()),
|
||||
Attributes: []*commonpb.KeyValue{
|
||||
@ -287,7 +287,7 @@ func TestExportSpans(t *testing.T) {
|
||||
},
|
||||
},
|
||||
Status: &tracepb.Status{
|
||||
Code: tracepb.Status_Ok,
|
||||
Code: tracepb.Status_STATUS_CODE_OK,
|
||||
Message: "Ok",
|
||||
},
|
||||
},
|
||||
@ -319,7 +319,7 @@ func TestExportSpans(t *testing.T) {
|
||||
TraceId: []byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2},
|
||||
SpanId: []byte{0, 0, 0, 0, 0, 0, 0, 1},
|
||||
Name: "parent process",
|
||||
Kind: tracepb.Span_SERVER,
|
||||
Kind: tracepb.Span_SPAN_KIND_SERVER,
|
||||
StartTimeUnixNano: uint64(startTime.UnixNano()),
|
||||
EndTimeUnixNano: uint64(endTime.UnixNano()),
|
||||
Attributes: []*commonpb.KeyValue{
|
||||
@ -341,7 +341,7 @@ func TestExportSpans(t *testing.T) {
|
||||
},
|
||||
},
|
||||
Status: &tracepb.Status{
|
||||
Code: tracepb.Status_UnknownError,
|
||||
Code: tracepb.Status_STATUS_CODE_UNKNOWN_ERROR,
|
||||
Message: "Unauthenticated",
|
||||
},
|
||||
},
|
||||
|
Loading…
x
Reference in New Issue
Block a user