repo
stringlengths 5
54
| path
stringlengths 4
155
| func_name
stringlengths 1
118
| original_string
stringlengths 52
85.5k
| language
stringclasses 1
value | code
stringlengths 52
85.5k
| code_tokens
sequence | docstring
stringlengths 6
2.61k
| docstring_tokens
sequence | sha
stringlengths 40
40
| url
stringlengths 85
252
| partition
stringclasses 1
value |
---|---|---|---|---|---|---|---|---|---|---|---|
anacrolix/torrent | torrent.go | bytesCompleted | func (t *Torrent) bytesCompleted() int64 {
if !t.haveInfo() {
return 0
}
return t.info.TotalLength() - t.bytesLeft()
} | go | func (t *Torrent) bytesCompleted() int64 {
if !t.haveInfo() {
return 0
}
return t.info.TotalLength() - t.bytesLeft()
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"bytesCompleted",
"(",
")",
"int64",
"{",
"if",
"!",
"t",
".",
"haveInfo",
"(",
")",
"{",
"return",
"0",
"\n",
"}",
"\n",
"return",
"t",
".",
"info",
".",
"TotalLength",
"(",
")",
"-",
"t",
".",
"bytesLeft",
"(",
")",
"\n",
"}"
] | // Don't call this before the info is available. | [
"Don",
"t",
"call",
"this",
"before",
"the",
"info",
"is",
"available",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1162-L1167 | train |
anacrolix/torrent | torrent.go | deleteConnection | func (t *Torrent) deleteConnection(c *connection) (ret bool) {
if !c.closed.IsSet() {
panic("connection is not closed")
// There are behaviours prevented by the closed state that will fail
// if the connection has been deleted.
}
_, ret = t.conns[c]
delete(t.conns, c)
torrent.Add("deleted connections", 1)
c.deleteAllRequests()
if len(t.conns) == 0 {
t.assertNoPendingRequests()
}
return
} | go | func (t *Torrent) deleteConnection(c *connection) (ret bool) {
if !c.closed.IsSet() {
panic("connection is not closed")
// There are behaviours prevented by the closed state that will fail
// if the connection has been deleted.
}
_, ret = t.conns[c]
delete(t.conns, c)
torrent.Add("deleted connections", 1)
c.deleteAllRequests()
if len(t.conns) == 0 {
t.assertNoPendingRequests()
}
return
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"deleteConnection",
"(",
"c",
"*",
"connection",
")",
"(",
"ret",
"bool",
")",
"{",
"if",
"!",
"c",
".",
"closed",
".",
"IsSet",
"(",
")",
"{",
"panic",
"(",
"\"",
"\"",
")",
"\n",
"// There are behaviours prevented by the closed state that will fail",
"// if the connection has been deleted.",
"}",
"\n",
"_",
",",
"ret",
"=",
"t",
".",
"conns",
"[",
"c",
"]",
"\n",
"delete",
"(",
"t",
".",
"conns",
",",
"c",
")",
"\n",
"torrent",
".",
"Add",
"(",
"\"",
"\"",
",",
"1",
")",
"\n",
"c",
".",
"deleteAllRequests",
"(",
")",
"\n",
"if",
"len",
"(",
"t",
".",
"conns",
")",
"==",
"0",
"{",
"t",
".",
"assertNoPendingRequests",
"(",
")",
"\n",
"}",
"\n",
"return",
"\n",
"}"
] | // Returns true if connection is removed from torrent.Conns. | [
"Returns",
"true",
"if",
"connection",
"is",
"removed",
"from",
"torrent",
".",
"Conns",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1176-L1190 | train |
anacrolix/torrent | torrent.go | seeding | func (t *Torrent) seeding() bool {
cl := t.cl
if t.closed.IsSet() {
return false
}
if cl.config.NoUpload {
return false
}
if !cl.config.Seed {
return false
}
if cl.config.DisableAggressiveUpload && t.needData() {
return false
}
return true
} | go | func (t *Torrent) seeding() bool {
cl := t.cl
if t.closed.IsSet() {
return false
}
if cl.config.NoUpload {
return false
}
if !cl.config.Seed {
return false
}
if cl.config.DisableAggressiveUpload && t.needData() {
return false
}
return true
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"seeding",
"(",
")",
"bool",
"{",
"cl",
":=",
"t",
".",
"cl",
"\n",
"if",
"t",
".",
"closed",
".",
"IsSet",
"(",
")",
"{",
"return",
"false",
"\n",
"}",
"\n",
"if",
"cl",
".",
"config",
".",
"NoUpload",
"{",
"return",
"false",
"\n",
"}",
"\n",
"if",
"!",
"cl",
".",
"config",
".",
"Seed",
"{",
"return",
"false",
"\n",
"}",
"\n",
"if",
"cl",
".",
"config",
".",
"DisableAggressiveUpload",
"&&",
"t",
".",
"needData",
"(",
")",
"{",
"return",
"false",
"\n",
"}",
"\n",
"return",
"true",
"\n",
"}"
] | // Returns whether the client should make effort to seed the torrent. | [
"Returns",
"whether",
"the",
"client",
"should",
"make",
"effort",
"to",
"seed",
"the",
"torrent",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1228-L1243 | train |
anacrolix/torrent | torrent.go | startMissingTrackerScrapers | func (t *Torrent) startMissingTrackerScrapers() {
if t.cl.config.DisableTrackers {
return
}
t.startScrapingTracker(t.metainfo.Announce)
for _, tier := range t.metainfo.AnnounceList {
for _, url := range tier {
t.startScrapingTracker(url)
}
}
} | go | func (t *Torrent) startMissingTrackerScrapers() {
if t.cl.config.DisableTrackers {
return
}
t.startScrapingTracker(t.metainfo.Announce)
for _, tier := range t.metainfo.AnnounceList {
for _, url := range tier {
t.startScrapingTracker(url)
}
}
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"startMissingTrackerScrapers",
"(",
")",
"{",
"if",
"t",
".",
"cl",
".",
"config",
".",
"DisableTrackers",
"{",
"return",
"\n",
"}",
"\n",
"t",
".",
"startScrapingTracker",
"(",
"t",
".",
"metainfo",
".",
"Announce",
")",
"\n",
"for",
"_",
",",
"tier",
":=",
"range",
"t",
".",
"metainfo",
".",
"AnnounceList",
"{",
"for",
"_",
",",
"url",
":=",
"range",
"tier",
"{",
"t",
".",
"startScrapingTracker",
"(",
"url",
")",
"\n",
"}",
"\n",
"}",
"\n",
"}"
] | // Adds and starts tracker scrapers for tracker URLs that aren't already
// running. | [
"Adds",
"and",
"starts",
"tracker",
"scrapers",
"for",
"tracker",
"URLs",
"that",
"aren",
"t",
"already",
"running",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1287-L1297 | train |
anacrolix/torrent | torrent.go | announceRequest | func (t *Torrent) announceRequest() tracker.AnnounceRequest {
// Note that IPAddress is not set. It's set for UDP inside the tracker
// code, since it's dependent on the network in use.
return tracker.AnnounceRequest{
Event: tracker.None,
NumWant: -1,
Port: uint16(t.cl.incomingPeerPort()),
PeerId: t.cl.peerID,
InfoHash: t.infoHash,
Key: t.cl.announceKey(),
// The following are vaguely described in BEP 3.
Left: t.bytesLeftAnnounce(),
Uploaded: t.stats.BytesWrittenData.Int64(),
// There's no mention of wasted or unwanted download in the BEP.
Downloaded: t.stats.BytesReadUsefulData.Int64(),
}
} | go | func (t *Torrent) announceRequest() tracker.AnnounceRequest {
// Note that IPAddress is not set. It's set for UDP inside the tracker
// code, since it's dependent on the network in use.
return tracker.AnnounceRequest{
Event: tracker.None,
NumWant: -1,
Port: uint16(t.cl.incomingPeerPort()),
PeerId: t.cl.peerID,
InfoHash: t.infoHash,
Key: t.cl.announceKey(),
// The following are vaguely described in BEP 3.
Left: t.bytesLeftAnnounce(),
Uploaded: t.stats.BytesWrittenData.Int64(),
// There's no mention of wasted or unwanted download in the BEP.
Downloaded: t.stats.BytesReadUsefulData.Int64(),
}
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"announceRequest",
"(",
")",
"tracker",
".",
"AnnounceRequest",
"{",
"// Note that IPAddress is not set. It's set for UDP inside the tracker",
"// code, since it's dependent on the network in use.",
"return",
"tracker",
".",
"AnnounceRequest",
"{",
"Event",
":",
"tracker",
".",
"None",
",",
"NumWant",
":",
"-",
"1",
",",
"Port",
":",
"uint16",
"(",
"t",
".",
"cl",
".",
"incomingPeerPort",
"(",
")",
")",
",",
"PeerId",
":",
"t",
".",
"cl",
".",
"peerID",
",",
"InfoHash",
":",
"t",
".",
"infoHash",
",",
"Key",
":",
"t",
".",
"cl",
".",
"announceKey",
"(",
")",
",",
"// The following are vaguely described in BEP 3.",
"Left",
":",
"t",
".",
"bytesLeftAnnounce",
"(",
")",
",",
"Uploaded",
":",
"t",
".",
"stats",
".",
"BytesWrittenData",
".",
"Int64",
"(",
")",
",",
"// There's no mention of wasted or unwanted download in the BEP.",
"Downloaded",
":",
"t",
".",
"stats",
".",
"BytesReadUsefulData",
".",
"Int64",
"(",
")",
",",
"}",
"\n",
"}"
] | // Returns an AnnounceRequest with fields filled out to defaults and current
// values. | [
"Returns",
"an",
"AnnounceRequest",
"with",
"fields",
"filled",
"out",
"to",
"defaults",
"and",
"current",
"values",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1301-L1319 | train |
anacrolix/torrent | torrent.go | consumeDhtAnnouncePeers | func (t *Torrent) consumeDhtAnnouncePeers(pvs <-chan dht.PeersValues) {
cl := t.cl
for v := range pvs {
cl.lock()
for _, cp := range v.Peers {
if cp.Port == 0 {
// Can't do anything with this.
continue
}
t.addPeer(Peer{
IP: cp.IP[:],
Port: cp.Port,
Source: peerSourceDHTGetPeers,
})
}
cl.unlock()
}
} | go | func (t *Torrent) consumeDhtAnnouncePeers(pvs <-chan dht.PeersValues) {
cl := t.cl
for v := range pvs {
cl.lock()
for _, cp := range v.Peers {
if cp.Port == 0 {
// Can't do anything with this.
continue
}
t.addPeer(Peer{
IP: cp.IP[:],
Port: cp.Port,
Source: peerSourceDHTGetPeers,
})
}
cl.unlock()
}
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"consumeDhtAnnouncePeers",
"(",
"pvs",
"<-",
"chan",
"dht",
".",
"PeersValues",
")",
"{",
"cl",
":=",
"t",
".",
"cl",
"\n",
"for",
"v",
":=",
"range",
"pvs",
"{",
"cl",
".",
"lock",
"(",
")",
"\n",
"for",
"_",
",",
"cp",
":=",
"range",
"v",
".",
"Peers",
"{",
"if",
"cp",
".",
"Port",
"==",
"0",
"{",
"// Can't do anything with this.",
"continue",
"\n",
"}",
"\n",
"t",
".",
"addPeer",
"(",
"Peer",
"{",
"IP",
":",
"cp",
".",
"IP",
"[",
":",
"]",
",",
"Port",
":",
"cp",
".",
"Port",
",",
"Source",
":",
"peerSourceDHTGetPeers",
",",
"}",
")",
"\n",
"}",
"\n",
"cl",
".",
"unlock",
"(",
")",
"\n",
"}",
"\n",
"}"
] | // Adds peers revealed in an announce until the announce ends, or we have
// enough peers. | [
"Adds",
"peers",
"revealed",
"in",
"an",
"announce",
"until",
"the",
"announce",
"ends",
"or",
"we",
"have",
"enough",
"peers",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1323-L1340 | train |
anacrolix/torrent | torrent.go | numTotalPeers | func (t *Torrent) numTotalPeers() int {
peers := make(map[string]struct{})
for conn := range t.conns {
ra := conn.conn.RemoteAddr()
if ra == nil {
// It's been closed and doesn't support RemoteAddr.
continue
}
peers[ra.String()] = struct{}{}
}
for addr := range t.halfOpen {
peers[addr] = struct{}{}
}
t.peers.Each(func(peer Peer) {
peers[fmt.Sprintf("%s:%d", peer.IP, peer.Port)] = struct{}{}
})
return len(peers)
} | go | func (t *Torrent) numTotalPeers() int {
peers := make(map[string]struct{})
for conn := range t.conns {
ra := conn.conn.RemoteAddr()
if ra == nil {
// It's been closed and doesn't support RemoteAddr.
continue
}
peers[ra.String()] = struct{}{}
}
for addr := range t.halfOpen {
peers[addr] = struct{}{}
}
t.peers.Each(func(peer Peer) {
peers[fmt.Sprintf("%s:%d", peer.IP, peer.Port)] = struct{}{}
})
return len(peers)
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"numTotalPeers",
"(",
")",
"int",
"{",
"peers",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"struct",
"{",
"}",
")",
"\n",
"for",
"conn",
":=",
"range",
"t",
".",
"conns",
"{",
"ra",
":=",
"conn",
".",
"conn",
".",
"RemoteAddr",
"(",
")",
"\n",
"if",
"ra",
"==",
"nil",
"{",
"// It's been closed and doesn't support RemoteAddr.",
"continue",
"\n",
"}",
"\n",
"peers",
"[",
"ra",
".",
"String",
"(",
")",
"]",
"=",
"struct",
"{",
"}",
"{",
"}",
"\n",
"}",
"\n",
"for",
"addr",
":=",
"range",
"t",
".",
"halfOpen",
"{",
"peers",
"[",
"addr",
"]",
"=",
"struct",
"{",
"}",
"{",
"}",
"\n",
"}",
"\n",
"t",
".",
"peers",
".",
"Each",
"(",
"func",
"(",
"peer",
"Peer",
")",
"{",
"peers",
"[",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"peer",
".",
"IP",
",",
"peer",
".",
"Port",
")",
"]",
"=",
"struct",
"{",
"}",
"{",
"}",
"\n",
"}",
")",
"\n",
"return",
"len",
"(",
"peers",
")",
"\n",
"}"
] | // The total number of peers in the torrent. | [
"The",
"total",
"number",
"of",
"peers",
"in",
"the",
"torrent",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1402-L1419 | train |
anacrolix/torrent | torrent.go | reconcileHandshakeStats | func (t *Torrent) reconcileHandshakeStats(c *connection) {
if c.stats != (ConnStats{
// Handshakes should only increment these fields:
BytesWritten: c.stats.BytesWritten,
BytesRead: c.stats.BytesRead,
}) {
panic("bad stats")
}
c.postHandshakeStats(func(cs *ConnStats) {
cs.BytesRead.Add(c.stats.BytesRead.Int64())
cs.BytesWritten.Add(c.stats.BytesWritten.Int64())
})
c.reconciledHandshakeStats = true
} | go | func (t *Torrent) reconcileHandshakeStats(c *connection) {
if c.stats != (ConnStats{
// Handshakes should only increment these fields:
BytesWritten: c.stats.BytesWritten,
BytesRead: c.stats.BytesRead,
}) {
panic("bad stats")
}
c.postHandshakeStats(func(cs *ConnStats) {
cs.BytesRead.Add(c.stats.BytesRead.Int64())
cs.BytesWritten.Add(c.stats.BytesWritten.Int64())
})
c.reconciledHandshakeStats = true
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"reconcileHandshakeStats",
"(",
"c",
"*",
"connection",
")",
"{",
"if",
"c",
".",
"stats",
"!=",
"(",
"ConnStats",
"{",
"// Handshakes should only increment these fields:",
"BytesWritten",
":",
"c",
".",
"stats",
".",
"BytesWritten",
",",
"BytesRead",
":",
"c",
".",
"stats",
".",
"BytesRead",
",",
"}",
")",
"{",
"panic",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"c",
".",
"postHandshakeStats",
"(",
"func",
"(",
"cs",
"*",
"ConnStats",
")",
"{",
"cs",
".",
"BytesRead",
".",
"Add",
"(",
"c",
".",
"stats",
".",
"BytesRead",
".",
"Int64",
"(",
")",
")",
"\n",
"cs",
".",
"BytesWritten",
".",
"Add",
"(",
"c",
".",
"stats",
".",
"BytesWritten",
".",
"Int64",
"(",
")",
")",
"\n",
"}",
")",
"\n",
"c",
".",
"reconciledHandshakeStats",
"=",
"true",
"\n",
"}"
] | // Reconcile bytes transferred before connection was associated with a
// torrent. | [
"Reconcile",
"bytes",
"transferred",
"before",
"connection",
"was",
"associated",
"with",
"a",
"torrent",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1423-L1436 | train |
anacrolix/torrent | torrent.go | addConnection | func (t *Torrent) addConnection(c *connection) (err error) {
defer func() {
if err == nil {
torrent.Add("added connections", 1)
}
}()
if t.closed.IsSet() {
return errors.New("torrent closed")
}
for c0 := range t.conns {
if c.PeerID != c0.PeerID {
continue
}
if !t.cl.config.dropDuplicatePeerIds {
continue
}
if left, ok := c.hasPreferredNetworkOver(c0); ok && left {
c0.Close()
t.deleteConnection(c0)
} else {
return errors.New("existing connection preferred")
}
}
if len(t.conns) >= t.maxEstablishedConns {
c := t.worstBadConn()
if c == nil {
return errors.New("don't want conns")
}
c.Close()
t.deleteConnection(c)
}
if len(t.conns) >= t.maxEstablishedConns {
panic(len(t.conns))
}
t.conns[c] = struct{}{}
return nil
} | go | func (t *Torrent) addConnection(c *connection) (err error) {
defer func() {
if err == nil {
torrent.Add("added connections", 1)
}
}()
if t.closed.IsSet() {
return errors.New("torrent closed")
}
for c0 := range t.conns {
if c.PeerID != c0.PeerID {
continue
}
if !t.cl.config.dropDuplicatePeerIds {
continue
}
if left, ok := c.hasPreferredNetworkOver(c0); ok && left {
c0.Close()
t.deleteConnection(c0)
} else {
return errors.New("existing connection preferred")
}
}
if len(t.conns) >= t.maxEstablishedConns {
c := t.worstBadConn()
if c == nil {
return errors.New("don't want conns")
}
c.Close()
t.deleteConnection(c)
}
if len(t.conns) >= t.maxEstablishedConns {
panic(len(t.conns))
}
t.conns[c] = struct{}{}
return nil
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"addConnection",
"(",
"c",
"*",
"connection",
")",
"(",
"err",
"error",
")",
"{",
"defer",
"func",
"(",
")",
"{",
"if",
"err",
"==",
"nil",
"{",
"torrent",
".",
"Add",
"(",
"\"",
"\"",
",",
"1",
")",
"\n",
"}",
"\n",
"}",
"(",
")",
"\n",
"if",
"t",
".",
"closed",
".",
"IsSet",
"(",
")",
"{",
"return",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"for",
"c0",
":=",
"range",
"t",
".",
"conns",
"{",
"if",
"c",
".",
"PeerID",
"!=",
"c0",
".",
"PeerID",
"{",
"continue",
"\n",
"}",
"\n",
"if",
"!",
"t",
".",
"cl",
".",
"config",
".",
"dropDuplicatePeerIds",
"{",
"continue",
"\n",
"}",
"\n",
"if",
"left",
",",
"ok",
":=",
"c",
".",
"hasPreferredNetworkOver",
"(",
"c0",
")",
";",
"ok",
"&&",
"left",
"{",
"c0",
".",
"Close",
"(",
")",
"\n",
"t",
".",
"deleteConnection",
"(",
"c0",
")",
"\n",
"}",
"else",
"{",
"return",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"len",
"(",
"t",
".",
"conns",
")",
">=",
"t",
".",
"maxEstablishedConns",
"{",
"c",
":=",
"t",
".",
"worstBadConn",
"(",
")",
"\n",
"if",
"c",
"==",
"nil",
"{",
"return",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"c",
".",
"Close",
"(",
")",
"\n",
"t",
".",
"deleteConnection",
"(",
"c",
")",
"\n",
"}",
"\n",
"if",
"len",
"(",
"t",
".",
"conns",
")",
">=",
"t",
".",
"maxEstablishedConns",
"{",
"panic",
"(",
"len",
"(",
"t",
".",
"conns",
")",
")",
"\n",
"}",
"\n",
"t",
".",
"conns",
"[",
"c",
"]",
"=",
"struct",
"{",
"}",
"{",
"}",
"\n",
"return",
"nil",
"\n",
"}"
] | // Returns true if the connection is added. | [
"Returns",
"true",
"if",
"the",
"connection",
"is",
"added",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1439-L1475 | train |
anacrolix/torrent | torrent.go | onIncompletePiece | func (t *Torrent) onIncompletePiece(piece pieceIndex) {
if t.pieceAllDirty(piece) {
t.pendAllChunkSpecs(piece)
}
if !t.wantPieceIndex(piece) {
// t.logger.Printf("piece %d incomplete and unwanted", piece)
return
}
// We could drop any connections that we told we have a piece that we
// don't here. But there's a test failure, and it seems clients don't care
// if you request pieces that you already claim to have. Pruning bad
// connections might just remove any connections that aren't treating us
// favourably anyway.
// for c := range t.conns {
// if c.sentHave(piece) {
// c.Drop()
// }
// }
for conn := range t.conns {
if conn.PeerHasPiece(piece) {
conn.updateRequests()
}
}
} | go | func (t *Torrent) onIncompletePiece(piece pieceIndex) {
if t.pieceAllDirty(piece) {
t.pendAllChunkSpecs(piece)
}
if !t.wantPieceIndex(piece) {
// t.logger.Printf("piece %d incomplete and unwanted", piece)
return
}
// We could drop any connections that we told we have a piece that we
// don't here. But there's a test failure, and it seems clients don't care
// if you request pieces that you already claim to have. Pruning bad
// connections might just remove any connections that aren't treating us
// favourably anyway.
// for c := range t.conns {
// if c.sentHave(piece) {
// c.Drop()
// }
// }
for conn := range t.conns {
if conn.PeerHasPiece(piece) {
conn.updateRequests()
}
}
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"onIncompletePiece",
"(",
"piece",
"pieceIndex",
")",
"{",
"if",
"t",
".",
"pieceAllDirty",
"(",
"piece",
")",
"{",
"t",
".",
"pendAllChunkSpecs",
"(",
"piece",
")",
"\n",
"}",
"\n",
"if",
"!",
"t",
".",
"wantPieceIndex",
"(",
"piece",
")",
"{",
"// t.logger.Printf(\"piece %d incomplete and unwanted\", piece)",
"return",
"\n",
"}",
"\n",
"// We could drop any connections that we told we have a piece that we",
"// don't here. But there's a test failure, and it seems clients don't care",
"// if you request pieces that you already claim to have. Pruning bad",
"// connections might just remove any connections that aren't treating us",
"// favourably anyway.",
"// for c := range t.conns {",
"// \tif c.sentHave(piece) {",
"// \t\tc.Drop()",
"// \t}",
"// }",
"for",
"conn",
":=",
"range",
"t",
".",
"conns",
"{",
"if",
"conn",
".",
"PeerHasPiece",
"(",
"piece",
")",
"{",
"conn",
".",
"updateRequests",
"(",
")",
"\n",
"}",
"\n",
"}",
"\n",
"}"
] | // Called when a piece is found to be not complete. | [
"Called",
"when",
"a",
"piece",
"is",
"found",
"to",
"be",
"not",
"complete",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1580-L1604 | train |
anacrolix/torrent | torrent.go | reapPieceTouchers | func (t *Torrent) reapPieceTouchers(piece pieceIndex) (ret []*connection) {
for c := range t.pieces[piece].dirtiers {
delete(c.peerTouchedPieces, piece)
ret = append(ret, c)
}
t.pieces[piece].dirtiers = nil
return
} | go | func (t *Torrent) reapPieceTouchers(piece pieceIndex) (ret []*connection) {
for c := range t.pieces[piece].dirtiers {
delete(c.peerTouchedPieces, piece)
ret = append(ret, c)
}
t.pieces[piece].dirtiers = nil
return
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"reapPieceTouchers",
"(",
"piece",
"pieceIndex",
")",
"(",
"ret",
"[",
"]",
"*",
"connection",
")",
"{",
"for",
"c",
":=",
"range",
"t",
".",
"pieces",
"[",
"piece",
"]",
".",
"dirtiers",
"{",
"delete",
"(",
"c",
".",
"peerTouchedPieces",
",",
"piece",
")",
"\n",
"ret",
"=",
"append",
"(",
"ret",
",",
"c",
")",
"\n",
"}",
"\n",
"t",
".",
"pieces",
"[",
"piece",
"]",
".",
"dirtiers",
"=",
"nil",
"\n",
"return",
"\n",
"}"
] | // Return the connections that touched a piece, and clear the entries while
// doing it. | [
"Return",
"the",
"connections",
"that",
"touched",
"a",
"piece",
"and",
"clear",
"the",
"entries",
"while",
"doing",
"it",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1641-L1648 | train |
anacrolix/torrent | torrent.go | queuePieceCheck | func (t *Torrent) queuePieceCheck(pieceIndex pieceIndex) {
piece := &t.pieces[pieceIndex]
if piece.queuedForHash() {
return
}
t.piecesQueuedForHash.Add(bitmap.BitIndex(pieceIndex))
t.publishPieceChange(pieceIndex)
t.updatePiecePriority(pieceIndex)
go t.verifyPiece(pieceIndex)
} | go | func (t *Torrent) queuePieceCheck(pieceIndex pieceIndex) {
piece := &t.pieces[pieceIndex]
if piece.queuedForHash() {
return
}
t.piecesQueuedForHash.Add(bitmap.BitIndex(pieceIndex))
t.publishPieceChange(pieceIndex)
t.updatePiecePriority(pieceIndex)
go t.verifyPiece(pieceIndex)
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"queuePieceCheck",
"(",
"pieceIndex",
"pieceIndex",
")",
"{",
"piece",
":=",
"&",
"t",
".",
"pieces",
"[",
"pieceIndex",
"]",
"\n",
"if",
"piece",
".",
"queuedForHash",
"(",
")",
"{",
"return",
"\n",
"}",
"\n",
"t",
".",
"piecesQueuedForHash",
".",
"Add",
"(",
"bitmap",
".",
"BitIndex",
"(",
"pieceIndex",
")",
")",
"\n",
"t",
".",
"publishPieceChange",
"(",
"pieceIndex",
")",
"\n",
"t",
".",
"updatePiecePriority",
"(",
"pieceIndex",
")",
"\n",
"go",
"t",
".",
"verifyPiece",
"(",
"pieceIndex",
")",
"\n",
"}"
] | // Currently doesn't really queue, but should in the future. | [
"Currently",
"doesn",
"t",
"really",
"queue",
"but",
"should",
"in",
"the",
"future",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1658-L1667 | train |
anacrolix/torrent | torrent.go | initiateConn | func (t *Torrent) initiateConn(peer Peer) {
if peer.Id == t.cl.peerID {
return
}
if t.cl.badPeerIPPort(peer.IP, peer.Port) {
return
}
addr := IpPort{peer.IP, uint16(peer.Port)}
if t.addrActive(addr.String()) {
return
}
t.halfOpen[addr.String()] = peer
go t.cl.outgoingConnection(t, addr, peer.Source)
} | go | func (t *Torrent) initiateConn(peer Peer) {
if peer.Id == t.cl.peerID {
return
}
if t.cl.badPeerIPPort(peer.IP, peer.Port) {
return
}
addr := IpPort{peer.IP, uint16(peer.Port)}
if t.addrActive(addr.String()) {
return
}
t.halfOpen[addr.String()] = peer
go t.cl.outgoingConnection(t, addr, peer.Source)
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"initiateConn",
"(",
"peer",
"Peer",
")",
"{",
"if",
"peer",
".",
"Id",
"==",
"t",
".",
"cl",
".",
"peerID",
"{",
"return",
"\n",
"}",
"\n",
"if",
"t",
".",
"cl",
".",
"badPeerIPPort",
"(",
"peer",
".",
"IP",
",",
"peer",
".",
"Port",
")",
"{",
"return",
"\n",
"}",
"\n",
"addr",
":=",
"IpPort",
"{",
"peer",
".",
"IP",
",",
"uint16",
"(",
"peer",
".",
"Port",
")",
"}",
"\n",
"if",
"t",
".",
"addrActive",
"(",
"addr",
".",
"String",
"(",
")",
")",
"{",
"return",
"\n",
"}",
"\n",
"t",
".",
"halfOpen",
"[",
"addr",
".",
"String",
"(",
")",
"]",
"=",
"peer",
"\n",
"go",
"t",
".",
"cl",
".",
"outgoingConnection",
"(",
"t",
",",
"addr",
",",
"peer",
".",
"Source",
")",
"\n",
"}"
] | // Start the process of connecting to the given peer for the given torrent if
// appropriate. | [
"Start",
"the",
"process",
"of",
"connecting",
"to",
"the",
"given",
"peer",
"for",
"the",
"given",
"torrent",
"if",
"appropriate",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1677-L1690 | train |
anacrolix/torrent | torrent.go | allStats | func (t *Torrent) allStats(f func(*ConnStats)) {
f(&t.stats)
f(&t.cl.stats)
} | go | func (t *Torrent) allStats(f func(*ConnStats)) {
f(&t.stats)
f(&t.cl.stats)
} | [
"func",
"(",
"t",
"*",
"Torrent",
")",
"allStats",
"(",
"f",
"func",
"(",
"*",
"ConnStats",
")",
")",
"{",
"f",
"(",
"&",
"t",
".",
"stats",
")",
"\n",
"f",
"(",
"&",
"t",
".",
"cl",
".",
"stats",
")",
"\n",
"}"
] | // All stats that include this Torrent. Useful when we want to increment
// ConnStats but not for every connection. | [
"All",
"stats",
"that",
"include",
"this",
"Torrent",
".",
"Useful",
"when",
"we",
"want",
"to",
"increment",
"ConnStats",
"but",
"not",
"for",
"every",
"connection",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1706-L1709 | train |
anacrolix/torrent | metainfo/info.go | BuildFromFilePath | func (info *Info) BuildFromFilePath(root string) (err error) {
info.Name = filepath.Base(root)
info.Files = nil
err = filepath.Walk(root, func(path string, fi os.FileInfo, err error) error {
if err != nil {
return err
}
if fi.IsDir() {
// Directories are implicit in torrent files.
return nil
} else if path == root {
// The root is a file.
info.Length = fi.Size()
return nil
}
relPath, err := filepath.Rel(root, path)
if err != nil {
return fmt.Errorf("error getting relative path: %s", err)
}
info.Files = append(info.Files, FileInfo{
Path: strings.Split(relPath, string(filepath.Separator)),
Length: fi.Size(),
})
return nil
})
if err != nil {
return
}
slices.Sort(info.Files, func(l, r FileInfo) bool {
return strings.Join(l.Path, "/") < strings.Join(r.Path, "/")
})
err = info.GeneratePieces(func(fi FileInfo) (io.ReadCloser, error) {
return os.Open(filepath.Join(root, strings.Join(fi.Path, string(filepath.Separator))))
})
if err != nil {
err = fmt.Errorf("error generating pieces: %s", err)
}
return
} | go | func (info *Info) BuildFromFilePath(root string) (err error) {
info.Name = filepath.Base(root)
info.Files = nil
err = filepath.Walk(root, func(path string, fi os.FileInfo, err error) error {
if err != nil {
return err
}
if fi.IsDir() {
// Directories are implicit in torrent files.
return nil
} else if path == root {
// The root is a file.
info.Length = fi.Size()
return nil
}
relPath, err := filepath.Rel(root, path)
if err != nil {
return fmt.Errorf("error getting relative path: %s", err)
}
info.Files = append(info.Files, FileInfo{
Path: strings.Split(relPath, string(filepath.Separator)),
Length: fi.Size(),
})
return nil
})
if err != nil {
return
}
slices.Sort(info.Files, func(l, r FileInfo) bool {
return strings.Join(l.Path, "/") < strings.Join(r.Path, "/")
})
err = info.GeneratePieces(func(fi FileInfo) (io.ReadCloser, error) {
return os.Open(filepath.Join(root, strings.Join(fi.Path, string(filepath.Separator))))
})
if err != nil {
err = fmt.Errorf("error generating pieces: %s", err)
}
return
} | [
"func",
"(",
"info",
"*",
"Info",
")",
"BuildFromFilePath",
"(",
"root",
"string",
")",
"(",
"err",
"error",
")",
"{",
"info",
".",
"Name",
"=",
"filepath",
".",
"Base",
"(",
"root",
")",
"\n",
"info",
".",
"Files",
"=",
"nil",
"\n",
"err",
"=",
"filepath",
".",
"Walk",
"(",
"root",
",",
"func",
"(",
"path",
"string",
",",
"fi",
"os",
".",
"FileInfo",
",",
"err",
"error",
")",
"error",
"{",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"if",
"fi",
".",
"IsDir",
"(",
")",
"{",
"// Directories are implicit in torrent files.",
"return",
"nil",
"\n",
"}",
"else",
"if",
"path",
"==",
"root",
"{",
"// The root is a file.",
"info",
".",
"Length",
"=",
"fi",
".",
"Size",
"(",
")",
"\n",
"return",
"nil",
"\n",
"}",
"\n",
"relPath",
",",
"err",
":=",
"filepath",
".",
"Rel",
"(",
"root",
",",
"path",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"info",
".",
"Files",
"=",
"append",
"(",
"info",
".",
"Files",
",",
"FileInfo",
"{",
"Path",
":",
"strings",
".",
"Split",
"(",
"relPath",
",",
"string",
"(",
"filepath",
".",
"Separator",
")",
")",
",",
"Length",
":",
"fi",
".",
"Size",
"(",
")",
",",
"}",
")",
"\n",
"return",
"nil",
"\n",
"}",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"slices",
".",
"Sort",
"(",
"info",
".",
"Files",
",",
"func",
"(",
"l",
",",
"r",
"FileInfo",
")",
"bool",
"{",
"return",
"strings",
".",
"Join",
"(",
"l",
".",
"Path",
",",
"\"",
"\"",
")",
"<",
"strings",
".",
"Join",
"(",
"r",
".",
"Path",
",",
"\"",
"\"",
")",
"\n",
"}",
")",
"\n",
"err",
"=",
"info",
".",
"GeneratePieces",
"(",
"func",
"(",
"fi",
"FileInfo",
")",
"(",
"io",
".",
"ReadCloser",
",",
"error",
")",
"{",
"return",
"os",
".",
"Open",
"(",
"filepath",
".",
"Join",
"(",
"root",
",",
"strings",
".",
"Join",
"(",
"fi",
".",
"Path",
",",
"string",
"(",
"filepath",
".",
"Separator",
")",
")",
")",
")",
"\n",
"}",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"err",
"=",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"return",
"\n",
"}"
] | // This is a helper that sets Files and Pieces from a root path and its
// children. | [
"This",
"is",
"a",
"helper",
"that",
"sets",
"Files",
"and",
"Pieces",
"from",
"a",
"root",
"path",
"and",
"its",
"children",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/metainfo/info.go#L29-L67 | train |
anacrolix/torrent | metainfo/info.go | writeFiles | func (info *Info) writeFiles(w io.Writer, open func(fi FileInfo) (io.ReadCloser, error)) error {
for _, fi := range info.UpvertedFiles() {
r, err := open(fi)
if err != nil {
return fmt.Errorf("error opening %v: %s", fi, err)
}
wn, err := io.CopyN(w, r, fi.Length)
r.Close()
if wn != fi.Length {
return fmt.Errorf("error copying %v: %s", fi, err)
}
}
return nil
} | go | func (info *Info) writeFiles(w io.Writer, open func(fi FileInfo) (io.ReadCloser, error)) error {
for _, fi := range info.UpvertedFiles() {
r, err := open(fi)
if err != nil {
return fmt.Errorf("error opening %v: %s", fi, err)
}
wn, err := io.CopyN(w, r, fi.Length)
r.Close()
if wn != fi.Length {
return fmt.Errorf("error copying %v: %s", fi, err)
}
}
return nil
} | [
"func",
"(",
"info",
"*",
"Info",
")",
"writeFiles",
"(",
"w",
"io",
".",
"Writer",
",",
"open",
"func",
"(",
"fi",
"FileInfo",
")",
"(",
"io",
".",
"ReadCloser",
",",
"error",
")",
")",
"error",
"{",
"for",
"_",
",",
"fi",
":=",
"range",
"info",
".",
"UpvertedFiles",
"(",
")",
"{",
"r",
",",
"err",
":=",
"open",
"(",
"fi",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"fi",
",",
"err",
")",
"\n",
"}",
"\n",
"wn",
",",
"err",
":=",
"io",
".",
"CopyN",
"(",
"w",
",",
"r",
",",
"fi",
".",
"Length",
")",
"\n",
"r",
".",
"Close",
"(",
")",
"\n",
"if",
"wn",
"!=",
"fi",
".",
"Length",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"fi",
",",
"err",
")",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"nil",
"\n",
"}"
] | // Concatenates all the files in the torrent into w. open is a function that
// gets at the contents of the given file. | [
"Concatenates",
"all",
"the",
"files",
"in",
"the",
"torrent",
"into",
"w",
".",
"open",
"is",
"a",
"function",
"that",
"gets",
"at",
"the",
"contents",
"of",
"the",
"given",
"file",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/metainfo/info.go#L71-L84 | train |
anacrolix/torrent | metainfo/info.go | UpvertedFiles | func (info *Info) UpvertedFiles() []FileInfo {
if len(info.Files) == 0 {
return []FileInfo{{
Length: info.Length,
// Callers should determine that Info.Name is the basename, and
// thus a regular file.
Path: nil,
}}
}
return info.Files
} | go | func (info *Info) UpvertedFiles() []FileInfo {
if len(info.Files) == 0 {
return []FileInfo{{
Length: info.Length,
// Callers should determine that Info.Name is the basename, and
// thus a regular file.
Path: nil,
}}
}
return info.Files
} | [
"func",
"(",
"info",
"*",
"Info",
")",
"UpvertedFiles",
"(",
")",
"[",
"]",
"FileInfo",
"{",
"if",
"len",
"(",
"info",
".",
"Files",
")",
"==",
"0",
"{",
"return",
"[",
"]",
"FileInfo",
"{",
"{",
"Length",
":",
"info",
".",
"Length",
",",
"// Callers should determine that Info.Name is the basename, and",
"// thus a regular file.",
"Path",
":",
"nil",
",",
"}",
"}",
"\n",
"}",
"\n",
"return",
"info",
".",
"Files",
"\n",
"}"
] | // The files field, converted up from the old single-file in the parent info
// dict if necessary. This is a helper to avoid having to conditionally handle
// single and multi-file torrent infos. | [
"The",
"files",
"field",
"converted",
"up",
"from",
"the",
"old",
"single",
"-",
"file",
"in",
"the",
"parent",
"info",
"dict",
"if",
"necessary",
".",
"This",
"is",
"a",
"helper",
"to",
"avoid",
"having",
"to",
"conditionally",
"handle",
"single",
"and",
"multi",
"-",
"file",
"torrent",
"infos",
"."
] | d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb | https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/metainfo/info.go#L142-L152 | train |
kubernetes-retired/heapster | metrics/util/label_copier.go | makeStoredLabels | func makeStoredLabels(labels []string) map[string]string {
storedLabels := make(map[string]string)
for _, s := range labels {
split := strings.SplitN(s, "=", 2)
if len(split) == 1 {
storedLabels[split[0]] = split[0]
} else {
storedLabels[split[1]] = split[0]
}
}
return storedLabels
} | go | func makeStoredLabels(labels []string) map[string]string {
storedLabels := make(map[string]string)
for _, s := range labels {
split := strings.SplitN(s, "=", 2)
if len(split) == 1 {
storedLabels[split[0]] = split[0]
} else {
storedLabels[split[1]] = split[0]
}
}
return storedLabels
} | [
"func",
"makeStoredLabels",
"(",
"labels",
"[",
"]",
"string",
")",
"map",
"[",
"string",
"]",
"string",
"{",
"storedLabels",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"string",
")",
"\n",
"for",
"_",
",",
"s",
":=",
"range",
"labels",
"{",
"split",
":=",
"strings",
".",
"SplitN",
"(",
"s",
",",
"\"",
"\"",
",",
"2",
")",
"\n",
"if",
"len",
"(",
"split",
")",
"==",
"1",
"{",
"storedLabels",
"[",
"split",
"[",
"0",
"]",
"]",
"=",
"split",
"[",
"0",
"]",
"\n",
"}",
"else",
"{",
"storedLabels",
"[",
"split",
"[",
"1",
"]",
"]",
"=",
"split",
"[",
"0",
"]",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"storedLabels",
"\n",
"}"
] | // makeStoredLabels converts labels into a map for quicker retrieval.
// Incoming labels, if desired, may contain mappings in format "newName=oldName" | [
"makeStoredLabels",
"converts",
"labels",
"into",
"a",
"map",
"for",
"quicker",
"retrieval",
".",
"Incoming",
"labels",
"if",
"desired",
"may",
"contain",
"mappings",
"in",
"format",
"newName",
"=",
"oldName"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/util/label_copier.go#L56-L67 | train |
kubernetes-retired/heapster | metrics/util/label_copier.go | makeIgnoredLabels | func makeIgnoredLabels(labels []string) map[string]string {
ignoredLabels := make(map[string]string)
for _, s := range labels {
ignoredLabels[s] = ""
}
return ignoredLabels
} | go | func makeIgnoredLabels(labels []string) map[string]string {
ignoredLabels := make(map[string]string)
for _, s := range labels {
ignoredLabels[s] = ""
}
return ignoredLabels
} | [
"func",
"makeIgnoredLabels",
"(",
"labels",
"[",
"]",
"string",
")",
"map",
"[",
"string",
"]",
"string",
"{",
"ignoredLabels",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"string",
")",
"\n",
"for",
"_",
",",
"s",
":=",
"range",
"labels",
"{",
"ignoredLabels",
"[",
"s",
"]",
"=",
"\"",
"\"",
"\n",
"}",
"\n",
"return",
"ignoredLabels",
"\n",
"}"
] | // makeIgnoredLabels converts label slice into a map for later use. | [
"makeIgnoredLabels",
"converts",
"label",
"slice",
"into",
"a",
"map",
"for",
"later",
"use",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/util/label_copier.go#L70-L76 | train |
kubernetes-retired/heapster | metrics/util/label_copier.go | NewLabelCopier | func NewLabelCopier(separator string, storedLabels, ignoredLabels []string) (*LabelCopier, error) {
return &LabelCopier{
labelSeparator: separator,
storedLabels: makeStoredLabels(storedLabels),
ignoredLabels: makeIgnoredLabels(ignoredLabels),
}, nil
} | go | func NewLabelCopier(separator string, storedLabels, ignoredLabels []string) (*LabelCopier, error) {
return &LabelCopier{
labelSeparator: separator,
storedLabels: makeStoredLabels(storedLabels),
ignoredLabels: makeIgnoredLabels(ignoredLabels),
}, nil
} | [
"func",
"NewLabelCopier",
"(",
"separator",
"string",
",",
"storedLabels",
",",
"ignoredLabels",
"[",
"]",
"string",
")",
"(",
"*",
"LabelCopier",
",",
"error",
")",
"{",
"return",
"&",
"LabelCopier",
"{",
"labelSeparator",
":",
"separator",
",",
"storedLabels",
":",
"makeStoredLabels",
"(",
"storedLabels",
")",
",",
"ignoredLabels",
":",
"makeIgnoredLabels",
"(",
"ignoredLabels",
")",
",",
"}",
",",
"nil",
"\n",
"}"
] | // NewLabelCopier creates a new instance of LabelCopier type | [
"NewLabelCopier",
"creates",
"a",
"new",
"instance",
"of",
"LabelCopier",
"type"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/util/label_copier.go#L79-L85 | train |
kubernetes-retired/heapster | events/sinks/influxdb/influxdb.go | getEventValue | func getEventValue(event *kube_api.Event) (string, error) {
// TODO: check whether indenting is required.
bytes, err := json.MarshalIndent(event, "", " ")
if err != nil {
return "", err
}
return string(bytes), nil
} | go | func getEventValue(event *kube_api.Event) (string, error) {
// TODO: check whether indenting is required.
bytes, err := json.MarshalIndent(event, "", " ")
if err != nil {
return "", err
}
return string(bytes), nil
} | [
"func",
"getEventValue",
"(",
"event",
"*",
"kube_api",
".",
"Event",
")",
"(",
"string",
",",
"error",
")",
"{",
"// TODO: check whether indenting is required.",
"bytes",
",",
"err",
":=",
"json",
".",
"MarshalIndent",
"(",
"event",
",",
"\"",
"\"",
",",
"\"",
"\"",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"\"",
"\"",
",",
"err",
"\n",
"}",
"\n",
"return",
"string",
"(",
"bytes",
")",
",",
"nil",
"\n",
"}"
] | // Generate point value for event | [
"Generate",
"point",
"value",
"for",
"event"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/events/sinks/influxdb/influxdb.go#L61-L68 | train |
kubernetes-retired/heapster | events/sinks/influxdb/influxdb.go | newSink | func newSink(c influxdb_common.InfluxdbConfig) core.EventSink {
client, err := influxdb_common.NewClient(c)
if err != nil {
glog.Errorf("issues while creating an InfluxDB sink: %v, will retry on use", err)
}
return &influxdbSink{
client: client, // can be nil
c: c,
}
} | go | func newSink(c influxdb_common.InfluxdbConfig) core.EventSink {
client, err := influxdb_common.NewClient(c)
if err != nil {
glog.Errorf("issues while creating an InfluxDB sink: %v, will retry on use", err)
}
return &influxdbSink{
client: client, // can be nil
c: c,
}
} | [
"func",
"newSink",
"(",
"c",
"influxdb_common",
".",
"InfluxdbConfig",
")",
"core",
".",
"EventSink",
"{",
"client",
",",
"err",
":=",
"influxdb_common",
".",
"NewClient",
"(",
"c",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"return",
"&",
"influxdbSink",
"{",
"client",
":",
"client",
",",
"// can be nil",
"c",
":",
"c",
",",
"}",
"\n",
"}"
] | // Returns a thread-safe implementation of core.EventSink for InfluxDB. | [
"Returns",
"a",
"thread",
"-",
"safe",
"implementation",
"of",
"core",
".",
"EventSink",
"for",
"InfluxDB",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/events/sinks/influxdb/influxdb.go#L230-L239 | train |
kubernetes-retired/heapster | metrics/sinks/opentsdb/driver.go | metricToPoint | func (tsdbSink *openTSDBSink) metricToPoint(name string, value core.MetricValue, timestamp time.Time, labels map[string]string) opentsdbclient.DataPoint {
seriesName := strings.Replace(toValidOpenTsdbName(name), "/", "_", -1)
if value.MetricType.String() != "" {
seriesName = fmt.Sprintf("%s_%s", seriesName, value.MetricType.String())
}
datapoint := opentsdbclient.DataPoint{
Metric: seriesName,
Tags: make(map[string]string, len(labels)),
Timestamp: timestamp.Unix(),
}
if value.ValueType == core.ValueInt64 {
datapoint.Value = value.IntValue
} else {
datapoint.Value = value.FloatValue
}
for key, value := range labels {
key = toValidOpenTsdbName(key)
value = toValidOpenTsdbName(value)
if value != "" {
datapoint.Tags[key] = value
}
}
tsdbSink.putDefaultTags(&datapoint)
return datapoint
} | go | func (tsdbSink *openTSDBSink) metricToPoint(name string, value core.MetricValue, timestamp time.Time, labels map[string]string) opentsdbclient.DataPoint {
seriesName := strings.Replace(toValidOpenTsdbName(name), "/", "_", -1)
if value.MetricType.String() != "" {
seriesName = fmt.Sprintf("%s_%s", seriesName, value.MetricType.String())
}
datapoint := opentsdbclient.DataPoint{
Metric: seriesName,
Tags: make(map[string]string, len(labels)),
Timestamp: timestamp.Unix(),
}
if value.ValueType == core.ValueInt64 {
datapoint.Value = value.IntValue
} else {
datapoint.Value = value.FloatValue
}
for key, value := range labels {
key = toValidOpenTsdbName(key)
value = toValidOpenTsdbName(value)
if value != "" {
datapoint.Tags[key] = value
}
}
tsdbSink.putDefaultTags(&datapoint)
return datapoint
} | [
"func",
"(",
"tsdbSink",
"*",
"openTSDBSink",
")",
"metricToPoint",
"(",
"name",
"string",
",",
"value",
"core",
".",
"MetricValue",
",",
"timestamp",
"time",
".",
"Time",
",",
"labels",
"map",
"[",
"string",
"]",
"string",
")",
"opentsdbclient",
".",
"DataPoint",
"{",
"seriesName",
":=",
"strings",
".",
"Replace",
"(",
"toValidOpenTsdbName",
"(",
"name",
")",
",",
"\"",
"\"",
",",
"\"",
"\"",
",",
"-",
"1",
")",
"\n\n",
"if",
"value",
".",
"MetricType",
".",
"String",
"(",
")",
"!=",
"\"",
"\"",
"{",
"seriesName",
"=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"seriesName",
",",
"value",
".",
"MetricType",
".",
"String",
"(",
")",
")",
"\n",
"}",
"\n\n",
"datapoint",
":=",
"opentsdbclient",
".",
"DataPoint",
"{",
"Metric",
":",
"seriesName",
",",
"Tags",
":",
"make",
"(",
"map",
"[",
"string",
"]",
"string",
",",
"len",
"(",
"labels",
")",
")",
",",
"Timestamp",
":",
"timestamp",
".",
"Unix",
"(",
")",
",",
"}",
"\n",
"if",
"value",
".",
"ValueType",
"==",
"core",
".",
"ValueInt64",
"{",
"datapoint",
".",
"Value",
"=",
"value",
".",
"IntValue",
"\n",
"}",
"else",
"{",
"datapoint",
".",
"Value",
"=",
"value",
".",
"FloatValue",
"\n",
"}",
"\n\n",
"for",
"key",
",",
"value",
":=",
"range",
"labels",
"{",
"key",
"=",
"toValidOpenTsdbName",
"(",
"key",
")",
"\n",
"value",
"=",
"toValidOpenTsdbName",
"(",
"value",
")",
"\n\n",
"if",
"value",
"!=",
"\"",
"\"",
"{",
"datapoint",
".",
"Tags",
"[",
"key",
"]",
"=",
"value",
"\n",
"}",
"\n",
"}",
"\n\n",
"tsdbSink",
".",
"putDefaultTags",
"(",
"&",
"datapoint",
")",
"\n",
"return",
"datapoint",
"\n",
"}"
] | // timeSeriesToPoint transfers the contents holding in the given pointer of sink_api.Timeseries
// into the instance of opentsdbclient.DataPoint | [
"timeSeriesToPoint",
"transfers",
"the",
"contents",
"holding",
"in",
"the",
"given",
"pointer",
"of",
"sink_api",
".",
"Timeseries",
"into",
"the",
"instance",
"of",
"opentsdbclient",
".",
"DataPoint"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/opentsdb/driver.go#L117-L146 | train |
kubernetes-retired/heapster | metrics/sinks/opentsdb/driver.go | putDefaultTags | func (tsdbSink *openTSDBSink) putDefaultTags(datapoint *opentsdbclient.DataPoint) {
datapoint.Tags[clusterNameTagName] = tsdbSink.clusterName
} | go | func (tsdbSink *openTSDBSink) putDefaultTags(datapoint *opentsdbclient.DataPoint) {
datapoint.Tags[clusterNameTagName] = tsdbSink.clusterName
} | [
"func",
"(",
"tsdbSink",
"*",
"openTSDBSink",
")",
"putDefaultTags",
"(",
"datapoint",
"*",
"opentsdbclient",
".",
"DataPoint",
")",
"{",
"datapoint",
".",
"Tags",
"[",
"clusterNameTagName",
"]",
"=",
"tsdbSink",
".",
"clusterName",
"\n",
"}"
] | // putDefaultTags just fills in the default key-value pair for the tags.
// OpenTSDB requires at least one non-empty tag otherwise the OpenTSDB will return error and the operation of putting
// datapoint will be failed. | [
"putDefaultTags",
"just",
"fills",
"in",
"the",
"default",
"key",
"-",
"value",
"pair",
"for",
"the",
"tags",
".",
"OpenTSDB",
"requires",
"at",
"least",
"one",
"non",
"-",
"empty",
"tag",
"otherwise",
"the",
"OpenTSDB",
"will",
"return",
"error",
"and",
"the",
"operation",
"of",
"putting",
"datapoint",
"will",
"be",
"failed",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/opentsdb/driver.go#L151-L153 | train |
kubernetes-retired/heapster | metrics/api/v1/api.go | NewApi | func NewApi(runningInKubernetes bool, metricSink *metricsink.MetricSink, historicalSource core.HistoricalSource, disableMetricExport bool) *Api {
gkeMetrics := make(map[string]core.MetricDescriptor)
gkeLabels := make(map[string]core.LabelDescriptor)
for _, val := range core.StandardMetrics {
gkeMetrics[val.Name] = val.MetricDescriptor
}
for _, val := range core.LabeledMetrics {
gkeMetrics[val.Name] = val.MetricDescriptor
}
gkeMetrics[core.MetricCpuLimit.Name] = core.MetricCpuLimit.MetricDescriptor
gkeMetrics[core.MetricMemoryLimit.Name] = core.MetricMemoryLimit.MetricDescriptor
for _, val := range core.CommonLabels() {
gkeLabels[val.Key] = val
}
for _, val := range core.ContainerLabels() {
gkeLabels[val.Key] = val
}
for _, val := range core.PodLabels() {
gkeLabels[val.Key] = val
}
return &Api{
runningInKubernetes: runningInKubernetes,
metricSink: metricSink,
historicalSource: historicalSource,
gkeMetrics: gkeMetrics,
gkeLabels: gkeLabels,
disabled: disableMetricExport,
}
} | go | func NewApi(runningInKubernetes bool, metricSink *metricsink.MetricSink, historicalSource core.HistoricalSource, disableMetricExport bool) *Api {
gkeMetrics := make(map[string]core.MetricDescriptor)
gkeLabels := make(map[string]core.LabelDescriptor)
for _, val := range core.StandardMetrics {
gkeMetrics[val.Name] = val.MetricDescriptor
}
for _, val := range core.LabeledMetrics {
gkeMetrics[val.Name] = val.MetricDescriptor
}
gkeMetrics[core.MetricCpuLimit.Name] = core.MetricCpuLimit.MetricDescriptor
gkeMetrics[core.MetricMemoryLimit.Name] = core.MetricMemoryLimit.MetricDescriptor
for _, val := range core.CommonLabels() {
gkeLabels[val.Key] = val
}
for _, val := range core.ContainerLabels() {
gkeLabels[val.Key] = val
}
for _, val := range core.PodLabels() {
gkeLabels[val.Key] = val
}
return &Api{
runningInKubernetes: runningInKubernetes,
metricSink: metricSink,
historicalSource: historicalSource,
gkeMetrics: gkeMetrics,
gkeLabels: gkeLabels,
disabled: disableMetricExport,
}
} | [
"func",
"NewApi",
"(",
"runningInKubernetes",
"bool",
",",
"metricSink",
"*",
"metricsink",
".",
"MetricSink",
",",
"historicalSource",
"core",
".",
"HistoricalSource",
",",
"disableMetricExport",
"bool",
")",
"*",
"Api",
"{",
"gkeMetrics",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"core",
".",
"MetricDescriptor",
")",
"\n",
"gkeLabels",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"core",
".",
"LabelDescriptor",
")",
"\n",
"for",
"_",
",",
"val",
":=",
"range",
"core",
".",
"StandardMetrics",
"{",
"gkeMetrics",
"[",
"val",
".",
"Name",
"]",
"=",
"val",
".",
"MetricDescriptor",
"\n",
"}",
"\n",
"for",
"_",
",",
"val",
":=",
"range",
"core",
".",
"LabeledMetrics",
"{",
"gkeMetrics",
"[",
"val",
".",
"Name",
"]",
"=",
"val",
".",
"MetricDescriptor",
"\n",
"}",
"\n",
"gkeMetrics",
"[",
"core",
".",
"MetricCpuLimit",
".",
"Name",
"]",
"=",
"core",
".",
"MetricCpuLimit",
".",
"MetricDescriptor",
"\n",
"gkeMetrics",
"[",
"core",
".",
"MetricMemoryLimit",
".",
"Name",
"]",
"=",
"core",
".",
"MetricMemoryLimit",
".",
"MetricDescriptor",
"\n\n",
"for",
"_",
",",
"val",
":=",
"range",
"core",
".",
"CommonLabels",
"(",
")",
"{",
"gkeLabels",
"[",
"val",
".",
"Key",
"]",
"=",
"val",
"\n",
"}",
"\n",
"for",
"_",
",",
"val",
":=",
"range",
"core",
".",
"ContainerLabels",
"(",
")",
"{",
"gkeLabels",
"[",
"val",
".",
"Key",
"]",
"=",
"val",
"\n",
"}",
"\n",
"for",
"_",
",",
"val",
":=",
"range",
"core",
".",
"PodLabels",
"(",
")",
"{",
"gkeLabels",
"[",
"val",
".",
"Key",
"]",
"=",
"val",
"\n",
"}",
"\n\n",
"return",
"&",
"Api",
"{",
"runningInKubernetes",
":",
"runningInKubernetes",
",",
"metricSink",
":",
"metricSink",
",",
"historicalSource",
":",
"historicalSource",
",",
"gkeMetrics",
":",
"gkeMetrics",
",",
"gkeLabels",
":",
"gkeLabels",
",",
"disabled",
":",
"disableMetricExport",
",",
"}",
"\n",
"}"
] | // Create a new Api to serve from the specified cache. | [
"Create",
"a",
"new",
"Api",
"to",
"serve",
"from",
"the",
"specified",
"cache",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/api.go#L42-L72 | train |
kubernetes-retired/heapster | metrics/api/v1/api.go | Register | func (a *Api) Register(container *restful.Container) {
ws := new(restful.WebService)
ws.Path("/api/v1/metric-export").
Doc("Exports the latest point for all Heapster metrics").
Produces(restful.MIME_JSON)
ws.Route(ws.GET("").
To(a.exportMetrics).
Doc("export the latest data point for all metrics").
Operation("exportMetrics").
Writes([]*types.Timeseries{}))
container.Add(ws)
ws = new(restful.WebService)
ws.Path("/api/v1/metric-export-schema").
Doc("Schema for metrics exported by heapster").
Produces(restful.MIME_JSON)
ws.Route(ws.GET("").
To(a.exportMetricsSchema).
Doc("export the schema for all metrics").
Operation("exportmetricsSchema").
Writes(types.TimeseriesSchema{}))
container.Add(ws)
if a.metricSink != nil {
a.RegisterModel(container)
}
if a.historicalSource != nil {
a.RegisterHistorical(container)
}
} | go | func (a *Api) Register(container *restful.Container) {
ws := new(restful.WebService)
ws.Path("/api/v1/metric-export").
Doc("Exports the latest point for all Heapster metrics").
Produces(restful.MIME_JSON)
ws.Route(ws.GET("").
To(a.exportMetrics).
Doc("export the latest data point for all metrics").
Operation("exportMetrics").
Writes([]*types.Timeseries{}))
container.Add(ws)
ws = new(restful.WebService)
ws.Path("/api/v1/metric-export-schema").
Doc("Schema for metrics exported by heapster").
Produces(restful.MIME_JSON)
ws.Route(ws.GET("").
To(a.exportMetricsSchema).
Doc("export the schema for all metrics").
Operation("exportmetricsSchema").
Writes(types.TimeseriesSchema{}))
container.Add(ws)
if a.metricSink != nil {
a.RegisterModel(container)
}
if a.historicalSource != nil {
a.RegisterHistorical(container)
}
} | [
"func",
"(",
"a",
"*",
"Api",
")",
"Register",
"(",
"container",
"*",
"restful",
".",
"Container",
")",
"{",
"ws",
":=",
"new",
"(",
"restful",
".",
"WebService",
")",
"\n",
"ws",
".",
"Path",
"(",
"\"",
"\"",
")",
".",
"Doc",
"(",
"\"",
"\"",
")",
".",
"Produces",
"(",
"restful",
".",
"MIME_JSON",
")",
"\n",
"ws",
".",
"Route",
"(",
"ws",
".",
"GET",
"(",
"\"",
"\"",
")",
".",
"To",
"(",
"a",
".",
"exportMetrics",
")",
".",
"Doc",
"(",
"\"",
"\"",
")",
".",
"Operation",
"(",
"\"",
"\"",
")",
".",
"Writes",
"(",
"[",
"]",
"*",
"types",
".",
"Timeseries",
"{",
"}",
")",
")",
"\n",
"container",
".",
"Add",
"(",
"ws",
")",
"\n",
"ws",
"=",
"new",
"(",
"restful",
".",
"WebService",
")",
"\n",
"ws",
".",
"Path",
"(",
"\"",
"\"",
")",
".",
"Doc",
"(",
"\"",
"\"",
")",
".",
"Produces",
"(",
"restful",
".",
"MIME_JSON",
")",
"\n",
"ws",
".",
"Route",
"(",
"ws",
".",
"GET",
"(",
"\"",
"\"",
")",
".",
"To",
"(",
"a",
".",
"exportMetricsSchema",
")",
".",
"Doc",
"(",
"\"",
"\"",
")",
".",
"Operation",
"(",
"\"",
"\"",
")",
".",
"Writes",
"(",
"types",
".",
"TimeseriesSchema",
"{",
"}",
")",
")",
"\n",
"container",
".",
"Add",
"(",
"ws",
")",
"\n\n",
"if",
"a",
".",
"metricSink",
"!=",
"nil",
"{",
"a",
".",
"RegisterModel",
"(",
"container",
")",
"\n",
"}",
"\n\n",
"if",
"a",
".",
"historicalSource",
"!=",
"nil",
"{",
"a",
".",
"RegisterHistorical",
"(",
"container",
")",
"\n",
"}",
"\n",
"}"
] | // Register the mainApi on the specified endpoint. | [
"Register",
"the",
"mainApi",
"on",
"the",
"specified",
"endpoint",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/api.go#L75-L104 | train |
kubernetes-retired/heapster | metrics/sinks/hawkular/driver.go | NewHawkularSink | func NewHawkularSink(u *url.URL) (core.DataSink, error) {
sink := &hawkularSink{
uri: u,
batchSize: 1000,
}
if err := sink.init(); err != nil {
return nil, err
}
metrics := make([]core.MetricDescriptor, 0, len(core.AllMetrics))
for _, metric := range core.AllMetrics {
metrics = append(metrics, metric.MetricDescriptor)
}
sink.Register(metrics)
return sink, nil
} | go | func NewHawkularSink(u *url.URL) (core.DataSink, error) {
sink := &hawkularSink{
uri: u,
batchSize: 1000,
}
if err := sink.init(); err != nil {
return nil, err
}
metrics := make([]core.MetricDescriptor, 0, len(core.AllMetrics))
for _, metric := range core.AllMetrics {
metrics = append(metrics, metric.MetricDescriptor)
}
sink.Register(metrics)
return sink, nil
} | [
"func",
"NewHawkularSink",
"(",
"u",
"*",
"url",
".",
"URL",
")",
"(",
"core",
".",
"DataSink",
",",
"error",
")",
"{",
"sink",
":=",
"&",
"hawkularSink",
"{",
"uri",
":",
"u",
",",
"batchSize",
":",
"1000",
",",
"}",
"\n",
"if",
"err",
":=",
"sink",
".",
"init",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n\n",
"metrics",
":=",
"make",
"(",
"[",
"]",
"core",
".",
"MetricDescriptor",
",",
"0",
",",
"len",
"(",
"core",
".",
"AllMetrics",
")",
")",
"\n",
"for",
"_",
",",
"metric",
":=",
"range",
"core",
".",
"AllMetrics",
"{",
"metrics",
"=",
"append",
"(",
"metrics",
",",
"metric",
".",
"MetricDescriptor",
")",
"\n",
"}",
"\n",
"sink",
".",
"Register",
"(",
"metrics",
")",
"\n",
"return",
"sink",
",",
"nil",
"\n",
"}"
] | // NewHawkularSink Creates and returns a new hawkularSink instance | [
"NewHawkularSink",
"Creates",
"and",
"returns",
"a",
"new",
"hawkularSink",
"instance"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/hawkular/driver.go#L178-L193 | train |
kubernetes-retired/heapster | common/elasticsearch/elasticsearch.go | SaveData | func (esSvc *ElasticSearchService) SaveData(date time.Time, typeName string, sinkData []interface{}) error {
if typeName == "" || len(sinkData) == 0 {
return nil
}
indexName := esSvc.Index(date)
// Use the IndexExists service to check if a specified index exists.
exists, err := esSvc.EsClient.IndexExists(indexName)
if err != nil {
return err
}
if !exists {
// Create a new index.
createIndex, err := esSvc.EsClient.CreateIndex(indexName, mapping)
if err != nil {
return err
}
ack := false
switch i := createIndex.(type) {
case *elastic2.IndicesCreateResult:
ack = i.Acknowledged
case *elastic5.IndicesCreateResult:
ack = i.Acknowledged
}
if !ack {
return errors.New("Failed to acknoledge index creation")
}
}
aliases, err := esSvc.EsClient.GetAliases(indexName)
if err != nil {
return err
}
aliasName := esSvc.IndexAlias(typeName)
hasAlias := false
switch a := aliases.(type) {
case *elastic2.AliasesResult:
hasAlias = a.Indices[indexName].HasAlias(aliasName)
case *elastic5.AliasesResult:
hasAlias = a.Indices[indexName].HasAlias(aliasName)
}
if !hasAlias {
createAlias, err := esSvc.EsClient.AddAlias(indexName, esSvc.IndexAlias(typeName))
if err != nil {
return err
}
ack := false
switch i := createAlias.(type) {
case *elastic2.AliasResult:
ack = i.Acknowledged
case *elastic5.AliasResult:
ack = i.Acknowledged
}
if !ack {
return errors.New("Failed to acknoledge index alias creation")
}
}
for _, data := range sinkData {
esSvc.EsClient.AddBulkReq(indexName, typeName, data)
}
return nil
} | go | func (esSvc *ElasticSearchService) SaveData(date time.Time, typeName string, sinkData []interface{}) error {
if typeName == "" || len(sinkData) == 0 {
return nil
}
indexName := esSvc.Index(date)
// Use the IndexExists service to check if a specified index exists.
exists, err := esSvc.EsClient.IndexExists(indexName)
if err != nil {
return err
}
if !exists {
// Create a new index.
createIndex, err := esSvc.EsClient.CreateIndex(indexName, mapping)
if err != nil {
return err
}
ack := false
switch i := createIndex.(type) {
case *elastic2.IndicesCreateResult:
ack = i.Acknowledged
case *elastic5.IndicesCreateResult:
ack = i.Acknowledged
}
if !ack {
return errors.New("Failed to acknoledge index creation")
}
}
aliases, err := esSvc.EsClient.GetAliases(indexName)
if err != nil {
return err
}
aliasName := esSvc.IndexAlias(typeName)
hasAlias := false
switch a := aliases.(type) {
case *elastic2.AliasesResult:
hasAlias = a.Indices[indexName].HasAlias(aliasName)
case *elastic5.AliasesResult:
hasAlias = a.Indices[indexName].HasAlias(aliasName)
}
if !hasAlias {
createAlias, err := esSvc.EsClient.AddAlias(indexName, esSvc.IndexAlias(typeName))
if err != nil {
return err
}
ack := false
switch i := createAlias.(type) {
case *elastic2.AliasResult:
ack = i.Acknowledged
case *elastic5.AliasResult:
ack = i.Acknowledged
}
if !ack {
return errors.New("Failed to acknoledge index alias creation")
}
}
for _, data := range sinkData {
esSvc.EsClient.AddBulkReq(indexName, typeName, data)
}
return nil
} | [
"func",
"(",
"esSvc",
"*",
"ElasticSearchService",
")",
"SaveData",
"(",
"date",
"time",
".",
"Time",
",",
"typeName",
"string",
",",
"sinkData",
"[",
"]",
"interface",
"{",
"}",
")",
"error",
"{",
"if",
"typeName",
"==",
"\"",
"\"",
"||",
"len",
"(",
"sinkData",
")",
"==",
"0",
"{",
"return",
"nil",
"\n",
"}",
"\n\n",
"indexName",
":=",
"esSvc",
".",
"Index",
"(",
"date",
")",
"\n\n",
"// Use the IndexExists service to check if a specified index exists.",
"exists",
",",
"err",
":=",
"esSvc",
".",
"EsClient",
".",
"IndexExists",
"(",
"indexName",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"if",
"!",
"exists",
"{",
"// Create a new index.",
"createIndex",
",",
"err",
":=",
"esSvc",
".",
"EsClient",
".",
"CreateIndex",
"(",
"indexName",
",",
"mapping",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"ack",
":=",
"false",
"\n",
"switch",
"i",
":=",
"createIndex",
".",
"(",
"type",
")",
"{",
"case",
"*",
"elastic2",
".",
"IndicesCreateResult",
":",
"ack",
"=",
"i",
".",
"Acknowledged",
"\n",
"case",
"*",
"elastic5",
".",
"IndicesCreateResult",
":",
"ack",
"=",
"i",
".",
"Acknowledged",
"\n",
"}",
"\n",
"if",
"!",
"ack",
"{",
"return",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"aliases",
",",
"err",
":=",
"esSvc",
".",
"EsClient",
".",
"GetAliases",
"(",
"indexName",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"aliasName",
":=",
"esSvc",
".",
"IndexAlias",
"(",
"typeName",
")",
"\n\n",
"hasAlias",
":=",
"false",
"\n",
"switch",
"a",
":=",
"aliases",
".",
"(",
"type",
")",
"{",
"case",
"*",
"elastic2",
".",
"AliasesResult",
":",
"hasAlias",
"=",
"a",
".",
"Indices",
"[",
"indexName",
"]",
".",
"HasAlias",
"(",
"aliasName",
")",
"\n",
"case",
"*",
"elastic5",
".",
"AliasesResult",
":",
"hasAlias",
"=",
"a",
".",
"Indices",
"[",
"indexName",
"]",
".",
"HasAlias",
"(",
"aliasName",
")",
"\n",
"}",
"\n",
"if",
"!",
"hasAlias",
"{",
"createAlias",
",",
"err",
":=",
"esSvc",
".",
"EsClient",
".",
"AddAlias",
"(",
"indexName",
",",
"esSvc",
".",
"IndexAlias",
"(",
"typeName",
")",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"ack",
":=",
"false",
"\n",
"switch",
"i",
":=",
"createAlias",
".",
"(",
"type",
")",
"{",
"case",
"*",
"elastic2",
".",
"AliasResult",
":",
"ack",
"=",
"i",
".",
"Acknowledged",
"\n",
"case",
"*",
"elastic5",
".",
"AliasResult",
":",
"ack",
"=",
"i",
".",
"Acknowledged",
"\n",
"}",
"\n",
"if",
"!",
"ack",
"{",
"return",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"data",
":=",
"range",
"sinkData",
"{",
"esSvc",
".",
"EsClient",
".",
"AddBulkReq",
"(",
"indexName",
",",
"typeName",
",",
"data",
")",
"\n",
"}",
"\n\n",
"return",
"nil",
"\n",
"}"
] | // SaveDataIntoES save metrics and events to ES by using ES client | [
"SaveDataIntoES",
"save",
"metrics",
"and",
"events",
"to",
"ES",
"by",
"using",
"ES",
"client"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/common/elasticsearch/elasticsearch.go#L54-L122 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb.go | newSink | func newSink(c influxdb_common.InfluxdbConfig) core.DataSink {
client, err := influxdb_common.NewClient(c)
if err != nil {
glog.Errorf("issues while creating an InfluxDB sink: %v, will retry on use", err)
}
return &influxdbSink{
client: client, // can be nil
c: c,
conChan: make(chan struct{}, c.Concurrency),
}
} | go | func newSink(c influxdb_common.InfluxdbConfig) core.DataSink {
client, err := influxdb_common.NewClient(c)
if err != nil {
glog.Errorf("issues while creating an InfluxDB sink: %v, will retry on use", err)
}
return &influxdbSink{
client: client, // can be nil
c: c,
conChan: make(chan struct{}, c.Concurrency),
}
} | [
"func",
"newSink",
"(",
"c",
"influxdb_common",
".",
"InfluxdbConfig",
")",
"core",
".",
"DataSink",
"{",
"client",
",",
"err",
":=",
"influxdb_common",
".",
"NewClient",
"(",
"c",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"return",
"&",
"influxdbSink",
"{",
"client",
":",
"client",
",",
"// can be nil",
"c",
":",
"c",
",",
"conChan",
":",
"make",
"(",
"chan",
"struct",
"{",
"}",
",",
"c",
".",
"Concurrency",
")",
",",
"}",
"\n",
"}"
] | // Returns a thread-compatible implementation of influxdb interactions. | [
"Returns",
"a",
"thread",
"-",
"compatible",
"implementation",
"of",
"influxdb",
"interactions",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb.go#L294-L304 | train |
kubernetes-retired/heapster | integration/framework.go | ParseRBAC | func (self *realKubeFramework) ParseRBAC(filePath string) (*rbacv1.ClusterRoleBinding, error) {
obj, err := self.loadRBACObject(filePath)
if err != nil {
return nil, err
}
rbac, ok := obj.(*rbacv1.ClusterRoleBinding)
if !ok {
return nil, fmt.Errorf("Failed to cast clusterrolebinding: %v", obj)
}
return rbac, nil
} | go | func (self *realKubeFramework) ParseRBAC(filePath string) (*rbacv1.ClusterRoleBinding, error) {
obj, err := self.loadRBACObject(filePath)
if err != nil {
return nil, err
}
rbac, ok := obj.(*rbacv1.ClusterRoleBinding)
if !ok {
return nil, fmt.Errorf("Failed to cast clusterrolebinding: %v", obj)
}
return rbac, nil
} | [
"func",
"(",
"self",
"*",
"realKubeFramework",
")",
"ParseRBAC",
"(",
"filePath",
"string",
")",
"(",
"*",
"rbacv1",
".",
"ClusterRoleBinding",
",",
"error",
")",
"{",
"obj",
",",
"err",
":=",
"self",
".",
"loadRBACObject",
"(",
"filePath",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n\n",
"rbac",
",",
"ok",
":=",
"obj",
".",
"(",
"*",
"rbacv1",
".",
"ClusterRoleBinding",
")",
"\n",
"if",
"!",
"ok",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"obj",
")",
"\n",
"}",
"\n",
"return",
"rbac",
",",
"nil",
"\n",
"}"
] | // Parses and Returns a RBAC object contained in 'filePath' | [
"Parses",
"and",
"Returns",
"a",
"RBAC",
"object",
"contained",
"in",
"filePath"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/integration/framework.go#L402-L413 | train |
kubernetes-retired/heapster | integration/framework.go | CreateRBAC | func (self *realKubeFramework) CreateRBAC(rbac *rbacv1.ClusterRoleBinding) error {
_, err := self.kubeClient.RbacV1().ClusterRoleBindings().Create(rbac)
return err
} | go | func (self *realKubeFramework) CreateRBAC(rbac *rbacv1.ClusterRoleBinding) error {
_, err := self.kubeClient.RbacV1().ClusterRoleBindings().Create(rbac)
return err
} | [
"func",
"(",
"self",
"*",
"realKubeFramework",
")",
"CreateRBAC",
"(",
"rbac",
"*",
"rbacv1",
".",
"ClusterRoleBinding",
")",
"error",
"{",
"_",
",",
"err",
":=",
"self",
".",
"kubeClient",
".",
"RbacV1",
"(",
")",
".",
"ClusterRoleBindings",
"(",
")",
".",
"Create",
"(",
"rbac",
")",
"\n",
"return",
"err",
"\n",
"}"
] | // CreateRBAC creates the RBAC object | [
"CreateRBAC",
"creates",
"the",
"RBAC",
"object"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/integration/framework.go#L416-L419 | train |
kubernetes-retired/heapster | integration/framework.go | ParseServiceAccount | func (self *realKubeFramework) ParseServiceAccount(filePath string) (*v1.ServiceAccount, error) {
obj, err := self.loadObject(filePath)
if err != nil {
return nil, err
}
sa, ok := obj.(*v1.ServiceAccount)
if !ok {
return nil, fmt.Errorf("Failed to cast serviceaccount: %v", obj)
}
return sa, nil
} | go | func (self *realKubeFramework) ParseServiceAccount(filePath string) (*v1.ServiceAccount, error) {
obj, err := self.loadObject(filePath)
if err != nil {
return nil, err
}
sa, ok := obj.(*v1.ServiceAccount)
if !ok {
return nil, fmt.Errorf("Failed to cast serviceaccount: %v", obj)
}
return sa, nil
} | [
"func",
"(",
"self",
"*",
"realKubeFramework",
")",
"ParseServiceAccount",
"(",
"filePath",
"string",
")",
"(",
"*",
"v1",
".",
"ServiceAccount",
",",
"error",
")",
"{",
"obj",
",",
"err",
":=",
"self",
".",
"loadObject",
"(",
"filePath",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n\n",
"sa",
",",
"ok",
":=",
"obj",
".",
"(",
"*",
"v1",
".",
"ServiceAccount",
")",
"\n",
"if",
"!",
"ok",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"obj",
")",
"\n",
"}",
"\n",
"return",
"sa",
",",
"nil",
"\n",
"}"
] | // Parses and Returns a ServiceAccount object contained in 'filePath' | [
"Parses",
"and",
"Returns",
"a",
"ServiceAccount",
"object",
"contained",
"in",
"filePath"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/integration/framework.go#L422-L433 | train |
kubernetes-retired/heapster | integration/framework.go | CreateServiceAccount | func (self *realKubeFramework) CreateServiceAccount(sa *v1.ServiceAccount) error {
_, err := self.kubeClient.CoreV1().ServiceAccounts(sa.Namespace).Create(sa)
return err
} | go | func (self *realKubeFramework) CreateServiceAccount(sa *v1.ServiceAccount) error {
_, err := self.kubeClient.CoreV1().ServiceAccounts(sa.Namespace).Create(sa)
return err
} | [
"func",
"(",
"self",
"*",
"realKubeFramework",
")",
"CreateServiceAccount",
"(",
"sa",
"*",
"v1",
".",
"ServiceAccount",
")",
"error",
"{",
"_",
",",
"err",
":=",
"self",
".",
"kubeClient",
".",
"CoreV1",
"(",
")",
".",
"ServiceAccounts",
"(",
"sa",
".",
"Namespace",
")",
".",
"Create",
"(",
"sa",
")",
"\n",
"return",
"err",
"\n",
"}"
] | // CreateServiceAccount creates the ServiceAccount object | [
"CreateServiceAccount",
"creates",
"the",
"ServiceAccount",
"object"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/integration/framework.go#L436-L439 | train |
kubernetes-retired/heapster | common/honeycomb/honeycomb.go | SendBatch | func (c *HoneycombClient) SendBatch(batch Batch) error {
if len(batch) == 0 {
// Nothing to send
return nil
}
errs := []string{}
for i := 0; i < len(batch); i += maxBatchSize {
offset := i + maxBatchSize
if offset > len(batch) {
offset = len(batch)
}
if err := c.sendBatch(batch[i:offset]); err != nil {
errs = append(errs, err.Error())
}
}
if len(errs) > 0 {
return errors.New(strings.Join(errs, "\n"))
}
return nil
} | go | func (c *HoneycombClient) SendBatch(batch Batch) error {
if len(batch) == 0 {
// Nothing to send
return nil
}
errs := []string{}
for i := 0; i < len(batch); i += maxBatchSize {
offset := i + maxBatchSize
if offset > len(batch) {
offset = len(batch)
}
if err := c.sendBatch(batch[i:offset]); err != nil {
errs = append(errs, err.Error())
}
}
if len(errs) > 0 {
return errors.New(strings.Join(errs, "\n"))
}
return nil
} | [
"func",
"(",
"c",
"*",
"HoneycombClient",
")",
"SendBatch",
"(",
"batch",
"Batch",
")",
"error",
"{",
"if",
"len",
"(",
"batch",
")",
"==",
"0",
"{",
"// Nothing to send",
"return",
"nil",
"\n",
"}",
"\n\n",
"errs",
":=",
"[",
"]",
"string",
"{",
"}",
"\n",
"for",
"i",
":=",
"0",
";",
"i",
"<",
"len",
"(",
"batch",
")",
";",
"i",
"+=",
"maxBatchSize",
"{",
"offset",
":=",
"i",
"+",
"maxBatchSize",
"\n",
"if",
"offset",
">",
"len",
"(",
"batch",
")",
"{",
"offset",
"=",
"len",
"(",
"batch",
")",
"\n",
"}",
"\n",
"if",
"err",
":=",
"c",
".",
"sendBatch",
"(",
"batch",
"[",
"i",
":",
"offset",
"]",
")",
";",
"err",
"!=",
"nil",
"{",
"errs",
"=",
"append",
"(",
"errs",
",",
"err",
".",
"Error",
"(",
")",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"if",
"len",
"(",
"errs",
")",
">",
"0",
"{",
"return",
"errors",
".",
"New",
"(",
"strings",
".",
"Join",
"(",
"errs",
",",
"\"",
"\\n",
"\"",
")",
")",
"\n",
"}",
"\n\n",
"return",
"nil",
"\n",
"}"
] | // SendBatch splits the top-level batch into sub-batches if needed. Otherwise,
// requests that are too large will be rejected by the Honeycomb API. | [
"SendBatch",
"splits",
"the",
"top",
"-",
"level",
"batch",
"into",
"sub",
"-",
"batches",
"if",
"needed",
".",
"Otherwise",
"requests",
"that",
"are",
"too",
"large",
"will",
"be",
"rejected",
"by",
"the",
"Honeycomb",
"API",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/common/honeycomb/honeycomb.go#L108-L130 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | checkSanitizedMetricName | func (sink *influxdbSink) checkSanitizedMetricName(name string) error {
if !metricAllowedChars.MatchString(name) {
return fmt.Errorf("Invalid metric name %q", name)
}
return nil
} | go | func (sink *influxdbSink) checkSanitizedMetricName(name string) error {
if !metricAllowedChars.MatchString(name) {
return fmt.Errorf("Invalid metric name %q", name)
}
return nil
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"checkSanitizedMetricName",
"(",
"name",
"string",
")",
"error",
"{",
"if",
"!",
"metricAllowedChars",
".",
"MatchString",
"(",
"name",
")",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"name",
")",
"\n",
"}",
"\n\n",
"return",
"nil",
"\n",
"}"
] | // checkSanitizedMetricName errors out if invalid characters are found in the metric name, since InfluxDB
// does not widely support bound parameters yet, and we need to sanitize our inputs. | [
"checkSanitizedMetricName",
"errors",
"out",
"if",
"invalid",
"characters",
"are",
"found",
"in",
"the",
"metric",
"name",
"since",
"InfluxDB",
"does",
"not",
"widely",
"support",
"bound",
"parameters",
"yet",
"and",
"we",
"need",
"to",
"sanitize",
"our",
"inputs",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L78-L84 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | checkSanitizedMetricLabels | func (sink *influxdbSink) checkSanitizedMetricLabels(labels map[string]string) error {
// label names have the same restrictions as metric names, here
for k, v := range labels {
if !metricAllowedChars.MatchString(k) {
return fmt.Errorf("Invalid label name %q", k)
}
// for metric values, we're somewhat more permissive. We allow any
// Printable unicode character, except quotation marks, which are used
// to delimit things.
if strings.ContainsRune(v, '"') || strings.ContainsRune(v, '\'') {
return fmt.Errorf("Invalid label value %q", v)
}
for _, runeVal := range v {
if !unicode.IsPrint(runeVal) {
return fmt.Errorf("Invalid label value %q", v)
}
}
}
return nil
} | go | func (sink *influxdbSink) checkSanitizedMetricLabels(labels map[string]string) error {
// label names have the same restrictions as metric names, here
for k, v := range labels {
if !metricAllowedChars.MatchString(k) {
return fmt.Errorf("Invalid label name %q", k)
}
// for metric values, we're somewhat more permissive. We allow any
// Printable unicode character, except quotation marks, which are used
// to delimit things.
if strings.ContainsRune(v, '"') || strings.ContainsRune(v, '\'') {
return fmt.Errorf("Invalid label value %q", v)
}
for _, runeVal := range v {
if !unicode.IsPrint(runeVal) {
return fmt.Errorf("Invalid label value %q", v)
}
}
}
return nil
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"checkSanitizedMetricLabels",
"(",
"labels",
"map",
"[",
"string",
"]",
"string",
")",
"error",
"{",
"// label names have the same restrictions as metric names, here",
"for",
"k",
",",
"v",
":=",
"range",
"labels",
"{",
"if",
"!",
"metricAllowedChars",
".",
"MatchString",
"(",
"k",
")",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"k",
")",
"\n",
"}",
"\n\n",
"// for metric values, we're somewhat more permissive. We allow any",
"// Printable unicode character, except quotation marks, which are used",
"// to delimit things.",
"if",
"strings",
".",
"ContainsRune",
"(",
"v",
",",
"'\"'",
")",
"||",
"strings",
".",
"ContainsRune",
"(",
"v",
",",
"'\\''",
")",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"v",
")",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"runeVal",
":=",
"range",
"v",
"{",
"if",
"!",
"unicode",
".",
"IsPrint",
"(",
"runeVal",
")",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"v",
")",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n\n",
"return",
"nil",
"\n",
"}"
] | // checkSanitizedMetricLabels errors out if invalid characters are found in the label name or label value, since
// InfluxDb does not widely support bound parameters yet, and we need to sanitize our inputs. | [
"checkSanitizedMetricLabels",
"errors",
"out",
"if",
"invalid",
"characters",
"are",
"found",
"in",
"the",
"label",
"name",
"or",
"label",
"value",
"since",
"InfluxDb",
"does",
"not",
"widely",
"support",
"bound",
"parameters",
"yet",
"and",
"we",
"need",
"to",
"sanitize",
"our",
"inputs",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L88-L110 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | aggregationFunc | func (sink *influxdbSink) aggregationFunc(aggregationName core.AggregationType, fieldName string) string {
switch aggregationName {
case core.AggregationTypeAverage:
return fmt.Sprintf("MEAN(%q)", fieldName)
case core.AggregationTypeMaximum:
return fmt.Sprintf("MAX(%q)", fieldName)
case core.AggregationTypeMinimum:
return fmt.Sprintf("MIN(%q)", fieldName)
case core.AggregationTypeMedian:
return fmt.Sprintf("MEDIAN(%q)", fieldName)
case core.AggregationTypeCount:
return fmt.Sprintf("COUNT(%q)", fieldName)
case core.AggregationTypePercentile50:
return fmt.Sprintf("PERCENTILE(%q, 50)", fieldName)
case core.AggregationTypePercentile95:
return fmt.Sprintf("PERCENTILE(%q, 95)", fieldName)
case core.AggregationTypePercentile99:
return fmt.Sprintf("PERCENTILE(%q, 99)", fieldName)
}
// This should have been checked by the API level, so something's seriously wrong here
panic(fmt.Sprintf("Unknown aggregation type %q", aggregationName))
} | go | func (sink *influxdbSink) aggregationFunc(aggregationName core.AggregationType, fieldName string) string {
switch aggregationName {
case core.AggregationTypeAverage:
return fmt.Sprintf("MEAN(%q)", fieldName)
case core.AggregationTypeMaximum:
return fmt.Sprintf("MAX(%q)", fieldName)
case core.AggregationTypeMinimum:
return fmt.Sprintf("MIN(%q)", fieldName)
case core.AggregationTypeMedian:
return fmt.Sprintf("MEDIAN(%q)", fieldName)
case core.AggregationTypeCount:
return fmt.Sprintf("COUNT(%q)", fieldName)
case core.AggregationTypePercentile50:
return fmt.Sprintf("PERCENTILE(%q, 50)", fieldName)
case core.AggregationTypePercentile95:
return fmt.Sprintf("PERCENTILE(%q, 95)", fieldName)
case core.AggregationTypePercentile99:
return fmt.Sprintf("PERCENTILE(%q, 99)", fieldName)
}
// This should have been checked by the API level, so something's seriously wrong here
panic(fmt.Sprintf("Unknown aggregation type %q", aggregationName))
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"aggregationFunc",
"(",
"aggregationName",
"core",
".",
"AggregationType",
",",
"fieldName",
"string",
")",
"string",
"{",
"switch",
"aggregationName",
"{",
"case",
"core",
".",
"AggregationTypeAverage",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
")",
"\n",
"case",
"core",
".",
"AggregationTypeMaximum",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
")",
"\n",
"case",
"core",
".",
"AggregationTypeMinimum",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
")",
"\n",
"case",
"core",
".",
"AggregationTypeMedian",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
")",
"\n",
"case",
"core",
".",
"AggregationTypeCount",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
")",
"\n",
"case",
"core",
".",
"AggregationTypePercentile50",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
")",
"\n",
"case",
"core",
".",
"AggregationTypePercentile95",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
")",
"\n",
"case",
"core",
".",
"AggregationTypePercentile99",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
")",
"\n",
"}",
"\n\n",
"// This should have been checked by the API level, so something's seriously wrong here",
"panic",
"(",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"aggregationName",
")",
")",
"\n",
"}"
] | // aggregationFunc converts an aggregation name into the equivalent call to an InfluxQL
// aggregation function | [
"aggregationFunc",
"converts",
"an",
"aggregation",
"name",
"into",
"the",
"equivalent",
"call",
"to",
"an",
"InfluxQL",
"aggregation",
"function"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L114-L136 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | keyToSelector | func (sink *influxdbSink) keyToSelector(key core.HistoricalKey) string {
typeSel := fmt.Sprintf("type = '%s'", key.ObjectType)
switch key.ObjectType {
case core.MetricSetTypeNode:
return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelNodename.Key, key.NodeName)
case core.MetricSetTypeSystemContainer:
return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelContainerName.Key, key.ContainerName, core.LabelNodename.Key, key.NodeName)
case core.MetricSetTypeCluster:
return typeSel
case core.MetricSetTypeNamespace:
return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName)
case core.MetricSetTypePod:
if key.PodId != "" {
return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelPodId.Key, key.PodId)
} else {
return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName, core.LabelPodName.Key, key.PodName)
}
case core.MetricSetTypePodContainer:
if key.PodId != "" {
return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelPodId.Key, key.PodId, core.LabelContainerName.Key, key.ContainerName)
} else {
return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s' AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName, core.LabelPodName.Key, key.PodName, core.LabelContainerName.Key, key.ContainerName)
}
}
// These are assigned by the API, so it shouldn't be possible to reach this unless things are really broken
panic(fmt.Sprintf("Unknown metric type %q", key.ObjectType))
} | go | func (sink *influxdbSink) keyToSelector(key core.HistoricalKey) string {
typeSel := fmt.Sprintf("type = '%s'", key.ObjectType)
switch key.ObjectType {
case core.MetricSetTypeNode:
return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelNodename.Key, key.NodeName)
case core.MetricSetTypeSystemContainer:
return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelContainerName.Key, key.ContainerName, core.LabelNodename.Key, key.NodeName)
case core.MetricSetTypeCluster:
return typeSel
case core.MetricSetTypeNamespace:
return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName)
case core.MetricSetTypePod:
if key.PodId != "" {
return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelPodId.Key, key.PodId)
} else {
return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName, core.LabelPodName.Key, key.PodName)
}
case core.MetricSetTypePodContainer:
if key.PodId != "" {
return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelPodId.Key, key.PodId, core.LabelContainerName.Key, key.ContainerName)
} else {
return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s' AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName, core.LabelPodName.Key, key.PodName, core.LabelContainerName.Key, key.ContainerName)
}
}
// These are assigned by the API, so it shouldn't be possible to reach this unless things are really broken
panic(fmt.Sprintf("Unknown metric type %q", key.ObjectType))
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"keyToSelector",
"(",
"key",
"core",
".",
"HistoricalKey",
")",
"string",
"{",
"typeSel",
":=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"key",
".",
"ObjectType",
")",
"\n",
"switch",
"key",
".",
"ObjectType",
"{",
"case",
"core",
".",
"MetricSetTypeNode",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"typeSel",
",",
"core",
".",
"LabelNodename",
".",
"Key",
",",
"key",
".",
"NodeName",
")",
"\n",
"case",
"core",
".",
"MetricSetTypeSystemContainer",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"typeSel",
",",
"core",
".",
"LabelContainerName",
".",
"Key",
",",
"key",
".",
"ContainerName",
",",
"core",
".",
"LabelNodename",
".",
"Key",
",",
"key",
".",
"NodeName",
")",
"\n",
"case",
"core",
".",
"MetricSetTypeCluster",
":",
"return",
"typeSel",
"\n",
"case",
"core",
".",
"MetricSetTypeNamespace",
":",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"typeSel",
",",
"core",
".",
"LabelNamespaceName",
".",
"Key",
",",
"key",
".",
"NamespaceName",
")",
"\n",
"case",
"core",
".",
"MetricSetTypePod",
":",
"if",
"key",
".",
"PodId",
"!=",
"\"",
"\"",
"{",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"typeSel",
",",
"core",
".",
"LabelPodId",
".",
"Key",
",",
"key",
".",
"PodId",
")",
"\n",
"}",
"else",
"{",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"typeSel",
",",
"core",
".",
"LabelNamespaceName",
".",
"Key",
",",
"key",
".",
"NamespaceName",
",",
"core",
".",
"LabelPodName",
".",
"Key",
",",
"key",
".",
"PodName",
")",
"\n",
"}",
"\n",
"case",
"core",
".",
"MetricSetTypePodContainer",
":",
"if",
"key",
".",
"PodId",
"!=",
"\"",
"\"",
"{",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"typeSel",
",",
"core",
".",
"LabelPodId",
".",
"Key",
",",
"key",
".",
"PodId",
",",
"core",
".",
"LabelContainerName",
".",
"Key",
",",
"key",
".",
"ContainerName",
")",
"\n",
"}",
"else",
"{",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"typeSel",
",",
"core",
".",
"LabelNamespaceName",
".",
"Key",
",",
"key",
".",
"NamespaceName",
",",
"core",
".",
"LabelPodName",
".",
"Key",
",",
"key",
".",
"PodName",
",",
"core",
".",
"LabelContainerName",
".",
"Key",
",",
"key",
".",
"ContainerName",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"// These are assigned by the API, so it shouldn't be possible to reach this unless things are really broken",
"panic",
"(",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"key",
".",
"ObjectType",
")",
")",
"\n",
"}"
] | // keyToSelector converts a HistoricalKey to an InfluxQL predicate | [
"keyToSelector",
"converts",
"a",
"HistoricalKey",
"to",
"an",
"InfluxQL",
"predicate"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L139-L166 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | labelsToPredicate | func (sink *influxdbSink) labelsToPredicate(labels map[string]string) string {
if len(labels) == 0 {
return ""
}
parts := make([]string, 0, len(labels))
for k, v := range labels {
parts = append(parts, fmt.Sprintf("%q = '%s'", k, v))
}
return strings.Join(parts, " AND ")
} | go | func (sink *influxdbSink) labelsToPredicate(labels map[string]string) string {
if len(labels) == 0 {
return ""
}
parts := make([]string, 0, len(labels))
for k, v := range labels {
parts = append(parts, fmt.Sprintf("%q = '%s'", k, v))
}
return strings.Join(parts, " AND ")
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"labelsToPredicate",
"(",
"labels",
"map",
"[",
"string",
"]",
"string",
")",
"string",
"{",
"if",
"len",
"(",
"labels",
")",
"==",
"0",
"{",
"return",
"\"",
"\"",
"\n",
"}",
"\n\n",
"parts",
":=",
"make",
"(",
"[",
"]",
"string",
",",
"0",
",",
"len",
"(",
"labels",
")",
")",
"\n",
"for",
"k",
",",
"v",
":=",
"range",
"labels",
"{",
"parts",
"=",
"append",
"(",
"parts",
",",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"k",
",",
"v",
")",
")",
"\n",
"}",
"\n\n",
"return",
"strings",
".",
"Join",
"(",
"parts",
",",
"\"",
"\"",
")",
"\n",
"}"
] | // labelsToPredicate composes an InfluxQL predicate based on the given map of labels | [
"labelsToPredicate",
"composes",
"an",
"InfluxQL",
"predicate",
"based",
"on",
"the",
"given",
"map",
"of",
"labels"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L169-L180 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | composeRawQuery | func (sink *influxdbSink) composeRawQuery(metricName string, labels map[string]string, metricKeys []core.HistoricalKey, start, end time.Time) string {
seriesName, fieldName := sink.metricToSeriesAndField(metricName)
queries := make([]string, len(metricKeys))
for i, key := range metricKeys {
pred := sink.keyToSelector(key)
if labels != nil {
pred += fmt.Sprintf(" AND %s", sink.labelsToPredicate(labels))
}
if !start.IsZero() {
pred += fmt.Sprintf(" AND time > '%s'", start.Format(time.RFC3339))
}
if !end.IsZero() {
pred += fmt.Sprintf(" AND time < '%s'", end.Format(time.RFC3339))
}
queries[i] = fmt.Sprintf("SELECT time, %q FROM %q WHERE %s", fieldName, seriesName, pred)
}
return strings.Join(queries, "; ")
} | go | func (sink *influxdbSink) composeRawQuery(metricName string, labels map[string]string, metricKeys []core.HistoricalKey, start, end time.Time) string {
seriesName, fieldName := sink.metricToSeriesAndField(metricName)
queries := make([]string, len(metricKeys))
for i, key := range metricKeys {
pred := sink.keyToSelector(key)
if labels != nil {
pred += fmt.Sprintf(" AND %s", sink.labelsToPredicate(labels))
}
if !start.IsZero() {
pred += fmt.Sprintf(" AND time > '%s'", start.Format(time.RFC3339))
}
if !end.IsZero() {
pred += fmt.Sprintf(" AND time < '%s'", end.Format(time.RFC3339))
}
queries[i] = fmt.Sprintf("SELECT time, %q FROM %q WHERE %s", fieldName, seriesName, pred)
}
return strings.Join(queries, "; ")
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"composeRawQuery",
"(",
"metricName",
"string",
",",
"labels",
"map",
"[",
"string",
"]",
"string",
",",
"metricKeys",
"[",
"]",
"core",
".",
"HistoricalKey",
",",
"start",
",",
"end",
"time",
".",
"Time",
")",
"string",
"{",
"seriesName",
",",
"fieldName",
":=",
"sink",
".",
"metricToSeriesAndField",
"(",
"metricName",
")",
"\n\n",
"queries",
":=",
"make",
"(",
"[",
"]",
"string",
",",
"len",
"(",
"metricKeys",
")",
")",
"\n",
"for",
"i",
",",
"key",
":=",
"range",
"metricKeys",
"{",
"pred",
":=",
"sink",
".",
"keyToSelector",
"(",
"key",
")",
"\n",
"if",
"labels",
"!=",
"nil",
"{",
"pred",
"+=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"sink",
".",
"labelsToPredicate",
"(",
"labels",
")",
")",
"\n",
"}",
"\n",
"if",
"!",
"start",
".",
"IsZero",
"(",
")",
"{",
"pred",
"+=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"start",
".",
"Format",
"(",
"time",
".",
"RFC3339",
")",
")",
"\n",
"}",
"\n",
"if",
"!",
"end",
".",
"IsZero",
"(",
")",
"{",
"pred",
"+=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"end",
".",
"Format",
"(",
"time",
".",
"RFC3339",
")",
")",
"\n",
"}",
"\n",
"queries",
"[",
"i",
"]",
"=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"fieldName",
",",
"seriesName",
",",
"pred",
")",
"\n",
"}",
"\n\n",
"return",
"strings",
".",
"Join",
"(",
"queries",
",",
"\"",
"\"",
")",
"\n",
"}"
] | // composeRawQuery creates the InfluxQL query to fetch the given metric values | [
"composeRawQuery",
"creates",
"the",
"InfluxQL",
"query",
"to",
"fetch",
"the",
"given",
"metric",
"values"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L198-L217 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | parseRawQueryRow | func (sink *influxdbSink) parseRawQueryRow(rawRow influx_models.Row) ([]core.TimestampedMetricValue, error) {
vals := make([]core.TimestampedMetricValue, len(rawRow.Values))
wasInt := make(map[string]bool, 1)
for i, rawVal := range rawRow.Values {
val := core.TimestampedMetricValue{}
if ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil {
return nil, fmt.Errorf("Unable to parse timestamp %q in series %q", rawVal[0].(string), rawRow.Name)
} else {
val.Timestamp = ts
}
if err := tryParseMetricValue("value", rawVal, &val.MetricValue, 1, wasInt); err != nil {
glog.Errorf("Unable to parse field \"value\" in series %q: %v", rawRow.Name, err)
return nil, fmt.Errorf("Unable to parse values in series %q", rawRow.Name)
}
vals[i] = val
}
if wasInt["value"] {
for i := range vals {
vals[i].MetricValue.ValueType = core.ValueInt64
}
} else {
for i := range vals {
vals[i].MetricValue.ValueType = core.ValueFloat
}
}
return vals, nil
} | go | func (sink *influxdbSink) parseRawQueryRow(rawRow influx_models.Row) ([]core.TimestampedMetricValue, error) {
vals := make([]core.TimestampedMetricValue, len(rawRow.Values))
wasInt := make(map[string]bool, 1)
for i, rawVal := range rawRow.Values {
val := core.TimestampedMetricValue{}
if ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil {
return nil, fmt.Errorf("Unable to parse timestamp %q in series %q", rawVal[0].(string), rawRow.Name)
} else {
val.Timestamp = ts
}
if err := tryParseMetricValue("value", rawVal, &val.MetricValue, 1, wasInt); err != nil {
glog.Errorf("Unable to parse field \"value\" in series %q: %v", rawRow.Name, err)
return nil, fmt.Errorf("Unable to parse values in series %q", rawRow.Name)
}
vals[i] = val
}
if wasInt["value"] {
for i := range vals {
vals[i].MetricValue.ValueType = core.ValueInt64
}
} else {
for i := range vals {
vals[i].MetricValue.ValueType = core.ValueFloat
}
}
return vals, nil
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"parseRawQueryRow",
"(",
"rawRow",
"influx_models",
".",
"Row",
")",
"(",
"[",
"]",
"core",
".",
"TimestampedMetricValue",
",",
"error",
")",
"{",
"vals",
":=",
"make",
"(",
"[",
"]",
"core",
".",
"TimestampedMetricValue",
",",
"len",
"(",
"rawRow",
".",
"Values",
")",
")",
"\n",
"wasInt",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"bool",
",",
"1",
")",
"\n",
"for",
"i",
",",
"rawVal",
":=",
"range",
"rawRow",
".",
"Values",
"{",
"val",
":=",
"core",
".",
"TimestampedMetricValue",
"{",
"}",
"\n\n",
"if",
"ts",
",",
"err",
":=",
"time",
".",
"Parse",
"(",
"time",
".",
"RFC3339",
",",
"rawVal",
"[",
"0",
"]",
".",
"(",
"string",
")",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawVal",
"[",
"0",
"]",
".",
"(",
"string",
")",
",",
"rawRow",
".",
"Name",
")",
"\n",
"}",
"else",
"{",
"val",
".",
"Timestamp",
"=",
"ts",
"\n",
"}",
"\n\n",
"if",
"err",
":=",
"tryParseMetricValue",
"(",
"\"",
"\"",
",",
"rawVal",
",",
"&",
"val",
".",
"MetricValue",
",",
"1",
",",
"wasInt",
")",
";",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Errorf",
"(",
"\"",
"\\\"",
"\\\"",
"\"",
",",
"rawRow",
".",
"Name",
",",
"err",
")",
"\n",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawRow",
".",
"Name",
")",
"\n",
"}",
"\n\n",
"vals",
"[",
"i",
"]",
"=",
"val",
"\n",
"}",
"\n\n",
"if",
"wasInt",
"[",
"\"",
"\"",
"]",
"{",
"for",
"i",
":=",
"range",
"vals",
"{",
"vals",
"[",
"i",
"]",
".",
"MetricValue",
".",
"ValueType",
"=",
"core",
".",
"ValueInt64",
"\n",
"}",
"\n",
"}",
"else",
"{",
"for",
"i",
":=",
"range",
"vals",
"{",
"vals",
"[",
"i",
"]",
".",
"MetricValue",
".",
"ValueType",
"=",
"core",
".",
"ValueFloat",
"\n",
"}",
"\n",
"}",
"\n\n",
"return",
"vals",
",",
"nil",
"\n",
"}"
] | // parseRawQueryRow parses a set of timestamped metric values from unstructured JSON output into the
// appropriate Heapster form | [
"parseRawQueryRow",
"parses",
"a",
"set",
"of",
"timestamped",
"metric",
"values",
"from",
"unstructured",
"JSON",
"output",
"into",
"the",
"appropriate",
"Heapster",
"form"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L221-L252 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | composeAggregateQuery | func (sink *influxdbSink) composeAggregateQuery(metricName string, labels map[string]string, aggregations []core.AggregationType, metricKeys []core.HistoricalKey, start, end time.Time, bucketSize time.Duration) string {
seriesName, fieldName := sink.metricToSeriesAndField(metricName)
var bucketSizeNanoSeconds int64 = 0
if bucketSize != 0 {
bucketSizeNanoSeconds = int64(bucketSize.Nanoseconds() / int64(time.Microsecond/time.Nanosecond))
}
queries := make([]string, len(metricKeys))
for i, key := range metricKeys {
pred := sink.keyToSelector(key)
if labels != nil {
pred += fmt.Sprintf(" AND %s", sink.labelsToPredicate(labels))
}
if !start.IsZero() {
pred += fmt.Sprintf(" AND time > '%s'", start.Format(time.RFC3339))
}
if !end.IsZero() {
pred += fmt.Sprintf(" AND time < '%s'", end.Format(time.RFC3339))
}
aggParts := make([]string, len(aggregations))
for i, agg := range aggregations {
aggParts[i] = sink.aggregationFunc(agg, fieldName)
}
queries[i] = fmt.Sprintf("SELECT %s FROM %q WHERE %s", strings.Join(aggParts, ", "), seriesName, pred)
if bucketSize != 0 {
// group by time requires we have at least one time bound
if start.IsZero() && end.IsZero() {
queries[i] += fmt.Sprintf(" AND time < now()")
}
// fill(none) makes sure we skip data points will null values (otherwise we'll get a *bunch* of null
// values when we go back beyond the time where we started collecting data).
queries[i] += fmt.Sprintf(" GROUP BY time(%vu) fill(none)", bucketSizeNanoSeconds)
}
}
return strings.Join(queries, "; ")
} | go | func (sink *influxdbSink) composeAggregateQuery(metricName string, labels map[string]string, aggregations []core.AggregationType, metricKeys []core.HistoricalKey, start, end time.Time, bucketSize time.Duration) string {
seriesName, fieldName := sink.metricToSeriesAndField(metricName)
var bucketSizeNanoSeconds int64 = 0
if bucketSize != 0 {
bucketSizeNanoSeconds = int64(bucketSize.Nanoseconds() / int64(time.Microsecond/time.Nanosecond))
}
queries := make([]string, len(metricKeys))
for i, key := range metricKeys {
pred := sink.keyToSelector(key)
if labels != nil {
pred += fmt.Sprintf(" AND %s", sink.labelsToPredicate(labels))
}
if !start.IsZero() {
pred += fmt.Sprintf(" AND time > '%s'", start.Format(time.RFC3339))
}
if !end.IsZero() {
pred += fmt.Sprintf(" AND time < '%s'", end.Format(time.RFC3339))
}
aggParts := make([]string, len(aggregations))
for i, agg := range aggregations {
aggParts[i] = sink.aggregationFunc(agg, fieldName)
}
queries[i] = fmt.Sprintf("SELECT %s FROM %q WHERE %s", strings.Join(aggParts, ", "), seriesName, pred)
if bucketSize != 0 {
// group by time requires we have at least one time bound
if start.IsZero() && end.IsZero() {
queries[i] += fmt.Sprintf(" AND time < now()")
}
// fill(none) makes sure we skip data points will null values (otherwise we'll get a *bunch* of null
// values when we go back beyond the time where we started collecting data).
queries[i] += fmt.Sprintf(" GROUP BY time(%vu) fill(none)", bucketSizeNanoSeconds)
}
}
return strings.Join(queries, "; ")
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"composeAggregateQuery",
"(",
"metricName",
"string",
",",
"labels",
"map",
"[",
"string",
"]",
"string",
",",
"aggregations",
"[",
"]",
"core",
".",
"AggregationType",
",",
"metricKeys",
"[",
"]",
"core",
".",
"HistoricalKey",
",",
"start",
",",
"end",
"time",
".",
"Time",
",",
"bucketSize",
"time",
".",
"Duration",
")",
"string",
"{",
"seriesName",
",",
"fieldName",
":=",
"sink",
".",
"metricToSeriesAndField",
"(",
"metricName",
")",
"\n\n",
"var",
"bucketSizeNanoSeconds",
"int64",
"=",
"0",
"\n",
"if",
"bucketSize",
"!=",
"0",
"{",
"bucketSizeNanoSeconds",
"=",
"int64",
"(",
"bucketSize",
".",
"Nanoseconds",
"(",
")",
"/",
"int64",
"(",
"time",
".",
"Microsecond",
"/",
"time",
".",
"Nanosecond",
")",
")",
"\n",
"}",
"\n\n",
"queries",
":=",
"make",
"(",
"[",
"]",
"string",
",",
"len",
"(",
"metricKeys",
")",
")",
"\n",
"for",
"i",
",",
"key",
":=",
"range",
"metricKeys",
"{",
"pred",
":=",
"sink",
".",
"keyToSelector",
"(",
"key",
")",
"\n",
"if",
"labels",
"!=",
"nil",
"{",
"pred",
"+=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"sink",
".",
"labelsToPredicate",
"(",
"labels",
")",
")",
"\n",
"}",
"\n",
"if",
"!",
"start",
".",
"IsZero",
"(",
")",
"{",
"pred",
"+=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"start",
".",
"Format",
"(",
"time",
".",
"RFC3339",
")",
")",
"\n",
"}",
"\n",
"if",
"!",
"end",
".",
"IsZero",
"(",
")",
"{",
"pred",
"+=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"end",
".",
"Format",
"(",
"time",
".",
"RFC3339",
")",
")",
"\n",
"}",
"\n\n",
"aggParts",
":=",
"make",
"(",
"[",
"]",
"string",
",",
"len",
"(",
"aggregations",
")",
")",
"\n",
"for",
"i",
",",
"agg",
":=",
"range",
"aggregations",
"{",
"aggParts",
"[",
"i",
"]",
"=",
"sink",
".",
"aggregationFunc",
"(",
"agg",
",",
"fieldName",
")",
"\n",
"}",
"\n\n",
"queries",
"[",
"i",
"]",
"=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"strings",
".",
"Join",
"(",
"aggParts",
",",
"\"",
"\"",
")",
",",
"seriesName",
",",
"pred",
")",
"\n\n",
"if",
"bucketSize",
"!=",
"0",
"{",
"// group by time requires we have at least one time bound",
"if",
"start",
".",
"IsZero",
"(",
")",
"&&",
"end",
".",
"IsZero",
"(",
")",
"{",
"queries",
"[",
"i",
"]",
"+=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n\n",
"// fill(none) makes sure we skip data points will null values (otherwise we'll get a *bunch* of null",
"// values when we go back beyond the time where we started collecting data).",
"queries",
"[",
"i",
"]",
"+=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"bucketSizeNanoSeconds",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"return",
"strings",
".",
"Join",
"(",
"queries",
",",
"\"",
"\"",
")",
"\n",
"}"
] | // composeAggregateQuery creates the InfluxQL query to fetch the given aggregation values | [
"composeAggregateQuery",
"creates",
"the",
"InfluxQL",
"query",
"to",
"fetch",
"the",
"given",
"aggregation",
"values"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L337-L378 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | parseAggregateQueryRow | func (sink *influxdbSink) parseAggregateQueryRow(rawRow influx_models.Row, aggregationLookup map[core.AggregationType]int, bucketSize time.Duration) ([]core.TimestampedAggregationValue, error) {
vals := make([]core.TimestampedAggregationValue, len(rawRow.Values))
wasInt := make(map[string]bool, len(aggregationLookup))
for i, rawVal := range rawRow.Values {
val := core.TimestampedAggregationValue{
BucketSize: bucketSize,
AggregationValue: core.AggregationValue{
Aggregations: map[core.AggregationType]core.MetricValue{},
},
}
if ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil {
return nil, fmt.Errorf("Unable to parse timestamp %q in series %q", rawVal[0].(string), rawRow.Name)
} else {
val.Timestamp = ts
}
// The Influx client decods numeric fields to json.Number (a string), so we have to try decoding to both types of numbers
// Count is always a uint64
if countIndex, ok := aggregationLookup[core.AggregationTypeCount]; ok {
if err := json.Unmarshal([]byte(rawVal[countIndex].(json.Number).String()), &val.Count); err != nil {
glog.Errorf("Unable to parse count value in series %q: %v", rawRow.Name, err)
return nil, fmt.Errorf("Unable to parse values in series %q", rawRow.Name)
}
}
// The rest of the aggregation values can be either float or int, so attempt to parse both
if err := populateAggregations(rawRow.Name, rawVal, &val, aggregationLookup, wasInt); err != nil {
return nil, err
}
vals[i] = val
}
// figure out whether each aggregation was full of float values, or int values
setAggregationValueTypes(vals, wasInt)
return vals, nil
} | go | func (sink *influxdbSink) parseAggregateQueryRow(rawRow influx_models.Row, aggregationLookup map[core.AggregationType]int, bucketSize time.Duration) ([]core.TimestampedAggregationValue, error) {
vals := make([]core.TimestampedAggregationValue, len(rawRow.Values))
wasInt := make(map[string]bool, len(aggregationLookup))
for i, rawVal := range rawRow.Values {
val := core.TimestampedAggregationValue{
BucketSize: bucketSize,
AggregationValue: core.AggregationValue{
Aggregations: map[core.AggregationType]core.MetricValue{},
},
}
if ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil {
return nil, fmt.Errorf("Unable to parse timestamp %q in series %q", rawVal[0].(string), rawRow.Name)
} else {
val.Timestamp = ts
}
// The Influx client decods numeric fields to json.Number (a string), so we have to try decoding to both types of numbers
// Count is always a uint64
if countIndex, ok := aggregationLookup[core.AggregationTypeCount]; ok {
if err := json.Unmarshal([]byte(rawVal[countIndex].(json.Number).String()), &val.Count); err != nil {
glog.Errorf("Unable to parse count value in series %q: %v", rawRow.Name, err)
return nil, fmt.Errorf("Unable to parse values in series %q", rawRow.Name)
}
}
// The rest of the aggregation values can be either float or int, so attempt to parse both
if err := populateAggregations(rawRow.Name, rawVal, &val, aggregationLookup, wasInt); err != nil {
return nil, err
}
vals[i] = val
}
// figure out whether each aggregation was full of float values, or int values
setAggregationValueTypes(vals, wasInt)
return vals, nil
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"parseAggregateQueryRow",
"(",
"rawRow",
"influx_models",
".",
"Row",
",",
"aggregationLookup",
"map",
"[",
"core",
".",
"AggregationType",
"]",
"int",
",",
"bucketSize",
"time",
".",
"Duration",
")",
"(",
"[",
"]",
"core",
".",
"TimestampedAggregationValue",
",",
"error",
")",
"{",
"vals",
":=",
"make",
"(",
"[",
"]",
"core",
".",
"TimestampedAggregationValue",
",",
"len",
"(",
"rawRow",
".",
"Values",
")",
")",
"\n",
"wasInt",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"bool",
",",
"len",
"(",
"aggregationLookup",
")",
")",
"\n\n",
"for",
"i",
",",
"rawVal",
":=",
"range",
"rawRow",
".",
"Values",
"{",
"val",
":=",
"core",
".",
"TimestampedAggregationValue",
"{",
"BucketSize",
":",
"bucketSize",
",",
"AggregationValue",
":",
"core",
".",
"AggregationValue",
"{",
"Aggregations",
":",
"map",
"[",
"core",
".",
"AggregationType",
"]",
"core",
".",
"MetricValue",
"{",
"}",
",",
"}",
",",
"}",
"\n\n",
"if",
"ts",
",",
"err",
":=",
"time",
".",
"Parse",
"(",
"time",
".",
"RFC3339",
",",
"rawVal",
"[",
"0",
"]",
".",
"(",
"string",
")",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawVal",
"[",
"0",
"]",
".",
"(",
"string",
")",
",",
"rawRow",
".",
"Name",
")",
"\n",
"}",
"else",
"{",
"val",
".",
"Timestamp",
"=",
"ts",
"\n",
"}",
"\n\n",
"// The Influx client decods numeric fields to json.Number (a string), so we have to try decoding to both types of numbers",
"// Count is always a uint64",
"if",
"countIndex",
",",
"ok",
":=",
"aggregationLookup",
"[",
"core",
".",
"AggregationTypeCount",
"]",
";",
"ok",
"{",
"if",
"err",
":=",
"json",
".",
"Unmarshal",
"(",
"[",
"]",
"byte",
"(",
"rawVal",
"[",
"countIndex",
"]",
".",
"(",
"json",
".",
"Number",
")",
".",
"String",
"(",
")",
")",
",",
"&",
"val",
".",
"Count",
")",
";",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawRow",
".",
"Name",
",",
"err",
")",
"\n",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawRow",
".",
"Name",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"// The rest of the aggregation values can be either float or int, so attempt to parse both",
"if",
"err",
":=",
"populateAggregations",
"(",
"rawRow",
".",
"Name",
",",
"rawVal",
",",
"&",
"val",
",",
"aggregationLookup",
",",
"wasInt",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n\n",
"vals",
"[",
"i",
"]",
"=",
"val",
"\n",
"}",
"\n\n",
"// figure out whether each aggregation was full of float values, or int values",
"setAggregationValueTypes",
"(",
"vals",
",",
"wasInt",
")",
"\n\n",
"return",
"vals",
",",
"nil",
"\n",
"}"
] | // parseRawQueryRow parses a set of timestamped aggregation values from unstructured JSON output into the
// appropriate Heapster form | [
"parseRawQueryRow",
"parses",
"a",
"set",
"of",
"timestamped",
"aggregation",
"values",
"from",
"unstructured",
"JSON",
"output",
"into",
"the",
"appropriate",
"Heapster",
"form"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L382-L422 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | setAggregationValueIfPresent | func setAggregationValueIfPresent(aggName core.AggregationType, rawVal []interface{}, aggregations *core.AggregationValue, indexLookup map[core.AggregationType]int, wasInt map[string]bool) error {
if fieldIndex, ok := indexLookup[aggName]; ok {
targetValue := &core.MetricValue{}
if err := tryParseMetricValue(string(aggName), rawVal, targetValue, fieldIndex, wasInt); err != nil {
return err
}
aggregations.Aggregations[aggName] = *targetValue
}
return nil
} | go | func setAggregationValueIfPresent(aggName core.AggregationType, rawVal []interface{}, aggregations *core.AggregationValue, indexLookup map[core.AggregationType]int, wasInt map[string]bool) error {
if fieldIndex, ok := indexLookup[aggName]; ok {
targetValue := &core.MetricValue{}
if err := tryParseMetricValue(string(aggName), rawVal, targetValue, fieldIndex, wasInt); err != nil {
return err
}
aggregations.Aggregations[aggName] = *targetValue
}
return nil
} | [
"func",
"setAggregationValueIfPresent",
"(",
"aggName",
"core",
".",
"AggregationType",
",",
"rawVal",
"[",
"]",
"interface",
"{",
"}",
",",
"aggregations",
"*",
"core",
".",
"AggregationValue",
",",
"indexLookup",
"map",
"[",
"core",
".",
"AggregationType",
"]",
"int",
",",
"wasInt",
"map",
"[",
"string",
"]",
"bool",
")",
"error",
"{",
"if",
"fieldIndex",
",",
"ok",
":=",
"indexLookup",
"[",
"aggName",
"]",
";",
"ok",
"{",
"targetValue",
":=",
"&",
"core",
".",
"MetricValue",
"{",
"}",
"\n",
"if",
"err",
":=",
"tryParseMetricValue",
"(",
"string",
"(",
"aggName",
")",
",",
"rawVal",
",",
"targetValue",
",",
"fieldIndex",
",",
"wasInt",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"aggregations",
".",
"Aggregations",
"[",
"aggName",
"]",
"=",
"*",
"targetValue",
"\n",
"}",
"\n\n",
"return",
"nil",
"\n",
"}"
] | // setAggregationValueIfPresent checks if the given metric value is present in the list of raw values, and if so,
// copies it to the output format | [
"setAggregationValueIfPresent",
"checks",
"if",
"the",
"given",
"metric",
"value",
"is",
"present",
"in",
"the",
"list",
"of",
"raw",
"values",
"and",
"if",
"so",
"copies",
"it",
"to",
"the",
"output",
"format"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L524-L535 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | tryParseMetricValue | func tryParseMetricValue(aggName string, rawVal []interface{}, targetValue *core.MetricValue, fieldIndex int, wasInt map[string]bool) error {
// the Influx client decodes numeric fields to json.Number (a string), so we have to deal with that --
// assume, starting off, that values may be either float or int. Try int until we fail once, and always
// try float. At the end, figure out which is which.
var rv string
if rvN, ok := rawVal[fieldIndex].(json.Number); !ok {
return fmt.Errorf("Value %q of metric %q was not a json.Number", rawVal[fieldIndex], aggName)
} else {
rv = rvN.String()
}
tryInt := false
isInt, triedBefore := wasInt[aggName]
tryInt = isInt || !triedBefore
if tryInt {
if err := json.Unmarshal([]byte(rv), &targetValue.IntValue); err != nil {
wasInt[aggName] = false
} else {
wasInt[aggName] = true
}
}
if err := json.Unmarshal([]byte(rv), &targetValue.FloatValue); err != nil {
return err
}
return nil
} | go | func tryParseMetricValue(aggName string, rawVal []interface{}, targetValue *core.MetricValue, fieldIndex int, wasInt map[string]bool) error {
// the Influx client decodes numeric fields to json.Number (a string), so we have to deal with that --
// assume, starting off, that values may be either float or int. Try int until we fail once, and always
// try float. At the end, figure out which is which.
var rv string
if rvN, ok := rawVal[fieldIndex].(json.Number); !ok {
return fmt.Errorf("Value %q of metric %q was not a json.Number", rawVal[fieldIndex], aggName)
} else {
rv = rvN.String()
}
tryInt := false
isInt, triedBefore := wasInt[aggName]
tryInt = isInt || !triedBefore
if tryInt {
if err := json.Unmarshal([]byte(rv), &targetValue.IntValue); err != nil {
wasInt[aggName] = false
} else {
wasInt[aggName] = true
}
}
if err := json.Unmarshal([]byte(rv), &targetValue.FloatValue); err != nil {
return err
}
return nil
} | [
"func",
"tryParseMetricValue",
"(",
"aggName",
"string",
",",
"rawVal",
"[",
"]",
"interface",
"{",
"}",
",",
"targetValue",
"*",
"core",
".",
"MetricValue",
",",
"fieldIndex",
"int",
",",
"wasInt",
"map",
"[",
"string",
"]",
"bool",
")",
"error",
"{",
"// the Influx client decodes numeric fields to json.Number (a string), so we have to deal with that --",
"// assume, starting off, that values may be either float or int. Try int until we fail once, and always",
"// try float. At the end, figure out which is which.",
"var",
"rv",
"string",
"\n",
"if",
"rvN",
",",
"ok",
":=",
"rawVal",
"[",
"fieldIndex",
"]",
".",
"(",
"json",
".",
"Number",
")",
";",
"!",
"ok",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawVal",
"[",
"fieldIndex",
"]",
",",
"aggName",
")",
"\n",
"}",
"else",
"{",
"rv",
"=",
"rvN",
".",
"String",
"(",
")",
"\n",
"}",
"\n\n",
"tryInt",
":=",
"false",
"\n",
"isInt",
",",
"triedBefore",
":=",
"wasInt",
"[",
"aggName",
"]",
"\n",
"tryInt",
"=",
"isInt",
"||",
"!",
"triedBefore",
"\n\n",
"if",
"tryInt",
"{",
"if",
"err",
":=",
"json",
".",
"Unmarshal",
"(",
"[",
"]",
"byte",
"(",
"rv",
")",
",",
"&",
"targetValue",
".",
"IntValue",
")",
";",
"err",
"!=",
"nil",
"{",
"wasInt",
"[",
"aggName",
"]",
"=",
"false",
"\n",
"}",
"else",
"{",
"wasInt",
"[",
"aggName",
"]",
"=",
"true",
"\n",
"}",
"\n",
"}",
"\n\n",
"if",
"err",
":=",
"json",
".",
"Unmarshal",
"(",
"[",
"]",
"byte",
"(",
"rv",
")",
",",
"&",
"targetValue",
".",
"FloatValue",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"return",
"nil",
"\n",
"}"
] | // tryParseMetricValue attempts to parse a raw metric value into the appropriate go type. | [
"tryParseMetricValue",
"attempts",
"to",
"parse",
"a",
"raw",
"metric",
"value",
"into",
"the",
"appropriate",
"go",
"type",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L538-L567 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | GetMetricNames | func (sink *influxdbSink) GetMetricNames(metricKey core.HistoricalKey) ([]string, error) {
if err := sink.checkSanitizedKey(&metricKey); err != nil {
return nil, err
}
return sink.stringListQuery(fmt.Sprintf("SHOW MEASUREMENTS WHERE %s", sink.keyToSelector(metricKey)), "Unable to list available metrics")
} | go | func (sink *influxdbSink) GetMetricNames(metricKey core.HistoricalKey) ([]string, error) {
if err := sink.checkSanitizedKey(&metricKey); err != nil {
return nil, err
}
return sink.stringListQuery(fmt.Sprintf("SHOW MEASUREMENTS WHERE %s", sink.keyToSelector(metricKey)), "Unable to list available metrics")
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"GetMetricNames",
"(",
"metricKey",
"core",
".",
"HistoricalKey",
")",
"(",
"[",
"]",
"string",
",",
"error",
")",
"{",
"if",
"err",
":=",
"sink",
".",
"checkSanitizedKey",
"(",
"&",
"metricKey",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"return",
"sink",
".",
"stringListQuery",
"(",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"sink",
".",
"keyToSelector",
"(",
"metricKey",
")",
")",
",",
"\"",
"\"",
")",
"\n",
"}"
] | // GetMetricNames retrieves the available metric names for the given object | [
"GetMetricNames",
"retrieves",
"the",
"available",
"metric",
"names",
"for",
"the",
"given",
"object"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L570-L575 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | GetNodes | func (sink *influxdbSink) GetNodes() ([]string, error) {
return sink.stringListQuery(fmt.Sprintf("SHOW TAG VALUES WITH KEY = %s", core.LabelNodename.Key), "Unable to list all nodes")
} | go | func (sink *influxdbSink) GetNodes() ([]string, error) {
return sink.stringListQuery(fmt.Sprintf("SHOW TAG VALUES WITH KEY = %s", core.LabelNodename.Key), "Unable to list all nodes")
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"GetNodes",
"(",
")",
"(",
"[",
"]",
"string",
",",
"error",
")",
"{",
"return",
"sink",
".",
"stringListQuery",
"(",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"core",
".",
"LabelNodename",
".",
"Key",
")",
",",
"\"",
"\"",
")",
"\n",
"}"
] | // GetNodes retrieves the list of nodes in the cluster | [
"GetNodes",
"retrieves",
"the",
"list",
"of",
"nodes",
"in",
"the",
"cluster"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L578-L580 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | GetNamespaces | func (sink *influxdbSink) GetNamespaces() ([]string, error) {
return sink.stringListQuery(fmt.Sprintf("SHOW TAG VALUES WITH KEY = %s", core.LabelNamespaceName.Key), "Unable to list all namespaces")
} | go | func (sink *influxdbSink) GetNamespaces() ([]string, error) {
return sink.stringListQuery(fmt.Sprintf("SHOW TAG VALUES WITH KEY = %s", core.LabelNamespaceName.Key), "Unable to list all namespaces")
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"GetNamespaces",
"(",
")",
"(",
"[",
"]",
"string",
",",
"error",
")",
"{",
"return",
"sink",
".",
"stringListQuery",
"(",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"core",
".",
"LabelNamespaceName",
".",
"Key",
")",
",",
"\"",
"\"",
")",
"\n",
"}"
] | // GetNamespaces retrieves the list of namespaces in the cluster | [
"GetNamespaces",
"retrieves",
"the",
"list",
"of",
"namespaces",
"in",
"the",
"cluster"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L583-L585 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | GetPodsFromNamespace | func (sink *influxdbSink) GetPodsFromNamespace(namespace string) ([]string, error) {
if !nameAllowedChars.MatchString(namespace) {
return nil, fmt.Errorf("Invalid namespace name %q", namespace)
}
// This is a bit difficult for the influx query language, so we cheat a bit here --
// we just get all series for the uptime measurement for pods which match our namespace
// (any measurement should work here, though)
q := fmt.Sprintf("SHOW SERIES FROM %q WHERE %s = '%s' AND type = '%s'", core.MetricUptime.MetricDescriptor.Name, core.LabelNamespaceName.Key, namespace, core.MetricSetTypePod)
return sink.stringListQueryCol(q, core.LabelPodName.Key, fmt.Sprintf("Unable to list pods in namespace %q", namespace))
} | go | func (sink *influxdbSink) GetPodsFromNamespace(namespace string) ([]string, error) {
if !nameAllowedChars.MatchString(namespace) {
return nil, fmt.Errorf("Invalid namespace name %q", namespace)
}
// This is a bit difficult for the influx query language, so we cheat a bit here --
// we just get all series for the uptime measurement for pods which match our namespace
// (any measurement should work here, though)
q := fmt.Sprintf("SHOW SERIES FROM %q WHERE %s = '%s' AND type = '%s'", core.MetricUptime.MetricDescriptor.Name, core.LabelNamespaceName.Key, namespace, core.MetricSetTypePod)
return sink.stringListQueryCol(q, core.LabelPodName.Key, fmt.Sprintf("Unable to list pods in namespace %q", namespace))
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"GetPodsFromNamespace",
"(",
"namespace",
"string",
")",
"(",
"[",
"]",
"string",
",",
"error",
")",
"{",
"if",
"!",
"nameAllowedChars",
".",
"MatchString",
"(",
"namespace",
")",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"namespace",
")",
"\n",
"}",
"\n",
"// This is a bit difficult for the influx query language, so we cheat a bit here --",
"// we just get all series for the uptime measurement for pods which match our namespace",
"// (any measurement should work here, though)",
"q",
":=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"core",
".",
"MetricUptime",
".",
"MetricDescriptor",
".",
"Name",
",",
"core",
".",
"LabelNamespaceName",
".",
"Key",
",",
"namespace",
",",
"core",
".",
"MetricSetTypePod",
")",
"\n",
"return",
"sink",
".",
"stringListQueryCol",
"(",
"q",
",",
"core",
".",
"LabelPodName",
".",
"Key",
",",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"namespace",
")",
")",
"\n",
"}"
] | // GetPodsFromNamespace retrieves the list of pods in a given namespace | [
"GetPodsFromNamespace",
"retrieves",
"the",
"list",
"of",
"pods",
"in",
"a",
"given",
"namespace"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L588-L597 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | GetSystemContainersFromNode | func (sink *influxdbSink) GetSystemContainersFromNode(node string) ([]string, error) {
if !nameAllowedChars.MatchString(node) {
return nil, fmt.Errorf("Invalid node name %q", node)
}
// This is a bit difficult for the influx query language, so we cheat a bit here --
// we just get all series for the uptime measurement for system containers on our node
// (any measurement should work here, though)
q := fmt.Sprintf("SHOW SERIES FROM %q WHERE %s = '%s' AND type = '%s'", core.MetricUptime.MetricDescriptor.Name, core.LabelNodename.Key, node, core.MetricSetTypeSystemContainer)
return sink.stringListQueryCol(q, core.LabelContainerName.Key, fmt.Sprintf("Unable to list system containers on node %q", node))
} | go | func (sink *influxdbSink) GetSystemContainersFromNode(node string) ([]string, error) {
if !nameAllowedChars.MatchString(node) {
return nil, fmt.Errorf("Invalid node name %q", node)
}
// This is a bit difficult for the influx query language, so we cheat a bit here --
// we just get all series for the uptime measurement for system containers on our node
// (any measurement should work here, though)
q := fmt.Sprintf("SHOW SERIES FROM %q WHERE %s = '%s' AND type = '%s'", core.MetricUptime.MetricDescriptor.Name, core.LabelNodename.Key, node, core.MetricSetTypeSystemContainer)
return sink.stringListQueryCol(q, core.LabelContainerName.Key, fmt.Sprintf("Unable to list system containers on node %q", node))
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"GetSystemContainersFromNode",
"(",
"node",
"string",
")",
"(",
"[",
"]",
"string",
",",
"error",
")",
"{",
"if",
"!",
"nameAllowedChars",
".",
"MatchString",
"(",
"node",
")",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"node",
")",
"\n",
"}",
"\n",
"// This is a bit difficult for the influx query language, so we cheat a bit here --",
"// we just get all series for the uptime measurement for system containers on our node",
"// (any measurement should work here, though)",
"q",
":=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"core",
".",
"MetricUptime",
".",
"MetricDescriptor",
".",
"Name",
",",
"core",
".",
"LabelNodename",
".",
"Key",
",",
"node",
",",
"core",
".",
"MetricSetTypeSystemContainer",
")",
"\n",
"return",
"sink",
".",
"stringListQueryCol",
"(",
"q",
",",
"core",
".",
"LabelContainerName",
".",
"Key",
",",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"node",
")",
")",
"\n",
"}"
] | // GetSystemContainersFromNode retrieves the list of free containers for a given node | [
"GetSystemContainersFromNode",
"retrieves",
"the",
"list",
"of",
"free",
"containers",
"for",
"a",
"given",
"node"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L600-L609 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | stringListQueryCol | func (sink *influxdbSink) stringListQueryCol(q, colName string, errStr string) ([]string, error) {
sink.RLock()
defer sink.RUnlock()
resp, err := sink.runQuery(q)
if err != nil {
return nil, fmt.Errorf(errStr)
}
if len(resp[0].Series) < 1 {
return nil, fmt.Errorf(errStr)
}
colInd := -1
for i, col := range resp[0].Series[0].Columns {
if col == colName {
colInd = i
break
}
}
if colInd == -1 {
glog.Errorf("%s: results did not contain the %q column", errStr, core.LabelPodName.Key)
return nil, fmt.Errorf(errStr)
}
res := make([]string, len(resp[0].Series[0].Values))
for i, rv := range resp[0].Series[0].Values {
res[i] = rv[colInd].(string)
}
return res, nil
} | go | func (sink *influxdbSink) stringListQueryCol(q, colName string, errStr string) ([]string, error) {
sink.RLock()
defer sink.RUnlock()
resp, err := sink.runQuery(q)
if err != nil {
return nil, fmt.Errorf(errStr)
}
if len(resp[0].Series) < 1 {
return nil, fmt.Errorf(errStr)
}
colInd := -1
for i, col := range resp[0].Series[0].Columns {
if col == colName {
colInd = i
break
}
}
if colInd == -1 {
glog.Errorf("%s: results did not contain the %q column", errStr, core.LabelPodName.Key)
return nil, fmt.Errorf(errStr)
}
res := make([]string, len(resp[0].Series[0].Values))
for i, rv := range resp[0].Series[0].Values {
res[i] = rv[colInd].(string)
}
return res, nil
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"stringListQueryCol",
"(",
"q",
",",
"colName",
"string",
",",
"errStr",
"string",
")",
"(",
"[",
"]",
"string",
",",
"error",
")",
"{",
"sink",
".",
"RLock",
"(",
")",
"\n",
"defer",
"sink",
".",
"RUnlock",
"(",
")",
"\n\n",
"resp",
",",
"err",
":=",
"sink",
".",
"runQuery",
"(",
"q",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"errStr",
")",
"\n",
"}",
"\n\n",
"if",
"len",
"(",
"resp",
"[",
"0",
"]",
".",
"Series",
")",
"<",
"1",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"errStr",
")",
"\n",
"}",
"\n\n",
"colInd",
":=",
"-",
"1",
"\n",
"for",
"i",
",",
"col",
":=",
"range",
"resp",
"[",
"0",
"]",
".",
"Series",
"[",
"0",
"]",
".",
"Columns",
"{",
"if",
"col",
"==",
"colName",
"{",
"colInd",
"=",
"i",
"\n",
"break",
"\n",
"}",
"\n",
"}",
"\n\n",
"if",
"colInd",
"==",
"-",
"1",
"{",
"glog",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"errStr",
",",
"core",
".",
"LabelPodName",
".",
"Key",
")",
"\n",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"errStr",
")",
"\n",
"}",
"\n\n",
"res",
":=",
"make",
"(",
"[",
"]",
"string",
",",
"len",
"(",
"resp",
"[",
"0",
"]",
".",
"Series",
"[",
"0",
"]",
".",
"Values",
")",
")",
"\n",
"for",
"i",
",",
"rv",
":=",
"range",
"resp",
"[",
"0",
"]",
".",
"Series",
"[",
"0",
"]",
".",
"Values",
"{",
"res",
"[",
"i",
"]",
"=",
"rv",
"[",
"colInd",
"]",
".",
"(",
"string",
")",
"\n",
"}",
"\n",
"return",
"res",
",",
"nil",
"\n",
"}"
] | // stringListQueryCol runs the given query, and returns all results from the given column as a string list | [
"stringListQueryCol",
"runs",
"the",
"given",
"query",
"and",
"returns",
"all",
"results",
"from",
"the",
"given",
"column",
"as",
"a",
"string",
"list"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L612-L643 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | stringListQuery | func (sink *influxdbSink) stringListQuery(q string, errStr string) ([]string, error) {
sink.RLock()
defer sink.RUnlock()
resp, err := sink.runQuery(q)
if err != nil {
return nil, fmt.Errorf(errStr)
}
if len(resp[0].Series) < 1 {
return nil, fmt.Errorf(errStr)
}
res := make([]string, len(resp[0].Series[0].Values))
for i, rv := range resp[0].Series[0].Values {
res[i] = rv[0].(string)
}
return res, nil
} | go | func (sink *influxdbSink) stringListQuery(q string, errStr string) ([]string, error) {
sink.RLock()
defer sink.RUnlock()
resp, err := sink.runQuery(q)
if err != nil {
return nil, fmt.Errorf(errStr)
}
if len(resp[0].Series) < 1 {
return nil, fmt.Errorf(errStr)
}
res := make([]string, len(resp[0].Series[0].Values))
for i, rv := range resp[0].Series[0].Values {
res[i] = rv[0].(string)
}
return res, nil
} | [
"func",
"(",
"sink",
"*",
"influxdbSink",
")",
"stringListQuery",
"(",
"q",
"string",
",",
"errStr",
"string",
")",
"(",
"[",
"]",
"string",
",",
"error",
")",
"{",
"sink",
".",
"RLock",
"(",
")",
"\n",
"defer",
"sink",
".",
"RUnlock",
"(",
")",
"\n\n",
"resp",
",",
"err",
":=",
"sink",
".",
"runQuery",
"(",
"q",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"errStr",
")",
"\n",
"}",
"\n\n",
"if",
"len",
"(",
"resp",
"[",
"0",
"]",
".",
"Series",
")",
"<",
"1",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"errStr",
")",
"\n",
"}",
"\n\n",
"res",
":=",
"make",
"(",
"[",
"]",
"string",
",",
"len",
"(",
"resp",
"[",
"0",
"]",
".",
"Series",
"[",
"0",
"]",
".",
"Values",
")",
")",
"\n",
"for",
"i",
",",
"rv",
":=",
"range",
"resp",
"[",
"0",
"]",
".",
"Series",
"[",
"0",
"]",
".",
"Values",
"{",
"res",
"[",
"i",
"]",
"=",
"rv",
"[",
"0",
"]",
".",
"(",
"string",
")",
"\n",
"}",
"\n",
"return",
"res",
",",
"nil",
"\n",
"}"
] | // stringListQuery runs the given query, and returns all results from the first column as a string list | [
"stringListQuery",
"runs",
"the",
"given",
"query",
"and",
"returns",
"all",
"results",
"from",
"the",
"first",
"column",
"as",
"a",
"string",
"list"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L646-L664 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | populateAggregations | func populateAggregations(rawRowName string, rawVal []interface{}, val *core.TimestampedAggregationValue, aggregationLookup map[core.AggregationType]int, wasInt map[string]bool) error {
for _, aggregation := range core.MultiTypedAggregations {
if err := setAggregationValueIfPresent(aggregation, rawVal, &val.AggregationValue, aggregationLookup, wasInt); err != nil {
glog.Errorf("Unable to parse field %q in series %q: %v", aggregation, rawRowName, err)
return fmt.Errorf("Unable to parse values in series %q", rawRowName)
}
}
return nil
} | go | func populateAggregations(rawRowName string, rawVal []interface{}, val *core.TimestampedAggregationValue, aggregationLookup map[core.AggregationType]int, wasInt map[string]bool) error {
for _, aggregation := range core.MultiTypedAggregations {
if err := setAggregationValueIfPresent(aggregation, rawVal, &val.AggregationValue, aggregationLookup, wasInt); err != nil {
glog.Errorf("Unable to parse field %q in series %q: %v", aggregation, rawRowName, err)
return fmt.Errorf("Unable to parse values in series %q", rawRowName)
}
}
return nil
} | [
"func",
"populateAggregations",
"(",
"rawRowName",
"string",
",",
"rawVal",
"[",
"]",
"interface",
"{",
"}",
",",
"val",
"*",
"core",
".",
"TimestampedAggregationValue",
",",
"aggregationLookup",
"map",
"[",
"core",
".",
"AggregationType",
"]",
"int",
",",
"wasInt",
"map",
"[",
"string",
"]",
"bool",
")",
"error",
"{",
"for",
"_",
",",
"aggregation",
":=",
"range",
"core",
".",
"MultiTypedAggregations",
"{",
"if",
"err",
":=",
"setAggregationValueIfPresent",
"(",
"aggregation",
",",
"rawVal",
",",
"&",
"val",
".",
"AggregationValue",
",",
"aggregationLookup",
",",
"wasInt",
")",
";",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"aggregation",
",",
"rawRowName",
",",
"err",
")",
"\n",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawRowName",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"return",
"nil",
"\n",
"}"
] | // populateAggregations extracts aggregation values from a given data point | [
"populateAggregations",
"extracts",
"aggregation",
"values",
"from",
"a",
"given",
"data",
"point"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L700-L709 | train |
kubernetes-retired/heapster | metrics/sinks/influxdb/influxdb_historical.go | setAggregationValueTypes | func setAggregationValueTypes(vals []core.TimestampedAggregationValue, wasInt map[string]bool) {
for _, aggregation := range core.MultiTypedAggregations {
if isInt, ok := wasInt[string(aggregation)]; ok && isInt {
for i := range vals {
val := vals[i].Aggregations[aggregation]
val.ValueType = core.ValueInt64
vals[i].Aggregations[aggregation] = val
}
} else if ok {
for i := range vals {
val := vals[i].Aggregations[aggregation]
val.ValueType = core.ValueFloat
vals[i].Aggregations[aggregation] = val
}
}
}
} | go | func setAggregationValueTypes(vals []core.TimestampedAggregationValue, wasInt map[string]bool) {
for _, aggregation := range core.MultiTypedAggregations {
if isInt, ok := wasInt[string(aggregation)]; ok && isInt {
for i := range vals {
val := vals[i].Aggregations[aggregation]
val.ValueType = core.ValueInt64
vals[i].Aggregations[aggregation] = val
}
} else if ok {
for i := range vals {
val := vals[i].Aggregations[aggregation]
val.ValueType = core.ValueFloat
vals[i].Aggregations[aggregation] = val
}
}
}
} | [
"func",
"setAggregationValueTypes",
"(",
"vals",
"[",
"]",
"core",
".",
"TimestampedAggregationValue",
",",
"wasInt",
"map",
"[",
"string",
"]",
"bool",
")",
"{",
"for",
"_",
",",
"aggregation",
":=",
"range",
"core",
".",
"MultiTypedAggregations",
"{",
"if",
"isInt",
",",
"ok",
":=",
"wasInt",
"[",
"string",
"(",
"aggregation",
")",
"]",
";",
"ok",
"&&",
"isInt",
"{",
"for",
"i",
":=",
"range",
"vals",
"{",
"val",
":=",
"vals",
"[",
"i",
"]",
".",
"Aggregations",
"[",
"aggregation",
"]",
"\n",
"val",
".",
"ValueType",
"=",
"core",
".",
"ValueInt64",
"\n",
"vals",
"[",
"i",
"]",
".",
"Aggregations",
"[",
"aggregation",
"]",
"=",
"val",
"\n",
"}",
"\n",
"}",
"else",
"if",
"ok",
"{",
"for",
"i",
":=",
"range",
"vals",
"{",
"val",
":=",
"vals",
"[",
"i",
"]",
".",
"Aggregations",
"[",
"aggregation",
"]",
"\n",
"val",
".",
"ValueType",
"=",
"core",
".",
"ValueFloat",
"\n",
"vals",
"[",
"i",
"]",
".",
"Aggregations",
"[",
"aggregation",
"]",
"=",
"val",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"}"
] | // setAggregationValueTypes inspects a set of aggregation values and figures out whether each aggregation value
// returned as a float column, or an int column | [
"setAggregationValueTypes",
"inspects",
"a",
"set",
"of",
"aggregation",
"values",
"and",
"figures",
"out",
"whether",
"each",
"aggregation",
"value",
"returned",
"as",
"a",
"float",
"column",
"or",
"an",
"int",
"column"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L713-L729 | train |
kubernetes-retired/heapster | metrics/sinks/gcm/gcm.go | register | func (sink *gcmSink) register(metrics []core.Metric) error {
sink.Lock()
defer sink.Unlock()
if sink.registered {
return nil
}
for _, metric := range metrics {
metricName := fullMetricName(sink.project, metric.MetricDescriptor.Name)
metricType := fullMetricType(metric.MetricDescriptor.Name)
if _, err := sink.gcmService.Projects.MetricDescriptors.Delete(metricName).Do(); err != nil {
glog.Infof("[GCM] Deleting metric %v failed: %v", metricName, err)
}
labels := make([]*gcm.LabelDescriptor, 0)
// Node autoscaling metrics have special labels.
if core.IsNodeAutoscalingMetric(metric.MetricDescriptor.Name) {
// All and autoscaling. Do not populate for other filters.
if sink.metricFilter != metricsAll &&
sink.metricFilter != metricsOnlyAutoscaling {
continue
}
for _, l := range core.GcmNodeAutoscalingLabels() {
labels = append(labels, &gcm.LabelDescriptor{
Key: l.Key,
Description: l.Description,
})
}
} else {
// Only all.
if sink.metricFilter != metricsAll {
continue
}
for _, l := range core.GcmLabels() {
labels = append(labels, &gcm.LabelDescriptor{
Key: l.Key,
Description: l.Description,
})
}
}
var metricKind string
switch metric.MetricDescriptor.Type {
case core.MetricCumulative:
metricKind = "CUMULATIVE"
case core.MetricGauge:
metricKind = "GAUGE"
case core.MetricDelta:
metricKind = "DELTA"
}
var valueType string
switch metric.MetricDescriptor.ValueType {
case core.ValueInt64:
valueType = "INT64"
case core.ValueFloat:
valueType = "DOUBLE"
}
desc := &gcm.MetricDescriptor{
Name: metricName,
Description: metric.MetricDescriptor.Description,
Labels: labels,
MetricKind: metricKind,
ValueType: valueType,
Type: metricType,
}
if _, err := sink.gcmService.Projects.MetricDescriptors.Create(fullProjectName(sink.project), desc).Do(); err != nil {
glog.Errorf("Metric registration of %v failed: %v", desc.Name, err)
return err
}
}
sink.registered = true
return nil
} | go | func (sink *gcmSink) register(metrics []core.Metric) error {
sink.Lock()
defer sink.Unlock()
if sink.registered {
return nil
}
for _, metric := range metrics {
metricName := fullMetricName(sink.project, metric.MetricDescriptor.Name)
metricType := fullMetricType(metric.MetricDescriptor.Name)
if _, err := sink.gcmService.Projects.MetricDescriptors.Delete(metricName).Do(); err != nil {
glog.Infof("[GCM] Deleting metric %v failed: %v", metricName, err)
}
labels := make([]*gcm.LabelDescriptor, 0)
// Node autoscaling metrics have special labels.
if core.IsNodeAutoscalingMetric(metric.MetricDescriptor.Name) {
// All and autoscaling. Do not populate for other filters.
if sink.metricFilter != metricsAll &&
sink.metricFilter != metricsOnlyAutoscaling {
continue
}
for _, l := range core.GcmNodeAutoscalingLabels() {
labels = append(labels, &gcm.LabelDescriptor{
Key: l.Key,
Description: l.Description,
})
}
} else {
// Only all.
if sink.metricFilter != metricsAll {
continue
}
for _, l := range core.GcmLabels() {
labels = append(labels, &gcm.LabelDescriptor{
Key: l.Key,
Description: l.Description,
})
}
}
var metricKind string
switch metric.MetricDescriptor.Type {
case core.MetricCumulative:
metricKind = "CUMULATIVE"
case core.MetricGauge:
metricKind = "GAUGE"
case core.MetricDelta:
metricKind = "DELTA"
}
var valueType string
switch metric.MetricDescriptor.ValueType {
case core.ValueInt64:
valueType = "INT64"
case core.ValueFloat:
valueType = "DOUBLE"
}
desc := &gcm.MetricDescriptor{
Name: metricName,
Description: metric.MetricDescriptor.Description,
Labels: labels,
MetricKind: metricKind,
ValueType: valueType,
Type: metricType,
}
if _, err := sink.gcmService.Projects.MetricDescriptors.Create(fullProjectName(sink.project), desc).Do(); err != nil {
glog.Errorf("Metric registration of %v failed: %v", desc.Name, err)
return err
}
}
sink.registered = true
return nil
} | [
"func",
"(",
"sink",
"*",
"gcmSink",
")",
"register",
"(",
"metrics",
"[",
"]",
"core",
".",
"Metric",
")",
"error",
"{",
"sink",
".",
"Lock",
"(",
")",
"\n",
"defer",
"sink",
".",
"Unlock",
"(",
")",
"\n",
"if",
"sink",
".",
"registered",
"{",
"return",
"nil",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"metric",
":=",
"range",
"metrics",
"{",
"metricName",
":=",
"fullMetricName",
"(",
"sink",
".",
"project",
",",
"metric",
".",
"MetricDescriptor",
".",
"Name",
")",
"\n",
"metricType",
":=",
"fullMetricType",
"(",
"metric",
".",
"MetricDescriptor",
".",
"Name",
")",
"\n\n",
"if",
"_",
",",
"err",
":=",
"sink",
".",
"gcmService",
".",
"Projects",
".",
"MetricDescriptors",
".",
"Delete",
"(",
"metricName",
")",
".",
"Do",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Infof",
"(",
"\"",
"\"",
",",
"metricName",
",",
"err",
")",
"\n",
"}",
"\n",
"labels",
":=",
"make",
"(",
"[",
"]",
"*",
"gcm",
".",
"LabelDescriptor",
",",
"0",
")",
"\n\n",
"// Node autoscaling metrics have special labels.",
"if",
"core",
".",
"IsNodeAutoscalingMetric",
"(",
"metric",
".",
"MetricDescriptor",
".",
"Name",
")",
"{",
"// All and autoscaling. Do not populate for other filters.",
"if",
"sink",
".",
"metricFilter",
"!=",
"metricsAll",
"&&",
"sink",
".",
"metricFilter",
"!=",
"metricsOnlyAutoscaling",
"{",
"continue",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"l",
":=",
"range",
"core",
".",
"GcmNodeAutoscalingLabels",
"(",
")",
"{",
"labels",
"=",
"append",
"(",
"labels",
",",
"&",
"gcm",
".",
"LabelDescriptor",
"{",
"Key",
":",
"l",
".",
"Key",
",",
"Description",
":",
"l",
".",
"Description",
",",
"}",
")",
"\n",
"}",
"\n",
"}",
"else",
"{",
"// Only all.",
"if",
"sink",
".",
"metricFilter",
"!=",
"metricsAll",
"{",
"continue",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"l",
":=",
"range",
"core",
".",
"GcmLabels",
"(",
")",
"{",
"labels",
"=",
"append",
"(",
"labels",
",",
"&",
"gcm",
".",
"LabelDescriptor",
"{",
"Key",
":",
"l",
".",
"Key",
",",
"Description",
":",
"l",
".",
"Description",
",",
"}",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"var",
"metricKind",
"string",
"\n\n",
"switch",
"metric",
".",
"MetricDescriptor",
".",
"Type",
"{",
"case",
"core",
".",
"MetricCumulative",
":",
"metricKind",
"=",
"\"",
"\"",
"\n",
"case",
"core",
".",
"MetricGauge",
":",
"metricKind",
"=",
"\"",
"\"",
"\n",
"case",
"core",
".",
"MetricDelta",
":",
"metricKind",
"=",
"\"",
"\"",
"\n",
"}",
"\n\n",
"var",
"valueType",
"string",
"\n\n",
"switch",
"metric",
".",
"MetricDescriptor",
".",
"ValueType",
"{",
"case",
"core",
".",
"ValueInt64",
":",
"valueType",
"=",
"\"",
"\"",
"\n",
"case",
"core",
".",
"ValueFloat",
":",
"valueType",
"=",
"\"",
"\"",
"\n",
"}",
"\n\n",
"desc",
":=",
"&",
"gcm",
".",
"MetricDescriptor",
"{",
"Name",
":",
"metricName",
",",
"Description",
":",
"metric",
".",
"MetricDescriptor",
".",
"Description",
",",
"Labels",
":",
"labels",
",",
"MetricKind",
":",
"metricKind",
",",
"ValueType",
":",
"valueType",
",",
"Type",
":",
"metricType",
",",
"}",
"\n\n",
"if",
"_",
",",
"err",
":=",
"sink",
".",
"gcmService",
".",
"Projects",
".",
"MetricDescriptors",
".",
"Create",
"(",
"fullProjectName",
"(",
"sink",
".",
"project",
")",
",",
"desc",
")",
".",
"Do",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"desc",
".",
"Name",
",",
"err",
")",
"\n",
"return",
"err",
"\n",
"}",
"\n",
"}",
"\n",
"sink",
".",
"registered",
"=",
"true",
"\n",
"return",
"nil",
"\n",
"}"
] | // Adds the specified metrics or updates them if they already exist. | [
"Adds",
"the",
"specified",
"metrics",
"or",
"updates",
"them",
"if",
"they",
"already",
"exist",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/gcm/gcm.go#L217-L297 | train |
kubernetes-retired/heapster | metrics/sinks/manager.go | ExportData | func (this *sinkManager) ExportData(data *core.DataBatch) {
var wg sync.WaitGroup
for _, sh := range this.sinkHolders {
wg.Add(1)
go func(sh sinkHolder, wg *sync.WaitGroup) {
defer wg.Done()
glog.V(2).Infof("Pushing data to: %s", sh.sink.Name())
select {
case sh.dataBatchChannel <- data:
glog.V(2).Infof("Data push completed: %s", sh.sink.Name())
// everything ok
case <-time.After(this.exportDataTimeout):
glog.Warningf("Failed to push data to sink: %s", sh.sink.Name())
}
}(sh, &wg)
}
// Wait for all pushes to complete or timeout.
wg.Wait()
} | go | func (this *sinkManager) ExportData(data *core.DataBatch) {
var wg sync.WaitGroup
for _, sh := range this.sinkHolders {
wg.Add(1)
go func(sh sinkHolder, wg *sync.WaitGroup) {
defer wg.Done()
glog.V(2).Infof("Pushing data to: %s", sh.sink.Name())
select {
case sh.dataBatchChannel <- data:
glog.V(2).Infof("Data push completed: %s", sh.sink.Name())
// everything ok
case <-time.After(this.exportDataTimeout):
glog.Warningf("Failed to push data to sink: %s", sh.sink.Name())
}
}(sh, &wg)
}
// Wait for all pushes to complete or timeout.
wg.Wait()
} | [
"func",
"(",
"this",
"*",
"sinkManager",
")",
"ExportData",
"(",
"data",
"*",
"core",
".",
"DataBatch",
")",
"{",
"var",
"wg",
"sync",
".",
"WaitGroup",
"\n",
"for",
"_",
",",
"sh",
":=",
"range",
"this",
".",
"sinkHolders",
"{",
"wg",
".",
"Add",
"(",
"1",
")",
"\n",
"go",
"func",
"(",
"sh",
"sinkHolder",
",",
"wg",
"*",
"sync",
".",
"WaitGroup",
")",
"{",
"defer",
"wg",
".",
"Done",
"(",
")",
"\n",
"glog",
".",
"V",
"(",
"2",
")",
".",
"Infof",
"(",
"\"",
"\"",
",",
"sh",
".",
"sink",
".",
"Name",
"(",
")",
")",
"\n",
"select",
"{",
"case",
"sh",
".",
"dataBatchChannel",
"<-",
"data",
":",
"glog",
".",
"V",
"(",
"2",
")",
".",
"Infof",
"(",
"\"",
"\"",
",",
"sh",
".",
"sink",
".",
"Name",
"(",
")",
")",
"\n",
"// everything ok",
"case",
"<-",
"time",
".",
"After",
"(",
"this",
".",
"exportDataTimeout",
")",
":",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
",",
"sh",
".",
"sink",
".",
"Name",
"(",
")",
")",
"\n",
"}",
"\n",
"}",
"(",
"sh",
",",
"&",
"wg",
")",
"\n",
"}",
"\n",
"// Wait for all pushes to complete or timeout.",
"wg",
".",
"Wait",
"(",
")",
"\n",
"}"
] | // Guarantees that the export will complete in sinkExportDataTimeout. | [
"Guarantees",
"that",
"the",
"export",
"will",
"complete",
"in",
"sinkExportDataTimeout",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/manager.go#L106-L124 | train |
kubernetes-retired/heapster | metrics/sources/kubelet/kubelet_client.go | GetAllRawContainers | func (self *KubeletClient) GetAllRawContainers(host Host, start, end time.Time) ([]cadvisor.ContainerInfo, error) {
url := self.getUrl(host, "/stats/container/")
return self.getAllContainers(url, start, end)
} | go | func (self *KubeletClient) GetAllRawContainers(host Host, start, end time.Time) ([]cadvisor.ContainerInfo, error) {
url := self.getUrl(host, "/stats/container/")
return self.getAllContainers(url, start, end)
} | [
"func",
"(",
"self",
"*",
"KubeletClient",
")",
"GetAllRawContainers",
"(",
"host",
"Host",
",",
"start",
",",
"end",
"time",
".",
"Time",
")",
"(",
"[",
"]",
"cadvisor",
".",
"ContainerInfo",
",",
"error",
")",
"{",
"url",
":=",
"self",
".",
"getUrl",
"(",
"host",
",",
"\"",
"\"",
")",
"\n\n",
"return",
"self",
".",
"getAllContainers",
"(",
"url",
",",
"start",
",",
"end",
")",
"\n",
"}"
] | // Get stats for all non-Kubernetes containers. | [
"Get",
"stats",
"for",
"all",
"non",
"-",
"Kubernetes",
"containers",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/kubelet/kubelet_client.go#L152-L156 | train |
kubernetes-retired/heapster | metrics/util/metrics/http.go | InstrumentRouteFunc | func InstrumentRouteFunc(handlerName string, routeFunc restful.RouteFunction) restful.RouteFunction {
opts := prometheus.SummaryOpts{
Subsystem: "http",
ConstLabels: prometheus.Labels{"handler": handlerName},
}
reqCnt := prometheus.NewCounterVec(
prometheus.CounterOpts{
Subsystem: opts.Subsystem,
Name: "requests_total",
Help: "Total number of HTTP requests made.",
ConstLabels: opts.ConstLabels,
},
instLabels,
)
opts.Name = "request_duration_microseconds"
opts.Help = "The HTTP request latencies in microseconds."
reqDur := prometheus.NewSummary(opts)
opts.Name = "request_size_bytes"
opts.Help = "The HTTP request sizes in bytes."
reqSz := prometheus.NewSummary(opts)
opts.Name = "response_size_bytes"
opts.Help = "The HTTP response sizes in bytes."
resSz := prometheus.NewSummary(opts)
regReqCnt := prometheus.MustRegisterOrGet(reqCnt).(*prometheus.CounterVec)
regReqDur := prometheus.MustRegisterOrGet(reqDur).(prometheus.Summary)
regReqSz := prometheus.MustRegisterOrGet(reqSz).(prometheus.Summary)
regResSz := prometheus.MustRegisterOrGet(resSz).(prometheus.Summary)
return restful.RouteFunction(func(request *restful.Request, response *restful.Response) {
now := time.Now()
delegate := &responseWriterDelegator{ResponseWriter: response.ResponseWriter}
out := make(chan int)
urlLen := 0
if request.Request.URL != nil {
urlLen = len(request.Request.URL.String())
}
go computeApproximateRequestSize(request.Request, out, urlLen)
_, cn := response.ResponseWriter.(http.CloseNotifier)
_, fl := response.ResponseWriter.(http.Flusher)
_, hj := response.ResponseWriter.(http.Hijacker)
_, rf := response.ResponseWriter.(io.ReaderFrom)
var rw http.ResponseWriter
if cn && fl && hj && rf {
rw = &fancyResponseWriterDelegator{delegate}
} else {
rw = delegate
}
response.ResponseWriter = rw
routeFunc(request, response)
elapsed := float64(time.Since(now)) / float64(time.Microsecond)
method := strings.ToLower(request.Request.Method)
code := strconv.Itoa(delegate.status)
regReqCnt.WithLabelValues(method, code).Inc()
regReqDur.Observe(elapsed)
regResSz.Observe(float64(delegate.written))
regReqSz.Observe(float64(<-out))
})
} | go | func InstrumentRouteFunc(handlerName string, routeFunc restful.RouteFunction) restful.RouteFunction {
opts := prometheus.SummaryOpts{
Subsystem: "http",
ConstLabels: prometheus.Labels{"handler": handlerName},
}
reqCnt := prometheus.NewCounterVec(
prometheus.CounterOpts{
Subsystem: opts.Subsystem,
Name: "requests_total",
Help: "Total number of HTTP requests made.",
ConstLabels: opts.ConstLabels,
},
instLabels,
)
opts.Name = "request_duration_microseconds"
opts.Help = "The HTTP request latencies in microseconds."
reqDur := prometheus.NewSummary(opts)
opts.Name = "request_size_bytes"
opts.Help = "The HTTP request sizes in bytes."
reqSz := prometheus.NewSummary(opts)
opts.Name = "response_size_bytes"
opts.Help = "The HTTP response sizes in bytes."
resSz := prometheus.NewSummary(opts)
regReqCnt := prometheus.MustRegisterOrGet(reqCnt).(*prometheus.CounterVec)
regReqDur := prometheus.MustRegisterOrGet(reqDur).(prometheus.Summary)
regReqSz := prometheus.MustRegisterOrGet(reqSz).(prometheus.Summary)
regResSz := prometheus.MustRegisterOrGet(resSz).(prometheus.Summary)
return restful.RouteFunction(func(request *restful.Request, response *restful.Response) {
now := time.Now()
delegate := &responseWriterDelegator{ResponseWriter: response.ResponseWriter}
out := make(chan int)
urlLen := 0
if request.Request.URL != nil {
urlLen = len(request.Request.URL.String())
}
go computeApproximateRequestSize(request.Request, out, urlLen)
_, cn := response.ResponseWriter.(http.CloseNotifier)
_, fl := response.ResponseWriter.(http.Flusher)
_, hj := response.ResponseWriter.(http.Hijacker)
_, rf := response.ResponseWriter.(io.ReaderFrom)
var rw http.ResponseWriter
if cn && fl && hj && rf {
rw = &fancyResponseWriterDelegator{delegate}
} else {
rw = delegate
}
response.ResponseWriter = rw
routeFunc(request, response)
elapsed := float64(time.Since(now)) / float64(time.Microsecond)
method := strings.ToLower(request.Request.Method)
code := strconv.Itoa(delegate.status)
regReqCnt.WithLabelValues(method, code).Inc()
regReqDur.Observe(elapsed)
regResSz.Observe(float64(delegate.written))
regReqSz.Observe(float64(<-out))
})
} | [
"func",
"InstrumentRouteFunc",
"(",
"handlerName",
"string",
",",
"routeFunc",
"restful",
".",
"RouteFunction",
")",
"restful",
".",
"RouteFunction",
"{",
"opts",
":=",
"prometheus",
".",
"SummaryOpts",
"{",
"Subsystem",
":",
"\"",
"\"",
",",
"ConstLabels",
":",
"prometheus",
".",
"Labels",
"{",
"\"",
"\"",
":",
"handlerName",
"}",
",",
"}",
"\n\n",
"reqCnt",
":=",
"prometheus",
".",
"NewCounterVec",
"(",
"prometheus",
".",
"CounterOpts",
"{",
"Subsystem",
":",
"opts",
".",
"Subsystem",
",",
"Name",
":",
"\"",
"\"",
",",
"Help",
":",
"\"",
"\"",
",",
"ConstLabels",
":",
"opts",
".",
"ConstLabels",
",",
"}",
",",
"instLabels",
",",
")",
"\n\n",
"opts",
".",
"Name",
"=",
"\"",
"\"",
"\n",
"opts",
".",
"Help",
"=",
"\"",
"\"",
"\n",
"reqDur",
":=",
"prometheus",
".",
"NewSummary",
"(",
"opts",
")",
"\n\n",
"opts",
".",
"Name",
"=",
"\"",
"\"",
"\n",
"opts",
".",
"Help",
"=",
"\"",
"\"",
"\n",
"reqSz",
":=",
"prometheus",
".",
"NewSummary",
"(",
"opts",
")",
"\n\n",
"opts",
".",
"Name",
"=",
"\"",
"\"",
"\n",
"opts",
".",
"Help",
"=",
"\"",
"\"",
"\n",
"resSz",
":=",
"prometheus",
".",
"NewSummary",
"(",
"opts",
")",
"\n\n",
"regReqCnt",
":=",
"prometheus",
".",
"MustRegisterOrGet",
"(",
"reqCnt",
")",
".",
"(",
"*",
"prometheus",
".",
"CounterVec",
")",
"\n",
"regReqDur",
":=",
"prometheus",
".",
"MustRegisterOrGet",
"(",
"reqDur",
")",
".",
"(",
"prometheus",
".",
"Summary",
")",
"\n",
"regReqSz",
":=",
"prometheus",
".",
"MustRegisterOrGet",
"(",
"reqSz",
")",
".",
"(",
"prometheus",
".",
"Summary",
")",
"\n",
"regResSz",
":=",
"prometheus",
".",
"MustRegisterOrGet",
"(",
"resSz",
")",
".",
"(",
"prometheus",
".",
"Summary",
")",
"\n\n",
"return",
"restful",
".",
"RouteFunction",
"(",
"func",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"now",
":=",
"time",
".",
"Now",
"(",
")",
"\n\n",
"delegate",
":=",
"&",
"responseWriterDelegator",
"{",
"ResponseWriter",
":",
"response",
".",
"ResponseWriter",
"}",
"\n",
"out",
":=",
"make",
"(",
"chan",
"int",
")",
"\n",
"urlLen",
":=",
"0",
"\n",
"if",
"request",
".",
"Request",
".",
"URL",
"!=",
"nil",
"{",
"urlLen",
"=",
"len",
"(",
"request",
".",
"Request",
".",
"URL",
".",
"String",
"(",
")",
")",
"\n",
"}",
"\n",
"go",
"computeApproximateRequestSize",
"(",
"request",
".",
"Request",
",",
"out",
",",
"urlLen",
")",
"\n\n",
"_",
",",
"cn",
":=",
"response",
".",
"ResponseWriter",
".",
"(",
"http",
".",
"CloseNotifier",
")",
"\n",
"_",
",",
"fl",
":=",
"response",
".",
"ResponseWriter",
".",
"(",
"http",
".",
"Flusher",
")",
"\n",
"_",
",",
"hj",
":=",
"response",
".",
"ResponseWriter",
".",
"(",
"http",
".",
"Hijacker",
")",
"\n",
"_",
",",
"rf",
":=",
"response",
".",
"ResponseWriter",
".",
"(",
"io",
".",
"ReaderFrom",
")",
"\n",
"var",
"rw",
"http",
".",
"ResponseWriter",
"\n",
"if",
"cn",
"&&",
"fl",
"&&",
"hj",
"&&",
"rf",
"{",
"rw",
"=",
"&",
"fancyResponseWriterDelegator",
"{",
"delegate",
"}",
"\n",
"}",
"else",
"{",
"rw",
"=",
"delegate",
"\n",
"}",
"\n",
"response",
".",
"ResponseWriter",
"=",
"rw",
"\n\n",
"routeFunc",
"(",
"request",
",",
"response",
")",
"\n\n",
"elapsed",
":=",
"float64",
"(",
"time",
".",
"Since",
"(",
"now",
")",
")",
"/",
"float64",
"(",
"time",
".",
"Microsecond",
")",
"\n\n",
"method",
":=",
"strings",
".",
"ToLower",
"(",
"request",
".",
"Request",
".",
"Method",
")",
"\n",
"code",
":=",
"strconv",
".",
"Itoa",
"(",
"delegate",
".",
"status",
")",
"\n",
"regReqCnt",
".",
"WithLabelValues",
"(",
"method",
",",
"code",
")",
".",
"Inc",
"(",
")",
"\n",
"regReqDur",
".",
"Observe",
"(",
"elapsed",
")",
"\n",
"regResSz",
".",
"Observe",
"(",
"float64",
"(",
"delegate",
".",
"written",
")",
")",
"\n",
"regReqSz",
".",
"Observe",
"(",
"float64",
"(",
"<-",
"out",
")",
")",
"\n",
"}",
")",
"\n",
"}"
] | // InstrumentRouteFunc works like Prometheus' InstrumentHandlerFunc but wraps
// the go-restful RouteFunction instead of a HandlerFunc | [
"InstrumentRouteFunc",
"works",
"like",
"Prometheus",
"InstrumentHandlerFunc",
"but",
"wraps",
"the",
"go",
"-",
"restful",
"RouteFunction",
"instead",
"of",
"a",
"HandlerFunc"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/util/metrics/http.go#L36-L103 | train |
kubernetes-retired/heapster | metrics/core/ms_keys.go | PodContainerKey | func PodContainerKey(namespace, podName, containerName string) string {
return fmt.Sprintf("namespace:%s/pod:%s/container:%s", namespace, podName, containerName)
} | go | func PodContainerKey(namespace, podName, containerName string) string {
return fmt.Sprintf("namespace:%s/pod:%s/container:%s", namespace, podName, containerName)
} | [
"func",
"PodContainerKey",
"(",
"namespace",
",",
"podName",
",",
"containerName",
"string",
")",
"string",
"{",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"namespace",
",",
"podName",
",",
"containerName",
")",
"\n",
"}"
] | // MetricsSet keys are inside of DataBatch. The structure of the returned string is
// an implementation detail and no component should rely on it as it may change
// anytime. It only guaranteed that it is unique for the unique combination of
// passed parameters. | [
"MetricsSet",
"keys",
"are",
"inside",
"of",
"DataBatch",
".",
"The",
"structure",
"of",
"the",
"returned",
"string",
"is",
"an",
"implementation",
"detail",
"and",
"no",
"component",
"should",
"rely",
"on",
"it",
"as",
"it",
"may",
"change",
"anytime",
".",
"It",
"only",
"guaranteed",
"that",
"it",
"is",
"unique",
"for",
"the",
"unique",
"combination",
"of",
"passed",
"parameters",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/core/ms_keys.go#L26-L28 | train |
kubernetes-retired/heapster | metrics/processors/namespace_based_enricher.go | addNamespaceInfo | func (this *NamespaceBasedEnricher) addNamespaceInfo(metricSet *core.MetricSet) {
metricSetType, found := metricSet.Labels[core.LabelMetricSetType.Key]
if !found {
return
}
if metricSetType != core.MetricSetTypePodContainer &&
metricSetType != core.MetricSetTypePod &&
metricSetType != core.MetricSetTypeNamespace {
return
}
namespaceName, found := metricSet.Labels[core.LabelNamespaceName.Key]
if !found {
return
}
nsObj, exists, err := this.store.GetByKey(namespaceName)
if exists && err == nil {
namespace, ok := nsObj.(*kube_api.Namespace)
if ok {
metricSet.Labels[core.LabelPodNamespaceUID.Key] = string(namespace.UID)
} else {
glog.Errorf("Wrong namespace store content")
}
} else if err != nil {
glog.Warningf("Failed to get namespace %s: %v", namespaceName, err)
} else if !exists {
glog.Warningf("Namespace doesn't exist: %s", namespaceName)
}
} | go | func (this *NamespaceBasedEnricher) addNamespaceInfo(metricSet *core.MetricSet) {
metricSetType, found := metricSet.Labels[core.LabelMetricSetType.Key]
if !found {
return
}
if metricSetType != core.MetricSetTypePodContainer &&
metricSetType != core.MetricSetTypePod &&
metricSetType != core.MetricSetTypeNamespace {
return
}
namespaceName, found := metricSet.Labels[core.LabelNamespaceName.Key]
if !found {
return
}
nsObj, exists, err := this.store.GetByKey(namespaceName)
if exists && err == nil {
namespace, ok := nsObj.(*kube_api.Namespace)
if ok {
metricSet.Labels[core.LabelPodNamespaceUID.Key] = string(namespace.UID)
} else {
glog.Errorf("Wrong namespace store content")
}
} else if err != nil {
glog.Warningf("Failed to get namespace %s: %v", namespaceName, err)
} else if !exists {
glog.Warningf("Namespace doesn't exist: %s", namespaceName)
}
} | [
"func",
"(",
"this",
"*",
"NamespaceBasedEnricher",
")",
"addNamespaceInfo",
"(",
"metricSet",
"*",
"core",
".",
"MetricSet",
")",
"{",
"metricSetType",
",",
"found",
":=",
"metricSet",
".",
"Labels",
"[",
"core",
".",
"LabelMetricSetType",
".",
"Key",
"]",
"\n",
"if",
"!",
"found",
"{",
"return",
"\n",
"}",
"\n",
"if",
"metricSetType",
"!=",
"core",
".",
"MetricSetTypePodContainer",
"&&",
"metricSetType",
"!=",
"core",
".",
"MetricSetTypePod",
"&&",
"metricSetType",
"!=",
"core",
".",
"MetricSetTypeNamespace",
"{",
"return",
"\n",
"}",
"\n\n",
"namespaceName",
",",
"found",
":=",
"metricSet",
".",
"Labels",
"[",
"core",
".",
"LabelNamespaceName",
".",
"Key",
"]",
"\n",
"if",
"!",
"found",
"{",
"return",
"\n",
"}",
"\n\n",
"nsObj",
",",
"exists",
",",
"err",
":=",
"this",
".",
"store",
".",
"GetByKey",
"(",
"namespaceName",
")",
"\n",
"if",
"exists",
"&&",
"err",
"==",
"nil",
"{",
"namespace",
",",
"ok",
":=",
"nsObj",
".",
"(",
"*",
"kube_api",
".",
"Namespace",
")",
"\n",
"if",
"ok",
"{",
"metricSet",
".",
"Labels",
"[",
"core",
".",
"LabelPodNamespaceUID",
".",
"Key",
"]",
"=",
"string",
"(",
"namespace",
".",
"UID",
")",
"\n",
"}",
"else",
"{",
"glog",
".",
"Errorf",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"}",
"else",
"if",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
",",
"namespaceName",
",",
"err",
")",
"\n",
"}",
"else",
"if",
"!",
"exists",
"{",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
",",
"namespaceName",
")",
"\n",
"}",
"\n",
"}"
] | // Adds UID to all namespaced elements. | [
"Adds",
"UID",
"to",
"all",
"namespaced",
"elements",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/processors/namespace_based_enricher.go#L49-L78 | train |
kubernetes-retired/heapster | metrics/api/v1/model_handlers.go | availableClusterMetrics | func (a *Api) availableClusterMetrics(request *restful.Request, response *restful.Response) {
a.processMetricNamesRequest(core.ClusterKey(), response)
} | go | func (a *Api) availableClusterMetrics(request *restful.Request, response *restful.Response) {
a.processMetricNamesRequest(core.ClusterKey(), response)
} | [
"func",
"(",
"a",
"*",
"Api",
")",
"availableClusterMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"a",
".",
"processMetricNamesRequest",
"(",
"core",
".",
"ClusterKey",
"(",
")",
",",
"response",
")",
"\n",
"}"
] | // availableMetrics returns a list of available cluster metric names. | [
"availableMetrics",
"returns",
"a",
"list",
"of",
"available",
"cluster",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/model_handlers.go#L283-L285 | train |
kubernetes-retired/heapster | metrics/api/v1/model_handlers.go | availableNodeMetrics | func (a *Api) availableNodeMetrics(request *restful.Request, response *restful.Response) {
a.processMetricNamesRequest(core.NodeKey(request.PathParameter("node-name")), response)
} | go | func (a *Api) availableNodeMetrics(request *restful.Request, response *restful.Response) {
a.processMetricNamesRequest(core.NodeKey(request.PathParameter("node-name")), response)
} | [
"func",
"(",
"a",
"*",
"Api",
")",
"availableNodeMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"a",
".",
"processMetricNamesRequest",
"(",
"core",
".",
"NodeKey",
"(",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
")",
",",
"response",
")",
"\n",
"}"
] | // availableMetrics returns a list of available node metric names. | [
"availableMetrics",
"returns",
"a",
"list",
"of",
"available",
"node",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/model_handlers.go#L288-L290 | train |
kubernetes-retired/heapster | metrics/api/v1/model_handlers.go | availableNamespaceMetrics | func (a *Api) availableNamespaceMetrics(request *restful.Request, response *restful.Response) {
a.processMetricNamesRequest(core.NamespaceKey(request.PathParameter("namespace-name")), response)
} | go | func (a *Api) availableNamespaceMetrics(request *restful.Request, response *restful.Response) {
a.processMetricNamesRequest(core.NamespaceKey(request.PathParameter("namespace-name")), response)
} | [
"func",
"(",
"a",
"*",
"Api",
")",
"availableNamespaceMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"a",
".",
"processMetricNamesRequest",
"(",
"core",
".",
"NamespaceKey",
"(",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
")",
",",
"response",
")",
"\n",
"}"
] | // availableMetrics returns a list of available namespace metric names. | [
"availableMetrics",
"returns",
"a",
"list",
"of",
"available",
"namespace",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/model_handlers.go#L293-L295 | train |
kubernetes-retired/heapster | metrics/api/v1/model_handlers.go | parseTimeParam | func parseTimeParam(queryParam string, defaultValue time.Time) (time.Time, error) {
if queryParam != "" {
reqStamp, err := time.Parse(time.RFC3339, queryParam)
if err != nil {
return time.Time{}, fmt.Errorf("timestamp argument cannot be parsed: %s", err)
}
return reqStamp, nil
}
return defaultValue, nil
} | go | func parseTimeParam(queryParam string, defaultValue time.Time) (time.Time, error) {
if queryParam != "" {
reqStamp, err := time.Parse(time.RFC3339, queryParam)
if err != nil {
return time.Time{}, fmt.Errorf("timestamp argument cannot be parsed: %s", err)
}
return reqStamp, nil
}
return defaultValue, nil
} | [
"func",
"parseTimeParam",
"(",
"queryParam",
"string",
",",
"defaultValue",
"time",
".",
"Time",
")",
"(",
"time",
".",
"Time",
",",
"error",
")",
"{",
"if",
"queryParam",
"!=",
"\"",
"\"",
"{",
"reqStamp",
",",
"err",
":=",
"time",
".",
"Parse",
"(",
"time",
".",
"RFC3339",
",",
"queryParam",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"time",
".",
"Time",
"{",
"}",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"return",
"reqStamp",
",",
"nil",
"\n",
"}",
"\n",
"return",
"defaultValue",
",",
"nil",
"\n",
"}"
] | // parseRequestParam parses a time.Time from a named QueryParam, using the RFC3339 format. | [
"parseRequestParam",
"parses",
"a",
"time",
".",
"Time",
"from",
"a",
"named",
"QueryParam",
"using",
"the",
"RFC3339",
"format",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/model_handlers.go#L429-L438 | train |
kubernetes-retired/heapster | metrics/sinks/riemann/driver.go | appendEvent | func appendEvent(events []riemanngo.Event, sink *RiemannSink, host, name string, value interface{}, labels map[string]string, timestamp int64) []riemanngo.Event {
event := riemanngo.Event{
Time: timestamp,
Service: name,
Host: host,
Description: "",
Attributes: labels,
Metric: value,
Ttl: sink.config.Ttl,
State: sink.config.State,
Tags: sink.config.Tags,
}
// state everywhere
events = append(events, event)
if len(events) >= sink.config.BatchSize {
err := riemannCommon.SendData(sink.client, events)
if err != nil {
glog.Warningf("Error sending events to Riemann: %v", err)
// client will reconnect later
sink.client = nil
}
events = nil
}
return events
} | go | func appendEvent(events []riemanngo.Event, sink *RiemannSink, host, name string, value interface{}, labels map[string]string, timestamp int64) []riemanngo.Event {
event := riemanngo.Event{
Time: timestamp,
Service: name,
Host: host,
Description: "",
Attributes: labels,
Metric: value,
Ttl: sink.config.Ttl,
State: sink.config.State,
Tags: sink.config.Tags,
}
// state everywhere
events = append(events, event)
if len(events) >= sink.config.BatchSize {
err := riemannCommon.SendData(sink.client, events)
if err != nil {
glog.Warningf("Error sending events to Riemann: %v", err)
// client will reconnect later
sink.client = nil
}
events = nil
}
return events
} | [
"func",
"appendEvent",
"(",
"events",
"[",
"]",
"riemanngo",
".",
"Event",
",",
"sink",
"*",
"RiemannSink",
",",
"host",
",",
"name",
"string",
",",
"value",
"interface",
"{",
"}",
",",
"labels",
"map",
"[",
"string",
"]",
"string",
",",
"timestamp",
"int64",
")",
"[",
"]",
"riemanngo",
".",
"Event",
"{",
"event",
":=",
"riemanngo",
".",
"Event",
"{",
"Time",
":",
"timestamp",
",",
"Service",
":",
"name",
",",
"Host",
":",
"host",
",",
"Description",
":",
"\"",
"\"",
",",
"Attributes",
":",
"labels",
",",
"Metric",
":",
"value",
",",
"Ttl",
":",
"sink",
".",
"config",
".",
"Ttl",
",",
"State",
":",
"sink",
".",
"config",
".",
"State",
",",
"Tags",
":",
"sink",
".",
"config",
".",
"Tags",
",",
"}",
"\n",
"// state everywhere",
"events",
"=",
"append",
"(",
"events",
",",
"event",
")",
"\n",
"if",
"len",
"(",
"events",
")",
">=",
"sink",
".",
"config",
".",
"BatchSize",
"{",
"err",
":=",
"riemannCommon",
".",
"SendData",
"(",
"sink",
".",
"client",
",",
"events",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"// client will reconnect later",
"sink",
".",
"client",
"=",
"nil",
"\n",
"}",
"\n",
"events",
"=",
"nil",
"\n",
"}",
"\n",
"return",
"events",
"\n",
"}"
] | // Receives a list of riemanngo.Event, the sink, and parameters.
// Creates a new event using the parameters and the sink config, and add it into the Event list.
// Can send events if events is full
// Return the list. | [
"Receives",
"a",
"list",
"of",
"riemanngo",
".",
"Event",
"the",
"sink",
"and",
"parameters",
".",
"Creates",
"a",
"new",
"event",
"using",
"the",
"parameters",
"and",
"the",
"sink",
"config",
"and",
"add",
"it",
"into",
"the",
"Event",
"list",
".",
"Can",
"send",
"events",
"if",
"events",
"is",
"full",
"Return",
"the",
"list",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/riemann/driver.go#L61-L85 | train |
kubernetes-retired/heapster | metrics/sinks/riemann/driver.go | ExportData | func (sink *RiemannSink) ExportData(dataBatch *core.DataBatch) {
sink.Lock()
defer sink.Unlock()
if sink.client == nil {
// the client could be nil here, so we reconnect
client, err := riemannCommon.GetRiemannClient(sink.config)
if err != nil {
glog.Warningf("Riemann sink not connected: %v", err)
return
}
sink.client = client
}
var events []riemanngo.Event
for _, metricSet := range dataBatch.MetricSets {
host := metricSet.Labels[core.LabelHostname.Key]
for metricName, metricValue := range metricSet.MetricValues {
if value := metricValue.GetValue(); value != nil {
timestamp := dataBatch.Timestamp.Unix()
// creates an event and add it to dataEvent
events = appendEvent(events, sink, host, metricName, value, metricSet.Labels, timestamp)
}
}
for _, metric := range metricSet.LabeledMetrics {
if value := metric.GetValue(); value != nil {
labels := make(map[string]string)
for k, v := range metricSet.Labels {
labels[k] = v
}
for k, v := range metric.Labels {
labels[k] = v
}
timestamp := dataBatch.Timestamp.Unix()
// creates an event and add it to dataEvent
events = appendEvent(events, sink, host, metric.Name, value, labels, timestamp)
}
}
}
// Send events to Riemann if events is not empty
if len(events) > 0 {
err := riemannCommon.SendData(sink.client, events)
if err != nil {
glog.Warningf("Error sending events to Riemann: %v", err)
// client will reconnect later
sink.client = nil
}
}
} | go | func (sink *RiemannSink) ExportData(dataBatch *core.DataBatch) {
sink.Lock()
defer sink.Unlock()
if sink.client == nil {
// the client could be nil here, so we reconnect
client, err := riemannCommon.GetRiemannClient(sink.config)
if err != nil {
glog.Warningf("Riemann sink not connected: %v", err)
return
}
sink.client = client
}
var events []riemanngo.Event
for _, metricSet := range dataBatch.MetricSets {
host := metricSet.Labels[core.LabelHostname.Key]
for metricName, metricValue := range metricSet.MetricValues {
if value := metricValue.GetValue(); value != nil {
timestamp := dataBatch.Timestamp.Unix()
// creates an event and add it to dataEvent
events = appendEvent(events, sink, host, metricName, value, metricSet.Labels, timestamp)
}
}
for _, metric := range metricSet.LabeledMetrics {
if value := metric.GetValue(); value != nil {
labels := make(map[string]string)
for k, v := range metricSet.Labels {
labels[k] = v
}
for k, v := range metric.Labels {
labels[k] = v
}
timestamp := dataBatch.Timestamp.Unix()
// creates an event and add it to dataEvent
events = appendEvent(events, sink, host, metric.Name, value, labels, timestamp)
}
}
}
// Send events to Riemann if events is not empty
if len(events) > 0 {
err := riemannCommon.SendData(sink.client, events)
if err != nil {
glog.Warningf("Error sending events to Riemann: %v", err)
// client will reconnect later
sink.client = nil
}
}
} | [
"func",
"(",
"sink",
"*",
"RiemannSink",
")",
"ExportData",
"(",
"dataBatch",
"*",
"core",
".",
"DataBatch",
")",
"{",
"sink",
".",
"Lock",
"(",
")",
"\n",
"defer",
"sink",
".",
"Unlock",
"(",
")",
"\n\n",
"if",
"sink",
".",
"client",
"==",
"nil",
"{",
"// the client could be nil here, so we reconnect",
"client",
",",
"err",
":=",
"riemannCommon",
".",
"GetRiemannClient",
"(",
"sink",
".",
"config",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n",
"sink",
".",
"client",
"=",
"client",
"\n",
"}",
"\n\n",
"var",
"events",
"[",
"]",
"riemanngo",
".",
"Event",
"\n\n",
"for",
"_",
",",
"metricSet",
":=",
"range",
"dataBatch",
".",
"MetricSets",
"{",
"host",
":=",
"metricSet",
".",
"Labels",
"[",
"core",
".",
"LabelHostname",
".",
"Key",
"]",
"\n",
"for",
"metricName",
",",
"metricValue",
":=",
"range",
"metricSet",
".",
"MetricValues",
"{",
"if",
"value",
":=",
"metricValue",
".",
"GetValue",
"(",
")",
";",
"value",
"!=",
"nil",
"{",
"timestamp",
":=",
"dataBatch",
".",
"Timestamp",
".",
"Unix",
"(",
")",
"\n",
"// creates an event and add it to dataEvent",
"events",
"=",
"appendEvent",
"(",
"events",
",",
"sink",
",",
"host",
",",
"metricName",
",",
"value",
",",
"metricSet",
".",
"Labels",
",",
"timestamp",
")",
"\n",
"}",
"\n",
"}",
"\n",
"for",
"_",
",",
"metric",
":=",
"range",
"metricSet",
".",
"LabeledMetrics",
"{",
"if",
"value",
":=",
"metric",
".",
"GetValue",
"(",
")",
";",
"value",
"!=",
"nil",
"{",
"labels",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"string",
")",
"\n",
"for",
"k",
",",
"v",
":=",
"range",
"metricSet",
".",
"Labels",
"{",
"labels",
"[",
"k",
"]",
"=",
"v",
"\n",
"}",
"\n",
"for",
"k",
",",
"v",
":=",
"range",
"metric",
".",
"Labels",
"{",
"labels",
"[",
"k",
"]",
"=",
"v",
"\n",
"}",
"\n",
"timestamp",
":=",
"dataBatch",
".",
"Timestamp",
".",
"Unix",
"(",
")",
"\n",
"// creates an event and add it to dataEvent",
"events",
"=",
"appendEvent",
"(",
"events",
",",
"sink",
",",
"host",
",",
"metric",
".",
"Name",
",",
"value",
",",
"labels",
",",
"timestamp",
")",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"// Send events to Riemann if events is not empty",
"if",
"len",
"(",
"events",
")",
">",
"0",
"{",
"err",
":=",
"riemannCommon",
".",
"SendData",
"(",
"sink",
".",
"client",
",",
"events",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"// client will reconnect later",
"sink",
".",
"client",
"=",
"nil",
"\n",
"}",
"\n",
"}",
"\n",
"}"
] | // ExportData Send a collection of Timeseries to Riemann | [
"ExportData",
"Send",
"a",
"collection",
"of",
"Timeseries",
"to",
"Riemann"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/riemann/driver.go#L88-L137 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | availableClusterMetrics | func (a *HistoricalApi) availableClusterMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{ObjectType: core.MetricSetTypeCluster}
a.processMetricNamesRequest(key, response)
} | go | func (a *HistoricalApi) availableClusterMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{ObjectType: core.MetricSetTypeCluster}
a.processMetricNamesRequest(key, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"availableClusterMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeCluster",
"}",
"\n",
"a",
".",
"processMetricNamesRequest",
"(",
"key",
",",
"response",
")",
"\n",
"}"
] | // availableClusterMetrics returns a list of available cluster metric names. | [
"availableClusterMetrics",
"returns",
"a",
"list",
"of",
"available",
"cluster",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L259-L262 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | availableNodeMetrics | func (a *HistoricalApi) availableNodeMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeNode,
NodeName: request.PathParameter("node-name"),
}
a.processMetricNamesRequest(key, response)
} | go | func (a *HistoricalApi) availableNodeMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeNode,
NodeName: request.PathParameter("node-name"),
}
a.processMetricNamesRequest(key, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"availableNodeMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeNode",
",",
"NodeName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processMetricNamesRequest",
"(",
"key",
",",
"response",
")",
"\n",
"}"
] | // availableNodeMetrics returns a list of available node metric names. | [
"availableNodeMetrics",
"returns",
"a",
"list",
"of",
"available",
"node",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L265-L271 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | availableNamespaceMetrics | func (a *HistoricalApi) availableNamespaceMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeNamespace,
NamespaceName: request.PathParameter("namespace-name"),
}
a.processMetricNamesRequest(key, response)
} | go | func (a *HistoricalApi) availableNamespaceMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeNamespace,
NamespaceName: request.PathParameter("namespace-name"),
}
a.processMetricNamesRequest(key, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"availableNamespaceMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeNamespace",
",",
"NamespaceName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processMetricNamesRequest",
"(",
"key",
",",
"response",
")",
"\n",
"}"
] | // availableNamespaceMetrics returns a list of available namespace metric names. | [
"availableNamespaceMetrics",
"returns",
"a",
"list",
"of",
"available",
"namespace",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L274-L280 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | availablePodMetrics | func (a *HistoricalApi) availablePodMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
NamespaceName: request.PathParameter("namespace-name"),
PodName: request.PathParameter("pod-name"),
}
a.processMetricNamesRequest(key, response)
} | go | func (a *HistoricalApi) availablePodMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
NamespaceName: request.PathParameter("namespace-name"),
PodName: request.PathParameter("pod-name"),
}
a.processMetricNamesRequest(key, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"availablePodMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypePod",
",",
"NamespaceName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"PodName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processMetricNamesRequest",
"(",
"key",
",",
"response",
")",
"\n",
"}"
] | // availablePodMetrics returns a list of available pod metric names. | [
"availablePodMetrics",
"returns",
"a",
"list",
"of",
"available",
"pod",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L283-L290 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | availablePodContainerMetrics | func (a *HistoricalApi) availablePodContainerMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePodContainer,
NamespaceName: request.PathParameter("namespace-name"),
PodName: request.PathParameter("pod-name"),
ContainerName: request.PathParameter("container-name"),
}
a.processMetricNamesRequest(key, response)
} | go | func (a *HistoricalApi) availablePodContainerMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePodContainer,
NamespaceName: request.PathParameter("namespace-name"),
PodName: request.PathParameter("pod-name"),
ContainerName: request.PathParameter("container-name"),
}
a.processMetricNamesRequest(key, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"availablePodContainerMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypePodContainer",
",",
"NamespaceName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"PodName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"ContainerName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processMetricNamesRequest",
"(",
"key",
",",
"response",
")",
"\n",
"}"
] | // availablePodContainerMetrics returns a list of available pod container metric names. | [
"availablePodContainerMetrics",
"returns",
"a",
"list",
"of",
"available",
"pod",
"container",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L293-L301 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | availableFreeContainerMetrics | func (a *HistoricalApi) availableFreeContainerMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeSystemContainer,
NodeName: request.PathParameter("node-name"),
ContainerName: request.PathParameter("container-name"),
}
a.processMetricNamesRequest(key, response)
} | go | func (a *HistoricalApi) availableFreeContainerMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeSystemContainer,
NodeName: request.PathParameter("node-name"),
ContainerName: request.PathParameter("container-name"),
}
a.processMetricNamesRequest(key, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"availableFreeContainerMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeSystemContainer",
",",
"NodeName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"ContainerName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processMetricNamesRequest",
"(",
"key",
",",
"response",
")",
"\n",
"}"
] | // availableFreeContainerMetrics returns a list of available pod metric names. | [
"availableFreeContainerMetrics",
"returns",
"a",
"list",
"of",
"available",
"pod",
"metric",
"names",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L304-L311 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | nodeList | func (a *HistoricalApi) nodeList(request *restful.Request, response *restful.Response) {
if resp, err := a.historicalSource.GetNodes(); err != nil {
response.WriteError(http.StatusInternalServerError, err)
} else {
response.WriteEntity(resp)
}
} | go | func (a *HistoricalApi) nodeList(request *restful.Request, response *restful.Response) {
if resp, err := a.historicalSource.GetNodes(); err != nil {
response.WriteError(http.StatusInternalServerError, err)
} else {
response.WriteEntity(resp)
}
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"nodeList",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"if",
"resp",
",",
"err",
":=",
"a",
".",
"historicalSource",
".",
"GetNodes",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusInternalServerError",
",",
"err",
")",
"\n",
"}",
"else",
"{",
"response",
".",
"WriteEntity",
"(",
"resp",
")",
"\n",
"}",
"\n",
"}"
] | // nodeList lists all nodes for which we have metrics | [
"nodeList",
"lists",
"all",
"nodes",
"for",
"which",
"we",
"have",
"metrics"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L314-L320 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | namespaceList | func (a *HistoricalApi) namespaceList(request *restful.Request, response *restful.Response) {
if resp, err := a.historicalSource.GetNamespaces(); err != nil {
response.WriteError(http.StatusInternalServerError, err)
} else {
response.WriteEntity(resp)
}
} | go | func (a *HistoricalApi) namespaceList(request *restful.Request, response *restful.Response) {
if resp, err := a.historicalSource.GetNamespaces(); err != nil {
response.WriteError(http.StatusInternalServerError, err)
} else {
response.WriteEntity(resp)
}
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"namespaceList",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"if",
"resp",
",",
"err",
":=",
"a",
".",
"historicalSource",
".",
"GetNamespaces",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusInternalServerError",
",",
"err",
")",
"\n",
"}",
"else",
"{",
"response",
".",
"WriteEntity",
"(",
"resp",
")",
"\n",
"}",
"\n",
"}"
] | // namespaceList lists all namespaces for which we have metrics | [
"namespaceList",
"lists",
"all",
"namespaces",
"for",
"which",
"we",
"have",
"metrics"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L323-L329 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | namespacePodList | func (a *HistoricalApi) namespacePodList(request *restful.Request, response *restful.Response) {
if resp, err := a.historicalSource.GetPodsFromNamespace(request.PathParameter("namespace-name")); err != nil {
response.WriteError(http.StatusInternalServerError, err)
} else {
response.WriteEntity(resp)
}
} | go | func (a *HistoricalApi) namespacePodList(request *restful.Request, response *restful.Response) {
if resp, err := a.historicalSource.GetPodsFromNamespace(request.PathParameter("namespace-name")); err != nil {
response.WriteError(http.StatusInternalServerError, err)
} else {
response.WriteEntity(resp)
}
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"namespacePodList",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"if",
"resp",
",",
"err",
":=",
"a",
".",
"historicalSource",
".",
"GetPodsFromNamespace",
"(",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
")",
";",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusInternalServerError",
",",
"err",
")",
"\n",
"}",
"else",
"{",
"response",
".",
"WriteEntity",
"(",
"resp",
")",
"\n",
"}",
"\n",
"}"
] | // namespacePodList lists all pods for which we have metrics in a particular namespace | [
"namespacePodList",
"lists",
"all",
"pods",
"for",
"which",
"we",
"have",
"metrics",
"in",
"a",
"particular",
"namespace"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L332-L338 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | freeContainerMetrics | func (a *HistoricalApi) freeContainerMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeSystemContainer,
NodeName: request.PathParameter("node-name"),
ContainerName: request.PathParameter("container-name"),
}
a.processMetricRequest(key, request, response)
} | go | func (a *HistoricalApi) freeContainerMetrics(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeSystemContainer,
NodeName: request.PathParameter("node-name"),
ContainerName: request.PathParameter("container-name"),
}
a.processMetricRequest(key, request, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"freeContainerMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeSystemContainer",
",",
"NodeName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"ContainerName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processMetricRequest",
"(",
"key",
",",
"request",
",",
"response",
")",
"\n",
"}"
] | // freeContainerMetrics returns a metric timeseries for a metric of the Container entity.
// freeContainerMetrics addresses only free containers. | [
"freeContainerMetrics",
"returns",
"a",
"metric",
"timeseries",
"for",
"a",
"metric",
"of",
"the",
"Container",
"entity",
".",
"freeContainerMetrics",
"addresses",
"only",
"free",
"containers",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L393-L400 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | podListMetrics | func (a *HistoricalApi) podListMetrics(request *restful.Request, response *restful.Response) {
start, end, err := getStartEndTimeHistorical(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
keys := []core.HistoricalKey{}
if request.PathParameter("pod-id-list") != "" {
for _, podId := range strings.Split(request.PathParameter("pod-id-list"), ",") {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
PodId: podId,
}
keys = append(keys, key)
}
} else {
for _, podName := range strings.Split(request.PathParameter("pod-list"), ",") {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
NamespaceName: request.PathParameter("namespace-name"),
PodName: podName,
}
keys = append(keys, key)
}
}
labels, err := getLabels(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
metricName := request.PathParameter("metric-name")
convertedMetricName := convertMetricName(metricName)
var metrics map[core.HistoricalKey][]core.TimestampedMetricValue
if labels != nil {
metrics, err = a.historicalSource.GetLabeledMetric(convertedMetricName, labels, keys, start, end)
} else {
metrics, err = a.historicalSource.GetMetric(convertedMetricName, keys, start, end)
}
if err != nil {
response.WriteError(http.StatusInternalServerError, err)
return
}
result := types.MetricResultList{
Items: make([]types.MetricResult, 0, len(keys)),
}
for _, key := range keys {
result.Items = append(result.Items, exportTimestampedMetricValue(metrics[key]))
}
response.PrettyPrint(false)
response.WriteEntity(result)
} | go | func (a *HistoricalApi) podListMetrics(request *restful.Request, response *restful.Response) {
start, end, err := getStartEndTimeHistorical(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
keys := []core.HistoricalKey{}
if request.PathParameter("pod-id-list") != "" {
for _, podId := range strings.Split(request.PathParameter("pod-id-list"), ",") {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
PodId: podId,
}
keys = append(keys, key)
}
} else {
for _, podName := range strings.Split(request.PathParameter("pod-list"), ",") {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
NamespaceName: request.PathParameter("namespace-name"),
PodName: podName,
}
keys = append(keys, key)
}
}
labels, err := getLabels(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
metricName := request.PathParameter("metric-name")
convertedMetricName := convertMetricName(metricName)
var metrics map[core.HistoricalKey][]core.TimestampedMetricValue
if labels != nil {
metrics, err = a.historicalSource.GetLabeledMetric(convertedMetricName, labels, keys, start, end)
} else {
metrics, err = a.historicalSource.GetMetric(convertedMetricName, keys, start, end)
}
if err != nil {
response.WriteError(http.StatusInternalServerError, err)
return
}
result := types.MetricResultList{
Items: make([]types.MetricResult, 0, len(keys)),
}
for _, key := range keys {
result.Items = append(result.Items, exportTimestampedMetricValue(metrics[key]))
}
response.PrettyPrint(false)
response.WriteEntity(result)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"podListMetrics",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"start",
",",
"end",
",",
"err",
":=",
"getStartEndTimeHistorical",
"(",
"request",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusBadRequest",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"keys",
":=",
"[",
"]",
"core",
".",
"HistoricalKey",
"{",
"}",
"\n",
"if",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
"!=",
"\"",
"\"",
"{",
"for",
"_",
",",
"podId",
":=",
"range",
"strings",
".",
"Split",
"(",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"\"",
"\"",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypePod",
",",
"PodId",
":",
"podId",
",",
"}",
"\n",
"keys",
"=",
"append",
"(",
"keys",
",",
"key",
")",
"\n",
"}",
"\n",
"}",
"else",
"{",
"for",
"_",
",",
"podName",
":=",
"range",
"strings",
".",
"Split",
"(",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"\"",
"\"",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypePod",
",",
"NamespaceName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"PodName",
":",
"podName",
",",
"}",
"\n",
"keys",
"=",
"append",
"(",
"keys",
",",
"key",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"labels",
",",
"err",
":=",
"getLabels",
"(",
"request",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusBadRequest",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"metricName",
":=",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
"\n",
"convertedMetricName",
":=",
"convertMetricName",
"(",
"metricName",
")",
"\n\n",
"var",
"metrics",
"map",
"[",
"core",
".",
"HistoricalKey",
"]",
"[",
"]",
"core",
".",
"TimestampedMetricValue",
"\n",
"if",
"labels",
"!=",
"nil",
"{",
"metrics",
",",
"err",
"=",
"a",
".",
"historicalSource",
".",
"GetLabeledMetric",
"(",
"convertedMetricName",
",",
"labels",
",",
"keys",
",",
"start",
",",
"end",
")",
"\n",
"}",
"else",
"{",
"metrics",
",",
"err",
"=",
"a",
".",
"historicalSource",
".",
"GetMetric",
"(",
"convertedMetricName",
",",
"keys",
",",
"start",
",",
"end",
")",
"\n",
"}",
"\n\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusInternalServerError",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"result",
":=",
"types",
".",
"MetricResultList",
"{",
"Items",
":",
"make",
"(",
"[",
"]",
"types",
".",
"MetricResult",
",",
"0",
",",
"len",
"(",
"keys",
")",
")",
",",
"}",
"\n",
"for",
"_",
",",
"key",
":=",
"range",
"keys",
"{",
"result",
".",
"Items",
"=",
"append",
"(",
"result",
".",
"Items",
",",
"exportTimestampedMetricValue",
"(",
"metrics",
"[",
"key",
"]",
")",
")",
"\n",
"}",
"\n",
"response",
".",
"PrettyPrint",
"(",
"false",
")",
"\n",
"response",
".",
"WriteEntity",
"(",
"result",
")",
"\n",
"}"
] | // podListMetrics returns a list of metric timeseries for each for the listed nodes | [
"podListMetrics",
"returns",
"a",
"list",
"of",
"metric",
"timeseries",
"for",
"each",
"for",
"the",
"listed",
"nodes"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L403-L459 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | clusterAggregations | func (a *HistoricalApi) clusterAggregations(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{ObjectType: core.MetricSetTypeCluster}
a.processAggregationRequest(key, request, response)
} | go | func (a *HistoricalApi) clusterAggregations(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{ObjectType: core.MetricSetTypeCluster}
a.processAggregationRequest(key, request, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"clusterAggregations",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeCluster",
"}",
"\n",
"a",
".",
"processAggregationRequest",
"(",
"key",
",",
"request",
",",
"response",
")",
"\n",
"}"
] | // clusterAggregations returns a metric timeseries for a metric of the Cluster entity. | [
"clusterAggregations",
"returns",
"a",
"metric",
"timeseries",
"for",
"a",
"metric",
"of",
"the",
"Cluster",
"entity",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L482-L485 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | nodeAggregations | func (a *HistoricalApi) nodeAggregations(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeNode,
NodeName: request.PathParameter("node-name"),
}
a.processAggregationRequest(key, request, response)
} | go | func (a *HistoricalApi) nodeAggregations(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeNode,
NodeName: request.PathParameter("node-name"),
}
a.processAggregationRequest(key, request, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"nodeAggregations",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeNode",
",",
"NodeName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processAggregationRequest",
"(",
"key",
",",
"request",
",",
"response",
")",
"\n",
"}"
] | // nodeAggregations returns a metric timeseries for a metric of the Node entity. | [
"nodeAggregations",
"returns",
"a",
"metric",
"timeseries",
"for",
"a",
"metric",
"of",
"the",
"Node",
"entity",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L488-L494 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | namespaceAggregations | func (a *HistoricalApi) namespaceAggregations(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeNamespace,
NamespaceName: request.PathParameter("namespace-name"),
}
a.processAggregationRequest(key, request, response)
} | go | func (a *HistoricalApi) namespaceAggregations(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeNamespace,
NamespaceName: request.PathParameter("namespace-name"),
}
a.processAggregationRequest(key, request, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"namespaceAggregations",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeNamespace",
",",
"NamespaceName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processAggregationRequest",
"(",
"key",
",",
"request",
",",
"response",
")",
"\n",
"}"
] | // namespaceAggregations returns a metric timeseries for a metric of the Namespace entity. | [
"namespaceAggregations",
"returns",
"a",
"metric",
"timeseries",
"for",
"a",
"metric",
"of",
"the",
"Namespace",
"entity",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L497-L503 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | podContainerAggregations | func (a *HistoricalApi) podContainerAggregations(request *restful.Request, response *restful.Response) {
var key core.HistoricalKey
if request.PathParameter("pod-id") != "" {
key = core.HistoricalKey{
ObjectType: core.MetricSetTypePodContainer,
PodId: request.PathParameter("pod-id"),
ContainerName: request.PathParameter("container-name"),
}
} else {
key = core.HistoricalKey{
ObjectType: core.MetricSetTypePodContainer,
NamespaceName: request.PathParameter("namespace-name"),
PodName: request.PathParameter("pod-name"),
ContainerName: request.PathParameter("container-name"),
}
}
a.processAggregationRequest(key, request, response)
} | go | func (a *HistoricalApi) podContainerAggregations(request *restful.Request, response *restful.Response) {
var key core.HistoricalKey
if request.PathParameter("pod-id") != "" {
key = core.HistoricalKey{
ObjectType: core.MetricSetTypePodContainer,
PodId: request.PathParameter("pod-id"),
ContainerName: request.PathParameter("container-name"),
}
} else {
key = core.HistoricalKey{
ObjectType: core.MetricSetTypePodContainer,
NamespaceName: request.PathParameter("namespace-name"),
PodName: request.PathParameter("pod-name"),
ContainerName: request.PathParameter("container-name"),
}
}
a.processAggregationRequest(key, request, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"podContainerAggregations",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"var",
"key",
"core",
".",
"HistoricalKey",
"\n",
"if",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
"!=",
"\"",
"\"",
"{",
"key",
"=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypePodContainer",
",",
"PodId",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"ContainerName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"}",
"else",
"{",
"key",
"=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypePodContainer",
",",
"NamespaceName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"PodName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"ContainerName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"}",
"\n",
"a",
".",
"processAggregationRequest",
"(",
"key",
",",
"request",
",",
"response",
")",
"\n",
"}"
] | // podContainerAggregations returns a metric timeseries for a metric of a Pod Container entity. | [
"podContainerAggregations",
"returns",
"a",
"metric",
"timeseries",
"for",
"a",
"metric",
"of",
"a",
"Pod",
"Container",
"entity",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L524-L541 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | freeContainerAggregations | func (a *HistoricalApi) freeContainerAggregations(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeSystemContainer,
NodeName: request.PathParameter("node-name"),
ContainerName: request.PathParameter("container-name"),
}
a.processAggregationRequest(key, request, response)
} | go | func (a *HistoricalApi) freeContainerAggregations(request *restful.Request, response *restful.Response) {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypeSystemContainer,
NodeName: request.PathParameter("node-name"),
ContainerName: request.PathParameter("container-name"),
}
a.processAggregationRequest(key, request, response)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"freeContainerAggregations",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypeSystemContainer",
",",
"NodeName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"ContainerName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"}",
"\n",
"a",
".",
"processAggregationRequest",
"(",
"key",
",",
"request",
",",
"response",
")",
"\n",
"}"
] | // freeContainerAggregations returns a metric timeseries for a metric of the Container entity.
// freeContainerAggregations addresses only free containers. | [
"freeContainerAggregations",
"returns",
"a",
"metric",
"timeseries",
"for",
"a",
"metric",
"of",
"the",
"Container",
"entity",
".",
"freeContainerAggregations",
"addresses",
"only",
"free",
"containers",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L545-L552 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | podListAggregations | func (a *HistoricalApi) podListAggregations(request *restful.Request, response *restful.Response) {
start, end, err := getStartEndTimeHistorical(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
bucketSize, err := getBucketSize(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
aggregations, err := getAggregations(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
labels, err := getLabels(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
keys := []core.HistoricalKey{}
if request.PathParameter("pod-id-list") != "" {
for _, podId := range strings.Split(request.PathParameter("pod-id-list"), ",") {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
PodId: podId,
}
keys = append(keys, key)
}
} else {
for _, podName := range strings.Split(request.PathParameter("pod-list"), ",") {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
NamespaceName: request.PathParameter("namespace-name"),
PodName: podName,
}
keys = append(keys, key)
}
}
metricName := request.PathParameter("metric-name")
convertedMetricName := convertMetricName(metricName)
var metrics map[core.HistoricalKey][]core.TimestampedAggregationValue
if labels != nil {
metrics, err = a.historicalSource.GetLabeledAggregation(convertedMetricName, labels, aggregations, keys, start, end, bucketSize)
} else {
metrics, err = a.historicalSource.GetAggregation(convertedMetricName, aggregations, keys, start, end, bucketSize)
}
if err != nil {
response.WriteError(http.StatusInternalServerError, err)
return
}
result := types.MetricAggregationResultList{
Items: make([]types.MetricAggregationResult, 0, len(keys)),
}
for _, key := range keys {
result.Items = append(result.Items, exportTimestampedAggregationValue(metrics[key]))
}
response.PrettyPrint(false)
response.WriteEntity(result)
} | go | func (a *HistoricalApi) podListAggregations(request *restful.Request, response *restful.Response) {
start, end, err := getStartEndTimeHistorical(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
bucketSize, err := getBucketSize(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
aggregations, err := getAggregations(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
labels, err := getLabels(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
keys := []core.HistoricalKey{}
if request.PathParameter("pod-id-list") != "" {
for _, podId := range strings.Split(request.PathParameter("pod-id-list"), ",") {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
PodId: podId,
}
keys = append(keys, key)
}
} else {
for _, podName := range strings.Split(request.PathParameter("pod-list"), ",") {
key := core.HistoricalKey{
ObjectType: core.MetricSetTypePod,
NamespaceName: request.PathParameter("namespace-name"),
PodName: podName,
}
keys = append(keys, key)
}
}
metricName := request.PathParameter("metric-name")
convertedMetricName := convertMetricName(metricName)
var metrics map[core.HistoricalKey][]core.TimestampedAggregationValue
if labels != nil {
metrics, err = a.historicalSource.GetLabeledAggregation(convertedMetricName, labels, aggregations, keys, start, end, bucketSize)
} else {
metrics, err = a.historicalSource.GetAggregation(convertedMetricName, aggregations, keys, start, end, bucketSize)
}
if err != nil {
response.WriteError(http.StatusInternalServerError, err)
return
}
result := types.MetricAggregationResultList{
Items: make([]types.MetricAggregationResult, 0, len(keys)),
}
for _, key := range keys {
result.Items = append(result.Items, exportTimestampedAggregationValue(metrics[key]))
}
response.PrettyPrint(false)
response.WriteEntity(result)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"podListAggregations",
"(",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"start",
",",
"end",
",",
"err",
":=",
"getStartEndTimeHistorical",
"(",
"request",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusBadRequest",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n",
"bucketSize",
",",
"err",
":=",
"getBucketSize",
"(",
"request",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusBadRequest",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n",
"aggregations",
",",
"err",
":=",
"getAggregations",
"(",
"request",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusBadRequest",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n",
"labels",
",",
"err",
":=",
"getLabels",
"(",
"request",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusBadRequest",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n",
"keys",
":=",
"[",
"]",
"core",
".",
"HistoricalKey",
"{",
"}",
"\n",
"if",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
"!=",
"\"",
"\"",
"{",
"for",
"_",
",",
"podId",
":=",
"range",
"strings",
".",
"Split",
"(",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"\"",
"\"",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypePod",
",",
"PodId",
":",
"podId",
",",
"}",
"\n",
"keys",
"=",
"append",
"(",
"keys",
",",
"key",
")",
"\n",
"}",
"\n",
"}",
"else",
"{",
"for",
"_",
",",
"podName",
":=",
"range",
"strings",
".",
"Split",
"(",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"\"",
"\"",
")",
"{",
"key",
":=",
"core",
".",
"HistoricalKey",
"{",
"ObjectType",
":",
"core",
".",
"MetricSetTypePod",
",",
"NamespaceName",
":",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"PodName",
":",
"podName",
",",
"}",
"\n",
"keys",
"=",
"append",
"(",
"keys",
",",
"key",
")",
"\n",
"}",
"\n",
"}",
"\n",
"metricName",
":=",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
"\n",
"convertedMetricName",
":=",
"convertMetricName",
"(",
"metricName",
")",
"\n",
"var",
"metrics",
"map",
"[",
"core",
".",
"HistoricalKey",
"]",
"[",
"]",
"core",
".",
"TimestampedAggregationValue",
"\n",
"if",
"labels",
"!=",
"nil",
"{",
"metrics",
",",
"err",
"=",
"a",
".",
"historicalSource",
".",
"GetLabeledAggregation",
"(",
"convertedMetricName",
",",
"labels",
",",
"aggregations",
",",
"keys",
",",
"start",
",",
"end",
",",
"bucketSize",
")",
"\n",
"}",
"else",
"{",
"metrics",
",",
"err",
"=",
"a",
".",
"historicalSource",
".",
"GetAggregation",
"(",
"convertedMetricName",
",",
"aggregations",
",",
"keys",
",",
"start",
",",
"end",
",",
"bucketSize",
")",
"\n",
"}",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusInternalServerError",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"result",
":=",
"types",
".",
"MetricAggregationResultList",
"{",
"Items",
":",
"make",
"(",
"[",
"]",
"types",
".",
"MetricAggregationResult",
",",
"0",
",",
"len",
"(",
"keys",
")",
")",
",",
"}",
"\n",
"for",
"_",
",",
"key",
":=",
"range",
"keys",
"{",
"result",
".",
"Items",
"=",
"append",
"(",
"result",
".",
"Items",
",",
"exportTimestampedAggregationValue",
"(",
"metrics",
"[",
"key",
"]",
")",
")",
"\n",
"}",
"\n",
"response",
".",
"PrettyPrint",
"(",
"false",
")",
"\n",
"response",
".",
"WriteEntity",
"(",
"result",
")",
"\n",
"}"
] | // podListAggregations returns a list of metric timeseries for the specified pods. | [
"podListAggregations",
"returns",
"a",
"list",
"of",
"metric",
"timeseries",
"for",
"the",
"specified",
"pods",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L555-L616 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | processMetricRequest | func (a *HistoricalApi) processMetricRequest(key core.HistoricalKey, request *restful.Request, response *restful.Response) {
start, end, err := getStartEndTimeHistorical(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
labels, err := getLabels(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
metricName := request.PathParameter("metric-name")
convertedMetricName := convertMetricName(metricName)
var metrics map[core.HistoricalKey][]core.TimestampedMetricValue
if labels != nil {
metrics, err = a.historicalSource.GetLabeledMetric(convertedMetricName, labels, []core.HistoricalKey{key}, start, end)
} else {
metrics, err = a.historicalSource.GetMetric(convertedMetricName, []core.HistoricalKey{key}, start, end)
}
if err != nil {
response.WriteError(http.StatusInternalServerError, err)
return
}
converted := exportTimestampedMetricValue(metrics[key])
response.WriteEntity(converted)
} | go | func (a *HistoricalApi) processMetricRequest(key core.HistoricalKey, request *restful.Request, response *restful.Response) {
start, end, err := getStartEndTimeHistorical(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
labels, err := getLabels(request)
if err != nil {
response.WriteError(http.StatusBadRequest, err)
return
}
metricName := request.PathParameter("metric-name")
convertedMetricName := convertMetricName(metricName)
var metrics map[core.HistoricalKey][]core.TimestampedMetricValue
if labels != nil {
metrics, err = a.historicalSource.GetLabeledMetric(convertedMetricName, labels, []core.HistoricalKey{key}, start, end)
} else {
metrics, err = a.historicalSource.GetMetric(convertedMetricName, []core.HistoricalKey{key}, start, end)
}
if err != nil {
response.WriteError(http.StatusInternalServerError, err)
return
}
converted := exportTimestampedMetricValue(metrics[key])
response.WriteEntity(converted)
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"processMetricRequest",
"(",
"key",
"core",
".",
"HistoricalKey",
",",
"request",
"*",
"restful",
".",
"Request",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"start",
",",
"end",
",",
"err",
":=",
"getStartEndTimeHistorical",
"(",
"request",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusBadRequest",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n",
"labels",
",",
"err",
":=",
"getLabels",
"(",
"request",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusBadRequest",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n",
"metricName",
":=",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
"\n",
"convertedMetricName",
":=",
"convertMetricName",
"(",
"metricName",
")",
"\n\n",
"var",
"metrics",
"map",
"[",
"core",
".",
"HistoricalKey",
"]",
"[",
"]",
"core",
".",
"TimestampedMetricValue",
"\n",
"if",
"labels",
"!=",
"nil",
"{",
"metrics",
",",
"err",
"=",
"a",
".",
"historicalSource",
".",
"GetLabeledMetric",
"(",
"convertedMetricName",
",",
"labels",
",",
"[",
"]",
"core",
".",
"HistoricalKey",
"{",
"key",
"}",
",",
"start",
",",
"end",
")",
"\n",
"}",
"else",
"{",
"metrics",
",",
"err",
"=",
"a",
".",
"historicalSource",
".",
"GetMetric",
"(",
"convertedMetricName",
",",
"[",
"]",
"core",
".",
"HistoricalKey",
"{",
"key",
"}",
",",
"start",
",",
"end",
")",
"\n",
"}",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusInternalServerError",
",",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"converted",
":=",
"exportTimestampedMetricValue",
"(",
"metrics",
"[",
"key",
"]",
")",
"\n",
"response",
".",
"WriteEntity",
"(",
"converted",
")",
"\n",
"}"
] | // processMetricRequest retrieves a metric for the object at the requested key. | [
"processMetricRequest",
"retrieves",
"a",
"metric",
"for",
"the",
"object",
"at",
"the",
"requested",
"key",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L619-L646 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | processMetricNamesRequest | func (a *HistoricalApi) processMetricNamesRequest(key core.HistoricalKey, response *restful.Response) {
if resp, err := a.historicalSource.GetMetricNames(key); err != nil {
response.WriteError(http.StatusInternalServerError, err)
} else {
response.WriteEntity(resp)
}
} | go | func (a *HistoricalApi) processMetricNamesRequest(key core.HistoricalKey, response *restful.Response) {
if resp, err := a.historicalSource.GetMetricNames(key); err != nil {
response.WriteError(http.StatusInternalServerError, err)
} else {
response.WriteEntity(resp)
}
} | [
"func",
"(",
"a",
"*",
"HistoricalApi",
")",
"processMetricNamesRequest",
"(",
"key",
"core",
".",
"HistoricalKey",
",",
"response",
"*",
"restful",
".",
"Response",
")",
"{",
"if",
"resp",
",",
"err",
":=",
"a",
".",
"historicalSource",
".",
"GetMetricNames",
"(",
"key",
")",
";",
"err",
"!=",
"nil",
"{",
"response",
".",
"WriteError",
"(",
"http",
".",
"StatusInternalServerError",
",",
"err",
")",
"\n",
"}",
"else",
"{",
"response",
".",
"WriteEntity",
"(",
"resp",
")",
"\n",
"}",
"\n",
"}"
] | // processMetricNamesRequest retrieves the available metrics for the object at the specified key. | [
"processMetricNamesRequest",
"retrieves",
"the",
"available",
"metrics",
"for",
"the",
"object",
"at",
"the",
"specified",
"key",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L649-L655 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | getBucketSize | func getBucketSize(request *restful.Request) (time.Duration, error) {
rawSize := request.QueryParameter("bucket")
if rawSize == "" {
return 0, nil
}
if len(rawSize) < 2 {
return 0, fmt.Errorf("unable to parse bucket size: %q is too short to be a duration", rawSize)
}
var multiplier time.Duration
var num string
switch rawSize[len(rawSize)-1] {
case 's':
// could be s or ms
if len(rawSize) < 3 || rawSize[len(rawSize)-2] != 'm' {
multiplier = time.Second
num = rawSize[:len(rawSize)-1]
} else {
multiplier = time.Millisecond
num = rawSize[:len(rawSize)-2]
}
case 'h':
multiplier = time.Hour
num = rawSize[:len(rawSize)-1]
case 'd':
multiplier = 24 * time.Hour
num = rawSize[:len(rawSize)-1]
case 'm':
multiplier = time.Minute
num = rawSize[:len(rawSize)-1]
default:
return 0, fmt.Errorf("unable to parse bucket size: %q has no known duration suffix", rawSize)
}
parsedNum, err := strconv.ParseUint(num, 10, 64)
if err != nil {
return 0, err
}
return time.Duration(parsedNum) * multiplier, nil
} | go | func getBucketSize(request *restful.Request) (time.Duration, error) {
rawSize := request.QueryParameter("bucket")
if rawSize == "" {
return 0, nil
}
if len(rawSize) < 2 {
return 0, fmt.Errorf("unable to parse bucket size: %q is too short to be a duration", rawSize)
}
var multiplier time.Duration
var num string
switch rawSize[len(rawSize)-1] {
case 's':
// could be s or ms
if len(rawSize) < 3 || rawSize[len(rawSize)-2] != 'm' {
multiplier = time.Second
num = rawSize[:len(rawSize)-1]
} else {
multiplier = time.Millisecond
num = rawSize[:len(rawSize)-2]
}
case 'h':
multiplier = time.Hour
num = rawSize[:len(rawSize)-1]
case 'd':
multiplier = 24 * time.Hour
num = rawSize[:len(rawSize)-1]
case 'm':
multiplier = time.Minute
num = rawSize[:len(rawSize)-1]
default:
return 0, fmt.Errorf("unable to parse bucket size: %q has no known duration suffix", rawSize)
}
parsedNum, err := strconv.ParseUint(num, 10, 64)
if err != nil {
return 0, err
}
return time.Duration(parsedNum) * multiplier, nil
} | [
"func",
"getBucketSize",
"(",
"request",
"*",
"restful",
".",
"Request",
")",
"(",
"time",
".",
"Duration",
",",
"error",
")",
"{",
"rawSize",
":=",
"request",
".",
"QueryParameter",
"(",
"\"",
"\"",
")",
"\n",
"if",
"rawSize",
"==",
"\"",
"\"",
"{",
"return",
"0",
",",
"nil",
"\n",
"}",
"\n\n",
"if",
"len",
"(",
"rawSize",
")",
"<",
"2",
"{",
"return",
"0",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawSize",
")",
"\n",
"}",
"\n",
"var",
"multiplier",
"time",
".",
"Duration",
"\n",
"var",
"num",
"string",
"\n\n",
"switch",
"rawSize",
"[",
"len",
"(",
"rawSize",
")",
"-",
"1",
"]",
"{",
"case",
"'s'",
":",
"// could be s or ms",
"if",
"len",
"(",
"rawSize",
")",
"<",
"3",
"||",
"rawSize",
"[",
"len",
"(",
"rawSize",
")",
"-",
"2",
"]",
"!=",
"'m'",
"{",
"multiplier",
"=",
"time",
".",
"Second",
"\n",
"num",
"=",
"rawSize",
"[",
":",
"len",
"(",
"rawSize",
")",
"-",
"1",
"]",
"\n",
"}",
"else",
"{",
"multiplier",
"=",
"time",
".",
"Millisecond",
"\n",
"num",
"=",
"rawSize",
"[",
":",
"len",
"(",
"rawSize",
")",
"-",
"2",
"]",
"\n",
"}",
"\n",
"case",
"'h'",
":",
"multiplier",
"=",
"time",
".",
"Hour",
"\n",
"num",
"=",
"rawSize",
"[",
":",
"len",
"(",
"rawSize",
")",
"-",
"1",
"]",
"\n",
"case",
"'d'",
":",
"multiplier",
"=",
"24",
"*",
"time",
".",
"Hour",
"\n",
"num",
"=",
"rawSize",
"[",
":",
"len",
"(",
"rawSize",
")",
"-",
"1",
"]",
"\n",
"case",
"'m'",
":",
"multiplier",
"=",
"time",
".",
"Minute",
"\n",
"num",
"=",
"rawSize",
"[",
":",
"len",
"(",
"rawSize",
")",
"-",
"1",
"]",
"\n",
"default",
":",
"return",
"0",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"rawSize",
")",
"\n",
"}",
"\n\n",
"parsedNum",
",",
"err",
":=",
"strconv",
".",
"ParseUint",
"(",
"num",
",",
"10",
",",
"64",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"0",
",",
"err",
"\n",
"}",
"\n\n",
"return",
"time",
".",
"Duration",
"(",
"parsedNum",
")",
"*",
"multiplier",
",",
"nil",
"\n",
"}"
] | // getBucketSize parses the bucket size specifier into a | [
"getBucketSize",
"parses",
"the",
"bucket",
"size",
"specifier",
"into",
"a"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L698-L739 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | getAggregations | func getAggregations(request *restful.Request) ([]core.AggregationType, error) {
aggregationsRaw := strings.Split(request.PathParameter("aggregations"), ",")
if len(aggregationsRaw) == 0 {
return nil, fmt.Errorf("No aggregations specified")
}
aggregations := make([]core.AggregationType, len(aggregationsRaw))
for ind, aggNameRaw := range aggregationsRaw {
aggName := core.AggregationType(aggNameRaw)
if _, ok := core.AllAggregations[aggName]; !ok {
return nil, fmt.Errorf("Unknown aggregation %q", aggName)
}
aggregations[ind] = aggName
}
return aggregations, nil
} | go | func getAggregations(request *restful.Request) ([]core.AggregationType, error) {
aggregationsRaw := strings.Split(request.PathParameter("aggregations"), ",")
if len(aggregationsRaw) == 0 {
return nil, fmt.Errorf("No aggregations specified")
}
aggregations := make([]core.AggregationType, len(aggregationsRaw))
for ind, aggNameRaw := range aggregationsRaw {
aggName := core.AggregationType(aggNameRaw)
if _, ok := core.AllAggregations[aggName]; !ok {
return nil, fmt.Errorf("Unknown aggregation %q", aggName)
}
aggregations[ind] = aggName
}
return aggregations, nil
} | [
"func",
"getAggregations",
"(",
"request",
"*",
"restful",
".",
"Request",
")",
"(",
"[",
"]",
"core",
".",
"AggregationType",
",",
"error",
")",
"{",
"aggregationsRaw",
":=",
"strings",
".",
"Split",
"(",
"request",
".",
"PathParameter",
"(",
"\"",
"\"",
")",
",",
"\"",
"\"",
")",
"\n",
"if",
"len",
"(",
"aggregationsRaw",
")",
"==",
"0",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n\n",
"aggregations",
":=",
"make",
"(",
"[",
"]",
"core",
".",
"AggregationType",
",",
"len",
"(",
"aggregationsRaw",
")",
")",
"\n\n",
"for",
"ind",
",",
"aggNameRaw",
":=",
"range",
"aggregationsRaw",
"{",
"aggName",
":=",
"core",
".",
"AggregationType",
"(",
"aggNameRaw",
")",
"\n",
"if",
"_",
",",
"ok",
":=",
"core",
".",
"AllAggregations",
"[",
"aggName",
"]",
";",
"!",
"ok",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"aggName",
")",
"\n",
"}",
"\n",
"aggregations",
"[",
"ind",
"]",
"=",
"aggName",
"\n",
"}",
"\n\n",
"return",
"aggregations",
",",
"nil",
"\n",
"}"
] | // getAggregations extracts and validates the list of requested aggregations | [
"getAggregations",
"extracts",
"and",
"validates",
"the",
"list",
"of",
"requested",
"aggregations"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L742-L759 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | exportMetricValue | func exportMetricValue(value *core.MetricValue) *types.MetricValue {
if value == nil {
return nil
}
if value.ValueType == core.ValueInt64 {
return &types.MetricValue{
IntValue: &value.IntValue,
}
} else {
floatVal := float64(value.FloatValue)
return &types.MetricValue{
FloatValue: &floatVal,
}
}
} | go | func exportMetricValue(value *core.MetricValue) *types.MetricValue {
if value == nil {
return nil
}
if value.ValueType == core.ValueInt64 {
return &types.MetricValue{
IntValue: &value.IntValue,
}
} else {
floatVal := float64(value.FloatValue)
return &types.MetricValue{
FloatValue: &floatVal,
}
}
} | [
"func",
"exportMetricValue",
"(",
"value",
"*",
"core",
".",
"MetricValue",
")",
"*",
"types",
".",
"MetricValue",
"{",
"if",
"value",
"==",
"nil",
"{",
"return",
"nil",
"\n",
"}",
"\n\n",
"if",
"value",
".",
"ValueType",
"==",
"core",
".",
"ValueInt64",
"{",
"return",
"&",
"types",
".",
"MetricValue",
"{",
"IntValue",
":",
"&",
"value",
".",
"IntValue",
",",
"}",
"\n",
"}",
"else",
"{",
"floatVal",
":=",
"float64",
"(",
"value",
".",
"FloatValue",
")",
"\n",
"return",
"&",
"types",
".",
"MetricValue",
"{",
"FloatValue",
":",
"&",
"floatVal",
",",
"}",
"\n",
"}",
"\n",
"}"
] | // exportMetricValue converts a core.MetricValue into an API MetricValue | [
"exportMetricValue",
"converts",
"a",
"core",
".",
"MetricValue",
"into",
"an",
"API",
"MetricValue"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L762-L777 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | extractMetricValue | func extractMetricValue(aggregations *core.AggregationValue, aggName core.AggregationType) *types.MetricValue {
if inputVal, ok := aggregations.Aggregations[aggName]; ok {
return exportMetricValue(&inputVal)
} else {
return nil
}
} | go | func extractMetricValue(aggregations *core.AggregationValue, aggName core.AggregationType) *types.MetricValue {
if inputVal, ok := aggregations.Aggregations[aggName]; ok {
return exportMetricValue(&inputVal)
} else {
return nil
}
} | [
"func",
"extractMetricValue",
"(",
"aggregations",
"*",
"core",
".",
"AggregationValue",
",",
"aggName",
"core",
".",
"AggregationType",
")",
"*",
"types",
".",
"MetricValue",
"{",
"if",
"inputVal",
",",
"ok",
":=",
"aggregations",
".",
"Aggregations",
"[",
"aggName",
"]",
";",
"ok",
"{",
"return",
"exportMetricValue",
"(",
"&",
"inputVal",
")",
"\n",
"}",
"else",
"{",
"return",
"nil",
"\n",
"}",
"\n",
"}"
] | // extractMetricValue checks to see if the given metric was present in the results, and if so,
// returns it in API form | [
"extractMetricValue",
"checks",
"to",
"see",
"if",
"the",
"given",
"metric",
"was",
"present",
"in",
"the",
"results",
"and",
"if",
"so",
"returns",
"it",
"in",
"API",
"form"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L781-L787 | train |
kubernetes-retired/heapster | metrics/api/v1/historical_handlers.go | exportTimestampedAggregationValue | func exportTimestampedAggregationValue(values []core.TimestampedAggregationValue) types.MetricAggregationResult {
result := types.MetricAggregationResult{
Buckets: make([]types.MetricAggregationBucket, 0, len(values)),
BucketSize: 0,
}
for _, value := range values {
// just use the largest bucket size, since all bucket sizes should be uniform
// (except for the last one, which may be smaller)
if result.BucketSize < value.BucketSize {
result.BucketSize = value.BucketSize
}
bucket := types.MetricAggregationBucket{
Timestamp: value.Timestamp,
Count: value.Count,
Average: extractMetricValue(&value.AggregationValue, core.AggregationTypeAverage),
Maximum: extractMetricValue(&value.AggregationValue, core.AggregationTypeMaximum),
Minimum: extractMetricValue(&value.AggregationValue, core.AggregationTypeMinimum),
Median: extractMetricValue(&value.AggregationValue, core.AggregationTypeMedian),
Percentiles: make(map[string]types.MetricValue, 3),
}
if val, ok := value.Aggregations[core.AggregationTypePercentile50]; ok {
bucket.Percentiles["50"] = *exportMetricValue(&val)
}
if val, ok := value.Aggregations[core.AggregationTypePercentile95]; ok {
bucket.Percentiles["95"] = *exportMetricValue(&val)
}
if val, ok := value.Aggregations[core.AggregationTypePercentile99]; ok {
bucket.Percentiles["99"] = *exportMetricValue(&val)
}
result.Buckets = append(result.Buckets, bucket)
}
return result
} | go | func exportTimestampedAggregationValue(values []core.TimestampedAggregationValue) types.MetricAggregationResult {
result := types.MetricAggregationResult{
Buckets: make([]types.MetricAggregationBucket, 0, len(values)),
BucketSize: 0,
}
for _, value := range values {
// just use the largest bucket size, since all bucket sizes should be uniform
// (except for the last one, which may be smaller)
if result.BucketSize < value.BucketSize {
result.BucketSize = value.BucketSize
}
bucket := types.MetricAggregationBucket{
Timestamp: value.Timestamp,
Count: value.Count,
Average: extractMetricValue(&value.AggregationValue, core.AggregationTypeAverage),
Maximum: extractMetricValue(&value.AggregationValue, core.AggregationTypeMaximum),
Minimum: extractMetricValue(&value.AggregationValue, core.AggregationTypeMinimum),
Median: extractMetricValue(&value.AggregationValue, core.AggregationTypeMedian),
Percentiles: make(map[string]types.MetricValue, 3),
}
if val, ok := value.Aggregations[core.AggregationTypePercentile50]; ok {
bucket.Percentiles["50"] = *exportMetricValue(&val)
}
if val, ok := value.Aggregations[core.AggregationTypePercentile95]; ok {
bucket.Percentiles["95"] = *exportMetricValue(&val)
}
if val, ok := value.Aggregations[core.AggregationTypePercentile99]; ok {
bucket.Percentiles["99"] = *exportMetricValue(&val)
}
result.Buckets = append(result.Buckets, bucket)
}
return result
} | [
"func",
"exportTimestampedAggregationValue",
"(",
"values",
"[",
"]",
"core",
".",
"TimestampedAggregationValue",
")",
"types",
".",
"MetricAggregationResult",
"{",
"result",
":=",
"types",
".",
"MetricAggregationResult",
"{",
"Buckets",
":",
"make",
"(",
"[",
"]",
"types",
".",
"MetricAggregationBucket",
",",
"0",
",",
"len",
"(",
"values",
")",
")",
",",
"BucketSize",
":",
"0",
",",
"}",
"\n",
"for",
"_",
",",
"value",
":=",
"range",
"values",
"{",
"// just use the largest bucket size, since all bucket sizes should be uniform",
"// (except for the last one, which may be smaller)",
"if",
"result",
".",
"BucketSize",
"<",
"value",
".",
"BucketSize",
"{",
"result",
".",
"BucketSize",
"=",
"value",
".",
"BucketSize",
"\n",
"}",
"\n\n",
"bucket",
":=",
"types",
".",
"MetricAggregationBucket",
"{",
"Timestamp",
":",
"value",
".",
"Timestamp",
",",
"Count",
":",
"value",
".",
"Count",
",",
"Average",
":",
"extractMetricValue",
"(",
"&",
"value",
".",
"AggregationValue",
",",
"core",
".",
"AggregationTypeAverage",
")",
",",
"Maximum",
":",
"extractMetricValue",
"(",
"&",
"value",
".",
"AggregationValue",
",",
"core",
".",
"AggregationTypeMaximum",
")",
",",
"Minimum",
":",
"extractMetricValue",
"(",
"&",
"value",
".",
"AggregationValue",
",",
"core",
".",
"AggregationTypeMinimum",
")",
",",
"Median",
":",
"extractMetricValue",
"(",
"&",
"value",
".",
"AggregationValue",
",",
"core",
".",
"AggregationTypeMedian",
")",
",",
"Percentiles",
":",
"make",
"(",
"map",
"[",
"string",
"]",
"types",
".",
"MetricValue",
",",
"3",
")",
",",
"}",
"\n\n",
"if",
"val",
",",
"ok",
":=",
"value",
".",
"Aggregations",
"[",
"core",
".",
"AggregationTypePercentile50",
"]",
";",
"ok",
"{",
"bucket",
".",
"Percentiles",
"[",
"\"",
"\"",
"]",
"=",
"*",
"exportMetricValue",
"(",
"&",
"val",
")",
"\n",
"}",
"\n",
"if",
"val",
",",
"ok",
":=",
"value",
".",
"Aggregations",
"[",
"core",
".",
"AggregationTypePercentile95",
"]",
";",
"ok",
"{",
"bucket",
".",
"Percentiles",
"[",
"\"",
"\"",
"]",
"=",
"*",
"exportMetricValue",
"(",
"&",
"val",
")",
"\n",
"}",
"\n",
"if",
"val",
",",
"ok",
":=",
"value",
".",
"Aggregations",
"[",
"core",
".",
"AggregationTypePercentile99",
"]",
";",
"ok",
"{",
"bucket",
".",
"Percentiles",
"[",
"\"",
"\"",
"]",
"=",
"*",
"exportMetricValue",
"(",
"&",
"val",
")",
"\n",
"}",
"\n\n",
"result",
".",
"Buckets",
"=",
"append",
"(",
"result",
".",
"Buckets",
",",
"bucket",
")",
"\n",
"}",
"\n",
"return",
"result",
"\n",
"}"
] | // exportTimestampedAggregationValue converts a core.TimestampedAggregationValue into an API MetricAggregationResult | [
"exportTimestampedAggregationValue",
"converts",
"a",
"core",
".",
"TimestampedAggregationValue",
"into",
"an",
"API",
"MetricAggregationResult"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L790-L828 | train |
kubernetes-retired/heapster | metrics/sources/summary/summary.go | decodeSummary | func (this *summaryMetricsSource) decodeSummary(summary *stats.Summary) map[string]*MetricSet {
glog.V(9).Infof("Begin summary decode")
result := map[string]*MetricSet{}
labels := map[string]string{
LabelNodename.Key: this.node.NodeName,
LabelHostname.Key: this.node.HostName,
LabelHostID.Key: this.node.HostID,
}
this.decodeNodeStats(result, labels, &summary.Node)
for _, pod := range summary.Pods {
this.decodePodStats(result, labels, &pod)
}
glog.V(9).Infof("End summary decode")
return result
} | go | func (this *summaryMetricsSource) decodeSummary(summary *stats.Summary) map[string]*MetricSet {
glog.V(9).Infof("Begin summary decode")
result := map[string]*MetricSet{}
labels := map[string]string{
LabelNodename.Key: this.node.NodeName,
LabelHostname.Key: this.node.HostName,
LabelHostID.Key: this.node.HostID,
}
this.decodeNodeStats(result, labels, &summary.Node)
for _, pod := range summary.Pods {
this.decodePodStats(result, labels, &pod)
}
glog.V(9).Infof("End summary decode")
return result
} | [
"func",
"(",
"this",
"*",
"summaryMetricsSource",
")",
"decodeSummary",
"(",
"summary",
"*",
"stats",
".",
"Summary",
")",
"map",
"[",
"string",
"]",
"*",
"MetricSet",
"{",
"glog",
".",
"V",
"(",
"9",
")",
".",
"Infof",
"(",
"\"",
"\"",
")",
"\n",
"result",
":=",
"map",
"[",
"string",
"]",
"*",
"MetricSet",
"{",
"}",
"\n\n",
"labels",
":=",
"map",
"[",
"string",
"]",
"string",
"{",
"LabelNodename",
".",
"Key",
":",
"this",
".",
"node",
".",
"NodeName",
",",
"LabelHostname",
".",
"Key",
":",
"this",
".",
"node",
".",
"HostName",
",",
"LabelHostID",
".",
"Key",
":",
"this",
".",
"node",
".",
"HostID",
",",
"}",
"\n\n",
"this",
".",
"decodeNodeStats",
"(",
"result",
",",
"labels",
",",
"&",
"summary",
".",
"Node",
")",
"\n",
"for",
"_",
",",
"pod",
":=",
"range",
"summary",
".",
"Pods",
"{",
"this",
".",
"decodePodStats",
"(",
"result",
",",
"labels",
",",
"&",
"pod",
")",
"\n",
"}",
"\n\n",
"glog",
".",
"V",
"(",
"9",
")",
".",
"Infof",
"(",
"\"",
"\"",
")",
"\n",
"return",
"result",
"\n",
"}"
] | // decodeSummary translates the kubelet statsSummary API into the flattened heapster MetricSet API. | [
"decodeSummary",
"translates",
"the",
"kubelet",
"statsSummary",
"API",
"into",
"the",
"flattened",
"heapster",
"MetricSet",
"API",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L120-L137 | train |
kubernetes-retired/heapster | metrics/sources/summary/summary.go | cloneLabels | func (this *summaryMetricsSource) cloneLabels(labels map[string]string) map[string]string {
clone := make(map[string]string, len(labels))
for k, v := range labels {
clone[k] = v
}
return clone
} | go | func (this *summaryMetricsSource) cloneLabels(labels map[string]string) map[string]string {
clone := make(map[string]string, len(labels))
for k, v := range labels {
clone[k] = v
}
return clone
} | [
"func",
"(",
"this",
"*",
"summaryMetricsSource",
")",
"cloneLabels",
"(",
"labels",
"map",
"[",
"string",
"]",
"string",
")",
"map",
"[",
"string",
"]",
"string",
"{",
"clone",
":=",
"make",
"(",
"map",
"[",
"string",
"]",
"string",
",",
"len",
"(",
"labels",
")",
")",
"\n",
"for",
"k",
",",
"v",
":=",
"range",
"labels",
"{",
"clone",
"[",
"k",
"]",
"=",
"v",
"\n",
"}",
"\n",
"return",
"clone",
"\n",
"}"
] | // Convenience method for labels deep copy. | [
"Convenience",
"method",
"for",
"labels",
"deep",
"copy",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L140-L146 | train |
kubernetes-retired/heapster | metrics/sources/summary/summary.go | addIntMetric | func (this *summaryMetricsSource) addIntMetric(metrics *MetricSet, metric *Metric, value *uint64) {
if value == nil {
glog.V(9).Infof("skipping metric %s because the value was nil", metric.Name)
return
}
val := MetricValue{
ValueType: ValueInt64,
MetricType: metric.Type,
IntValue: int64(*value),
}
metrics.MetricValues[metric.Name] = val
} | go | func (this *summaryMetricsSource) addIntMetric(metrics *MetricSet, metric *Metric, value *uint64) {
if value == nil {
glog.V(9).Infof("skipping metric %s because the value was nil", metric.Name)
return
}
val := MetricValue{
ValueType: ValueInt64,
MetricType: metric.Type,
IntValue: int64(*value),
}
metrics.MetricValues[metric.Name] = val
} | [
"func",
"(",
"this",
"*",
"summaryMetricsSource",
")",
"addIntMetric",
"(",
"metrics",
"*",
"MetricSet",
",",
"metric",
"*",
"Metric",
",",
"value",
"*",
"uint64",
")",
"{",
"if",
"value",
"==",
"nil",
"{",
"glog",
".",
"V",
"(",
"9",
")",
".",
"Infof",
"(",
"\"",
"\"",
",",
"metric",
".",
"Name",
")",
"\n",
"return",
"\n",
"}",
"\n",
"val",
":=",
"MetricValue",
"{",
"ValueType",
":",
"ValueInt64",
",",
"MetricType",
":",
"metric",
".",
"Type",
",",
"IntValue",
":",
"int64",
"(",
"*",
"value",
")",
",",
"}",
"\n",
"metrics",
".",
"MetricValues",
"[",
"metric",
".",
"Name",
"]",
"=",
"val",
"\n",
"}"
] | // addIntMetric is a convenience method for adding the metric and value to the metric set. | [
"addIntMetric",
"is",
"a",
"convenience",
"method",
"for",
"adding",
"the",
"metric",
"and",
"value",
"to",
"the",
"metric",
"set",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L370-L381 | train |
kubernetes-retired/heapster | metrics/sources/summary/summary.go | addLabeledIntMetric | func (this *summaryMetricsSource) addLabeledIntMetric(metrics *MetricSet, metric *Metric, labels map[string]string, value *uint64) {
if value == nil {
glog.V(9).Infof("skipping labeled metric %s (%v) because the value was nil", metric.Name, labels)
return
}
val := LabeledMetric{
Name: metric.Name,
Labels: labels,
MetricValue: MetricValue{
ValueType: ValueInt64,
MetricType: metric.Type,
IntValue: int64(*value),
},
}
metrics.LabeledMetrics = append(metrics.LabeledMetrics, val)
} | go | func (this *summaryMetricsSource) addLabeledIntMetric(metrics *MetricSet, metric *Metric, labels map[string]string, value *uint64) {
if value == nil {
glog.V(9).Infof("skipping labeled metric %s (%v) because the value was nil", metric.Name, labels)
return
}
val := LabeledMetric{
Name: metric.Name,
Labels: labels,
MetricValue: MetricValue{
ValueType: ValueInt64,
MetricType: metric.Type,
IntValue: int64(*value),
},
}
metrics.LabeledMetrics = append(metrics.LabeledMetrics, val)
} | [
"func",
"(",
"this",
"*",
"summaryMetricsSource",
")",
"addLabeledIntMetric",
"(",
"metrics",
"*",
"MetricSet",
",",
"metric",
"*",
"Metric",
",",
"labels",
"map",
"[",
"string",
"]",
"string",
",",
"value",
"*",
"uint64",
")",
"{",
"if",
"value",
"==",
"nil",
"{",
"glog",
".",
"V",
"(",
"9",
")",
".",
"Infof",
"(",
"\"",
"\"",
",",
"metric",
".",
"Name",
",",
"labels",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"val",
":=",
"LabeledMetric",
"{",
"Name",
":",
"metric",
".",
"Name",
",",
"Labels",
":",
"labels",
",",
"MetricValue",
":",
"MetricValue",
"{",
"ValueType",
":",
"ValueInt64",
",",
"MetricType",
":",
"metric",
".",
"Type",
",",
"IntValue",
":",
"int64",
"(",
"*",
"value",
")",
",",
"}",
",",
"}",
"\n",
"metrics",
".",
"LabeledMetrics",
"=",
"append",
"(",
"metrics",
".",
"LabeledMetrics",
",",
"val",
")",
"\n",
"}"
] | // addLabeledIntMetric is a convenience method for adding the labeled metric and value to the metric set. | [
"addLabeledIntMetric",
"is",
"a",
"convenience",
"method",
"for",
"adding",
"the",
"labeled",
"metric",
"and",
"value",
"to",
"the",
"metric",
"set",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L384-L400 | train |
kubernetes-retired/heapster | metrics/sources/summary/summary.go | getSystemContainerName | func (this *summaryMetricsSource) getSystemContainerName(c *stats.ContainerStats) string {
if legacyName, ok := systemNameMap[c.Name]; ok {
return legacyName
}
return c.Name
} | go | func (this *summaryMetricsSource) getSystemContainerName(c *stats.ContainerStats) string {
if legacyName, ok := systemNameMap[c.Name]; ok {
return legacyName
}
return c.Name
} | [
"func",
"(",
"this",
"*",
"summaryMetricsSource",
")",
"getSystemContainerName",
"(",
"c",
"*",
"stats",
".",
"ContainerStats",
")",
"string",
"{",
"if",
"legacyName",
",",
"ok",
":=",
"systemNameMap",
"[",
"c",
".",
"Name",
"]",
";",
"ok",
"{",
"return",
"legacyName",
"\n",
"}",
"\n",
"return",
"c",
".",
"Name",
"\n",
"}"
] | // Translate system container names to the legacy names for backwards compatibility. | [
"Translate",
"system",
"container",
"names",
"to",
"the",
"legacy",
"names",
"for",
"backwards",
"compatibility",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L403-L408 | train |
kubernetes-retired/heapster | common/riemann/riemann.go | GetRiemannClient | func GetRiemannClient(config RiemannConfig) (riemanngo.Client, error) {
glog.Infof("Connect Riemann client...")
client := riemanngo.NewTcpClient(config.Host)
runtime.SetFinalizer(client, func(c riemanngo.Client) { c.Close() })
// 5 seconds timeout
err := client.Connect(5)
if err != nil {
return nil, err
}
return client, nil
} | go | func GetRiemannClient(config RiemannConfig) (riemanngo.Client, error) {
glog.Infof("Connect Riemann client...")
client := riemanngo.NewTcpClient(config.Host)
runtime.SetFinalizer(client, func(c riemanngo.Client) { c.Close() })
// 5 seconds timeout
err := client.Connect(5)
if err != nil {
return nil, err
}
return client, nil
} | [
"func",
"GetRiemannClient",
"(",
"config",
"RiemannConfig",
")",
"(",
"riemanngo",
".",
"Client",
",",
"error",
")",
"{",
"glog",
".",
"Infof",
"(",
"\"",
"\"",
")",
"\n",
"client",
":=",
"riemanngo",
".",
"NewTcpClient",
"(",
"config",
".",
"Host",
")",
"\n",
"runtime",
".",
"SetFinalizer",
"(",
"client",
",",
"func",
"(",
"c",
"riemanngo",
".",
"Client",
")",
"{",
"c",
".",
"Close",
"(",
")",
"}",
")",
"\n",
"// 5 seconds timeout",
"err",
":=",
"client",
".",
"Connect",
"(",
"5",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"return",
"client",
",",
"nil",
"\n",
"}"
] | // Receives a sink, connect the riemann client. | [
"Receives",
"a",
"sink",
"connect",
"the",
"riemann",
"client",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/common/riemann/riemann.go#L102-L112 | train |
kubernetes-retired/heapster | common/riemann/riemann.go | SendData | func SendData(client riemanngo.Client, events []riemanngo.Event) error {
// do nothing if we are not connected
if client == nil {
glog.Warningf("Riemann sink not connected")
return nil
}
start := time.Now()
_, err := riemanngo.SendEvents(client, &events)
end := time.Now()
if err == nil {
glog.V(4).Infof("Exported %d events to riemann in %s", len(events), end.Sub(start))
return nil
} else {
glog.Warningf("There were errors sending events to Riemman, forcing reconnection. Error : %+v", err)
return err
}
} | go | func SendData(client riemanngo.Client, events []riemanngo.Event) error {
// do nothing if we are not connected
if client == nil {
glog.Warningf("Riemann sink not connected")
return nil
}
start := time.Now()
_, err := riemanngo.SendEvents(client, &events)
end := time.Now()
if err == nil {
glog.V(4).Infof("Exported %d events to riemann in %s", len(events), end.Sub(start))
return nil
} else {
glog.Warningf("There were errors sending events to Riemman, forcing reconnection. Error : %+v", err)
return err
}
} | [
"func",
"SendData",
"(",
"client",
"riemanngo",
".",
"Client",
",",
"events",
"[",
"]",
"riemanngo",
".",
"Event",
")",
"error",
"{",
"// do nothing if we are not connected",
"if",
"client",
"==",
"nil",
"{",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
")",
"\n",
"return",
"nil",
"\n",
"}",
"\n",
"start",
":=",
"time",
".",
"Now",
"(",
")",
"\n",
"_",
",",
"err",
":=",
"riemanngo",
".",
"SendEvents",
"(",
"client",
",",
"&",
"events",
")",
"\n",
"end",
":=",
"time",
".",
"Now",
"(",
")",
"\n",
"if",
"err",
"==",
"nil",
"{",
"glog",
".",
"V",
"(",
"4",
")",
".",
"Infof",
"(",
"\"",
"\"",
",",
"len",
"(",
"events",
")",
",",
"end",
".",
"Sub",
"(",
"start",
")",
")",
"\n",
"return",
"nil",
"\n",
"}",
"else",
"{",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"return",
"err",
"\n",
"}",
"\n",
"}"
] | // Send Events to Riemann using the client from the sink. | [
"Send",
"Events",
"to",
"Riemann",
"using",
"the",
"client",
"from",
"the",
"sink",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/common/riemann/riemann.go#L115-L131 | train |
kubernetes-retired/heapster | events/sinks/manager.go | ExportEvents | func (this *sinkManager) ExportEvents(data *core.EventBatch) {
var wg sync.WaitGroup
for _, sh := range this.sinkHolders {
wg.Add(1)
go func(sh sinkHolder, wg *sync.WaitGroup) {
defer wg.Done()
glog.V(2).Infof("Pushing events to: %s", sh.sink.Name())
select {
case sh.eventBatchChannel <- data:
glog.V(2).Infof("Data events completed: %s", sh.sink.Name())
// everything ok
case <-time.After(this.exportEventsTimeout):
glog.Warningf("Failed to events data to sink: %s", sh.sink.Name())
}
}(sh, &wg)
}
// Wait for all pushes to complete or timeout.
wg.Wait()
} | go | func (this *sinkManager) ExportEvents(data *core.EventBatch) {
var wg sync.WaitGroup
for _, sh := range this.sinkHolders {
wg.Add(1)
go func(sh sinkHolder, wg *sync.WaitGroup) {
defer wg.Done()
glog.V(2).Infof("Pushing events to: %s", sh.sink.Name())
select {
case sh.eventBatchChannel <- data:
glog.V(2).Infof("Data events completed: %s", sh.sink.Name())
// everything ok
case <-time.After(this.exportEventsTimeout):
glog.Warningf("Failed to events data to sink: %s", sh.sink.Name())
}
}(sh, &wg)
}
// Wait for all pushes to complete or timeout.
wg.Wait()
} | [
"func",
"(",
"this",
"*",
"sinkManager",
")",
"ExportEvents",
"(",
"data",
"*",
"core",
".",
"EventBatch",
")",
"{",
"var",
"wg",
"sync",
".",
"WaitGroup",
"\n",
"for",
"_",
",",
"sh",
":=",
"range",
"this",
".",
"sinkHolders",
"{",
"wg",
".",
"Add",
"(",
"1",
")",
"\n",
"go",
"func",
"(",
"sh",
"sinkHolder",
",",
"wg",
"*",
"sync",
".",
"WaitGroup",
")",
"{",
"defer",
"wg",
".",
"Done",
"(",
")",
"\n",
"glog",
".",
"V",
"(",
"2",
")",
".",
"Infof",
"(",
"\"",
"\"",
",",
"sh",
".",
"sink",
".",
"Name",
"(",
")",
")",
"\n",
"select",
"{",
"case",
"sh",
".",
"eventBatchChannel",
"<-",
"data",
":",
"glog",
".",
"V",
"(",
"2",
")",
".",
"Infof",
"(",
"\"",
"\"",
",",
"sh",
".",
"sink",
".",
"Name",
"(",
")",
")",
"\n",
"// everything ok",
"case",
"<-",
"time",
".",
"After",
"(",
"this",
".",
"exportEventsTimeout",
")",
":",
"glog",
".",
"Warningf",
"(",
"\"",
"\"",
",",
"sh",
".",
"sink",
".",
"Name",
"(",
")",
")",
"\n",
"}",
"\n",
"}",
"(",
"sh",
",",
"&",
"wg",
")",
"\n",
"}",
"\n",
"// Wait for all pushes to complete or timeout.",
"wg",
".",
"Wait",
"(",
")",
"\n",
"}"
] | // Guarantees that the export will complete in exportEventsTimeout. | [
"Guarantees",
"that",
"the",
"export",
"will",
"complete",
"in",
"exportEventsTimeout",
"."
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/events/sinks/manager.go#L96-L114 | train |
kubernetes-retired/heapster | metrics/sinks/hawkular/client.go | cache | func (h *hawkularSink) cache(md *metrics.MetricDefinition) {
h.pushToCache(md.ID, hashDefinition(md))
} | go | func (h *hawkularSink) cache(md *metrics.MetricDefinition) {
h.pushToCache(md.ID, hashDefinition(md))
} | [
"func",
"(",
"h",
"*",
"hawkularSink",
")",
"cache",
"(",
"md",
"*",
"metrics",
".",
"MetricDefinition",
")",
"{",
"h",
".",
"pushToCache",
"(",
"md",
".",
"ID",
",",
"hashDefinition",
"(",
"md",
")",
")",
"\n",
"}"
] | // cache inserts the item to the cache | [
"cache",
"inserts",
"the",
"item",
"to",
"the",
"cache"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/hawkular/client.go#L50-L52 | train |
kubernetes-retired/heapster | metrics/sinks/hawkular/client.go | pushToCache | func (h *hawkularSink) pushToCache(key string, hash uint64) {
h.regLock.Lock()
h.expReg[key] = &expiringItem{
hash: hash,
ttl: h.runId,
}
h.regLock.Unlock()
} | go | func (h *hawkularSink) pushToCache(key string, hash uint64) {
h.regLock.Lock()
h.expReg[key] = &expiringItem{
hash: hash,
ttl: h.runId,
}
h.regLock.Unlock()
} | [
"func",
"(",
"h",
"*",
"hawkularSink",
")",
"pushToCache",
"(",
"key",
"string",
",",
"hash",
"uint64",
")",
"{",
"h",
".",
"regLock",
".",
"Lock",
"(",
")",
"\n",
"h",
".",
"expReg",
"[",
"key",
"]",
"=",
"&",
"expiringItem",
"{",
"hash",
":",
"hash",
",",
"ttl",
":",
"h",
".",
"runId",
",",
"}",
"\n",
"h",
".",
"regLock",
".",
"Unlock",
"(",
")",
"\n",
"}"
] | // toCache inserts the item and updates the TTL in the cache to current time | [
"toCache",
"inserts",
"the",
"item",
"and",
"updates",
"the",
"TTL",
"in",
"the",
"cache",
"to",
"current",
"time"
] | e1e83412787b60d8a70088f09a2cb12339b305c3 | https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/hawkular/client.go#L55-L62 | train |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.