repo
stringlengths
5
54
path
stringlengths
4
155
func_name
stringlengths
1
118
original_string
stringlengths
52
85.5k
language
stringclasses
1 value
code
stringlengths
52
85.5k
code_tokens
sequence
docstring
stringlengths
6
2.61k
docstring_tokens
sequence
sha
stringlengths
40
40
url
stringlengths
85
252
partition
stringclasses
1 value
anacrolix/torrent
torrent.go
bytesCompleted
func (t *Torrent) bytesCompleted() int64 { if !t.haveInfo() { return 0 } return t.info.TotalLength() - t.bytesLeft() }
go
func (t *Torrent) bytesCompleted() int64 { if !t.haveInfo() { return 0 } return t.info.TotalLength() - t.bytesLeft() }
[ "func", "(", "t", "*", "Torrent", ")", "bytesCompleted", "(", ")", "int64", "{", "if", "!", "t", ".", "haveInfo", "(", ")", "{", "return", "0", "\n", "}", "\n", "return", "t", ".", "info", ".", "TotalLength", "(", ")", "-", "t", ".", "bytesLeft", "(", ")", "\n", "}" ]
// Don't call this before the info is available.
[ "Don", "t", "call", "this", "before", "the", "info", "is", "available", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1162-L1167
train
anacrolix/torrent
torrent.go
deleteConnection
func (t *Torrent) deleteConnection(c *connection) (ret bool) { if !c.closed.IsSet() { panic("connection is not closed") // There are behaviours prevented by the closed state that will fail // if the connection has been deleted. } _, ret = t.conns[c] delete(t.conns, c) torrent.Add("deleted connections", 1) c.deleteAllRequests() if len(t.conns) == 0 { t.assertNoPendingRequests() } return }
go
func (t *Torrent) deleteConnection(c *connection) (ret bool) { if !c.closed.IsSet() { panic("connection is not closed") // There are behaviours prevented by the closed state that will fail // if the connection has been deleted. } _, ret = t.conns[c] delete(t.conns, c) torrent.Add("deleted connections", 1) c.deleteAllRequests() if len(t.conns) == 0 { t.assertNoPendingRequests() } return }
[ "func", "(", "t", "*", "Torrent", ")", "deleteConnection", "(", "c", "*", "connection", ")", "(", "ret", "bool", ")", "{", "if", "!", "c", ".", "closed", ".", "IsSet", "(", ")", "{", "panic", "(", "\"", "\"", ")", "\n", "// There are behaviours prevented by the closed state that will fail", "// if the connection has been deleted.", "}", "\n", "_", ",", "ret", "=", "t", ".", "conns", "[", "c", "]", "\n", "delete", "(", "t", ".", "conns", ",", "c", ")", "\n", "torrent", ".", "Add", "(", "\"", "\"", ",", "1", ")", "\n", "c", ".", "deleteAllRequests", "(", ")", "\n", "if", "len", "(", "t", ".", "conns", ")", "==", "0", "{", "t", ".", "assertNoPendingRequests", "(", ")", "\n", "}", "\n", "return", "\n", "}" ]
// Returns true if connection is removed from torrent.Conns.
[ "Returns", "true", "if", "connection", "is", "removed", "from", "torrent", ".", "Conns", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1176-L1190
train
anacrolix/torrent
torrent.go
seeding
func (t *Torrent) seeding() bool { cl := t.cl if t.closed.IsSet() { return false } if cl.config.NoUpload { return false } if !cl.config.Seed { return false } if cl.config.DisableAggressiveUpload && t.needData() { return false } return true }
go
func (t *Torrent) seeding() bool { cl := t.cl if t.closed.IsSet() { return false } if cl.config.NoUpload { return false } if !cl.config.Seed { return false } if cl.config.DisableAggressiveUpload && t.needData() { return false } return true }
[ "func", "(", "t", "*", "Torrent", ")", "seeding", "(", ")", "bool", "{", "cl", ":=", "t", ".", "cl", "\n", "if", "t", ".", "closed", ".", "IsSet", "(", ")", "{", "return", "false", "\n", "}", "\n", "if", "cl", ".", "config", ".", "NoUpload", "{", "return", "false", "\n", "}", "\n", "if", "!", "cl", ".", "config", ".", "Seed", "{", "return", "false", "\n", "}", "\n", "if", "cl", ".", "config", ".", "DisableAggressiveUpload", "&&", "t", ".", "needData", "(", ")", "{", "return", "false", "\n", "}", "\n", "return", "true", "\n", "}" ]
// Returns whether the client should make effort to seed the torrent.
[ "Returns", "whether", "the", "client", "should", "make", "effort", "to", "seed", "the", "torrent", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1228-L1243
train
anacrolix/torrent
torrent.go
startMissingTrackerScrapers
func (t *Torrent) startMissingTrackerScrapers() { if t.cl.config.DisableTrackers { return } t.startScrapingTracker(t.metainfo.Announce) for _, tier := range t.metainfo.AnnounceList { for _, url := range tier { t.startScrapingTracker(url) } } }
go
func (t *Torrent) startMissingTrackerScrapers() { if t.cl.config.DisableTrackers { return } t.startScrapingTracker(t.metainfo.Announce) for _, tier := range t.metainfo.AnnounceList { for _, url := range tier { t.startScrapingTracker(url) } } }
[ "func", "(", "t", "*", "Torrent", ")", "startMissingTrackerScrapers", "(", ")", "{", "if", "t", ".", "cl", ".", "config", ".", "DisableTrackers", "{", "return", "\n", "}", "\n", "t", ".", "startScrapingTracker", "(", "t", ".", "metainfo", ".", "Announce", ")", "\n", "for", "_", ",", "tier", ":=", "range", "t", ".", "metainfo", ".", "AnnounceList", "{", "for", "_", ",", "url", ":=", "range", "tier", "{", "t", ".", "startScrapingTracker", "(", "url", ")", "\n", "}", "\n", "}", "\n", "}" ]
// Adds and starts tracker scrapers for tracker URLs that aren't already // running.
[ "Adds", "and", "starts", "tracker", "scrapers", "for", "tracker", "URLs", "that", "aren", "t", "already", "running", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1287-L1297
train
anacrolix/torrent
torrent.go
announceRequest
func (t *Torrent) announceRequest() tracker.AnnounceRequest { // Note that IPAddress is not set. It's set for UDP inside the tracker // code, since it's dependent on the network in use. return tracker.AnnounceRequest{ Event: tracker.None, NumWant: -1, Port: uint16(t.cl.incomingPeerPort()), PeerId: t.cl.peerID, InfoHash: t.infoHash, Key: t.cl.announceKey(), // The following are vaguely described in BEP 3. Left: t.bytesLeftAnnounce(), Uploaded: t.stats.BytesWrittenData.Int64(), // There's no mention of wasted or unwanted download in the BEP. Downloaded: t.stats.BytesReadUsefulData.Int64(), } }
go
func (t *Torrent) announceRequest() tracker.AnnounceRequest { // Note that IPAddress is not set. It's set for UDP inside the tracker // code, since it's dependent on the network in use. return tracker.AnnounceRequest{ Event: tracker.None, NumWant: -1, Port: uint16(t.cl.incomingPeerPort()), PeerId: t.cl.peerID, InfoHash: t.infoHash, Key: t.cl.announceKey(), // The following are vaguely described in BEP 3. Left: t.bytesLeftAnnounce(), Uploaded: t.stats.BytesWrittenData.Int64(), // There's no mention of wasted or unwanted download in the BEP. Downloaded: t.stats.BytesReadUsefulData.Int64(), } }
[ "func", "(", "t", "*", "Torrent", ")", "announceRequest", "(", ")", "tracker", ".", "AnnounceRequest", "{", "// Note that IPAddress is not set. It's set for UDP inside the tracker", "// code, since it's dependent on the network in use.", "return", "tracker", ".", "AnnounceRequest", "{", "Event", ":", "tracker", ".", "None", ",", "NumWant", ":", "-", "1", ",", "Port", ":", "uint16", "(", "t", ".", "cl", ".", "incomingPeerPort", "(", ")", ")", ",", "PeerId", ":", "t", ".", "cl", ".", "peerID", ",", "InfoHash", ":", "t", ".", "infoHash", ",", "Key", ":", "t", ".", "cl", ".", "announceKey", "(", ")", ",", "// The following are vaguely described in BEP 3.", "Left", ":", "t", ".", "bytesLeftAnnounce", "(", ")", ",", "Uploaded", ":", "t", ".", "stats", ".", "BytesWrittenData", ".", "Int64", "(", ")", ",", "// There's no mention of wasted or unwanted download in the BEP.", "Downloaded", ":", "t", ".", "stats", ".", "BytesReadUsefulData", ".", "Int64", "(", ")", ",", "}", "\n", "}" ]
// Returns an AnnounceRequest with fields filled out to defaults and current // values.
[ "Returns", "an", "AnnounceRequest", "with", "fields", "filled", "out", "to", "defaults", "and", "current", "values", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1301-L1319
train
anacrolix/torrent
torrent.go
consumeDhtAnnouncePeers
func (t *Torrent) consumeDhtAnnouncePeers(pvs <-chan dht.PeersValues) { cl := t.cl for v := range pvs { cl.lock() for _, cp := range v.Peers { if cp.Port == 0 { // Can't do anything with this. continue } t.addPeer(Peer{ IP: cp.IP[:], Port: cp.Port, Source: peerSourceDHTGetPeers, }) } cl.unlock() } }
go
func (t *Torrent) consumeDhtAnnouncePeers(pvs <-chan dht.PeersValues) { cl := t.cl for v := range pvs { cl.lock() for _, cp := range v.Peers { if cp.Port == 0 { // Can't do anything with this. continue } t.addPeer(Peer{ IP: cp.IP[:], Port: cp.Port, Source: peerSourceDHTGetPeers, }) } cl.unlock() } }
[ "func", "(", "t", "*", "Torrent", ")", "consumeDhtAnnouncePeers", "(", "pvs", "<-", "chan", "dht", ".", "PeersValues", ")", "{", "cl", ":=", "t", ".", "cl", "\n", "for", "v", ":=", "range", "pvs", "{", "cl", ".", "lock", "(", ")", "\n", "for", "_", ",", "cp", ":=", "range", "v", ".", "Peers", "{", "if", "cp", ".", "Port", "==", "0", "{", "// Can't do anything with this.", "continue", "\n", "}", "\n", "t", ".", "addPeer", "(", "Peer", "{", "IP", ":", "cp", ".", "IP", "[", ":", "]", ",", "Port", ":", "cp", ".", "Port", ",", "Source", ":", "peerSourceDHTGetPeers", ",", "}", ")", "\n", "}", "\n", "cl", ".", "unlock", "(", ")", "\n", "}", "\n", "}" ]
// Adds peers revealed in an announce until the announce ends, or we have // enough peers.
[ "Adds", "peers", "revealed", "in", "an", "announce", "until", "the", "announce", "ends", "or", "we", "have", "enough", "peers", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1323-L1340
train
anacrolix/torrent
torrent.go
numTotalPeers
func (t *Torrent) numTotalPeers() int { peers := make(map[string]struct{}) for conn := range t.conns { ra := conn.conn.RemoteAddr() if ra == nil { // It's been closed and doesn't support RemoteAddr. continue } peers[ra.String()] = struct{}{} } for addr := range t.halfOpen { peers[addr] = struct{}{} } t.peers.Each(func(peer Peer) { peers[fmt.Sprintf("%s:%d", peer.IP, peer.Port)] = struct{}{} }) return len(peers) }
go
func (t *Torrent) numTotalPeers() int { peers := make(map[string]struct{}) for conn := range t.conns { ra := conn.conn.RemoteAddr() if ra == nil { // It's been closed and doesn't support RemoteAddr. continue } peers[ra.String()] = struct{}{} } for addr := range t.halfOpen { peers[addr] = struct{}{} } t.peers.Each(func(peer Peer) { peers[fmt.Sprintf("%s:%d", peer.IP, peer.Port)] = struct{}{} }) return len(peers) }
[ "func", "(", "t", "*", "Torrent", ")", "numTotalPeers", "(", ")", "int", "{", "peers", ":=", "make", "(", "map", "[", "string", "]", "struct", "{", "}", ")", "\n", "for", "conn", ":=", "range", "t", ".", "conns", "{", "ra", ":=", "conn", ".", "conn", ".", "RemoteAddr", "(", ")", "\n", "if", "ra", "==", "nil", "{", "// It's been closed and doesn't support RemoteAddr.", "continue", "\n", "}", "\n", "peers", "[", "ra", ".", "String", "(", ")", "]", "=", "struct", "{", "}", "{", "}", "\n", "}", "\n", "for", "addr", ":=", "range", "t", ".", "halfOpen", "{", "peers", "[", "addr", "]", "=", "struct", "{", "}", "{", "}", "\n", "}", "\n", "t", ".", "peers", ".", "Each", "(", "func", "(", "peer", "Peer", ")", "{", "peers", "[", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "peer", ".", "IP", ",", "peer", ".", "Port", ")", "]", "=", "struct", "{", "}", "{", "}", "\n", "}", ")", "\n", "return", "len", "(", "peers", ")", "\n", "}" ]
// The total number of peers in the torrent.
[ "The", "total", "number", "of", "peers", "in", "the", "torrent", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1402-L1419
train
anacrolix/torrent
torrent.go
reconcileHandshakeStats
func (t *Torrent) reconcileHandshakeStats(c *connection) { if c.stats != (ConnStats{ // Handshakes should only increment these fields: BytesWritten: c.stats.BytesWritten, BytesRead: c.stats.BytesRead, }) { panic("bad stats") } c.postHandshakeStats(func(cs *ConnStats) { cs.BytesRead.Add(c.stats.BytesRead.Int64()) cs.BytesWritten.Add(c.stats.BytesWritten.Int64()) }) c.reconciledHandshakeStats = true }
go
func (t *Torrent) reconcileHandshakeStats(c *connection) { if c.stats != (ConnStats{ // Handshakes should only increment these fields: BytesWritten: c.stats.BytesWritten, BytesRead: c.stats.BytesRead, }) { panic("bad stats") } c.postHandshakeStats(func(cs *ConnStats) { cs.BytesRead.Add(c.stats.BytesRead.Int64()) cs.BytesWritten.Add(c.stats.BytesWritten.Int64()) }) c.reconciledHandshakeStats = true }
[ "func", "(", "t", "*", "Torrent", ")", "reconcileHandshakeStats", "(", "c", "*", "connection", ")", "{", "if", "c", ".", "stats", "!=", "(", "ConnStats", "{", "// Handshakes should only increment these fields:", "BytesWritten", ":", "c", ".", "stats", ".", "BytesWritten", ",", "BytesRead", ":", "c", ".", "stats", ".", "BytesRead", ",", "}", ")", "{", "panic", "(", "\"", "\"", ")", "\n", "}", "\n", "c", ".", "postHandshakeStats", "(", "func", "(", "cs", "*", "ConnStats", ")", "{", "cs", ".", "BytesRead", ".", "Add", "(", "c", ".", "stats", ".", "BytesRead", ".", "Int64", "(", ")", ")", "\n", "cs", ".", "BytesWritten", ".", "Add", "(", "c", ".", "stats", ".", "BytesWritten", ".", "Int64", "(", ")", ")", "\n", "}", ")", "\n", "c", ".", "reconciledHandshakeStats", "=", "true", "\n", "}" ]
// Reconcile bytes transferred before connection was associated with a // torrent.
[ "Reconcile", "bytes", "transferred", "before", "connection", "was", "associated", "with", "a", "torrent", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1423-L1436
train
anacrolix/torrent
torrent.go
addConnection
func (t *Torrent) addConnection(c *connection) (err error) { defer func() { if err == nil { torrent.Add("added connections", 1) } }() if t.closed.IsSet() { return errors.New("torrent closed") } for c0 := range t.conns { if c.PeerID != c0.PeerID { continue } if !t.cl.config.dropDuplicatePeerIds { continue } if left, ok := c.hasPreferredNetworkOver(c0); ok && left { c0.Close() t.deleteConnection(c0) } else { return errors.New("existing connection preferred") } } if len(t.conns) >= t.maxEstablishedConns { c := t.worstBadConn() if c == nil { return errors.New("don't want conns") } c.Close() t.deleteConnection(c) } if len(t.conns) >= t.maxEstablishedConns { panic(len(t.conns)) } t.conns[c] = struct{}{} return nil }
go
func (t *Torrent) addConnection(c *connection) (err error) { defer func() { if err == nil { torrent.Add("added connections", 1) } }() if t.closed.IsSet() { return errors.New("torrent closed") } for c0 := range t.conns { if c.PeerID != c0.PeerID { continue } if !t.cl.config.dropDuplicatePeerIds { continue } if left, ok := c.hasPreferredNetworkOver(c0); ok && left { c0.Close() t.deleteConnection(c0) } else { return errors.New("existing connection preferred") } } if len(t.conns) >= t.maxEstablishedConns { c := t.worstBadConn() if c == nil { return errors.New("don't want conns") } c.Close() t.deleteConnection(c) } if len(t.conns) >= t.maxEstablishedConns { panic(len(t.conns)) } t.conns[c] = struct{}{} return nil }
[ "func", "(", "t", "*", "Torrent", ")", "addConnection", "(", "c", "*", "connection", ")", "(", "err", "error", ")", "{", "defer", "func", "(", ")", "{", "if", "err", "==", "nil", "{", "torrent", ".", "Add", "(", "\"", "\"", ",", "1", ")", "\n", "}", "\n", "}", "(", ")", "\n", "if", "t", ".", "closed", ".", "IsSet", "(", ")", "{", "return", "errors", ".", "New", "(", "\"", "\"", ")", "\n", "}", "\n", "for", "c0", ":=", "range", "t", ".", "conns", "{", "if", "c", ".", "PeerID", "!=", "c0", ".", "PeerID", "{", "continue", "\n", "}", "\n", "if", "!", "t", ".", "cl", ".", "config", ".", "dropDuplicatePeerIds", "{", "continue", "\n", "}", "\n", "if", "left", ",", "ok", ":=", "c", ".", "hasPreferredNetworkOver", "(", "c0", ")", ";", "ok", "&&", "left", "{", "c0", ".", "Close", "(", ")", "\n", "t", ".", "deleteConnection", "(", "c0", ")", "\n", "}", "else", "{", "return", "errors", ".", "New", "(", "\"", "\"", ")", "\n", "}", "\n", "}", "\n", "if", "len", "(", "t", ".", "conns", ")", ">=", "t", ".", "maxEstablishedConns", "{", "c", ":=", "t", ".", "worstBadConn", "(", ")", "\n", "if", "c", "==", "nil", "{", "return", "errors", ".", "New", "(", "\"", "\"", ")", "\n", "}", "\n", "c", ".", "Close", "(", ")", "\n", "t", ".", "deleteConnection", "(", "c", ")", "\n", "}", "\n", "if", "len", "(", "t", ".", "conns", ")", ">=", "t", ".", "maxEstablishedConns", "{", "panic", "(", "len", "(", "t", ".", "conns", ")", ")", "\n", "}", "\n", "t", ".", "conns", "[", "c", "]", "=", "struct", "{", "}", "{", "}", "\n", "return", "nil", "\n", "}" ]
// Returns true if the connection is added.
[ "Returns", "true", "if", "the", "connection", "is", "added", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1439-L1475
train
anacrolix/torrent
torrent.go
onIncompletePiece
func (t *Torrent) onIncompletePiece(piece pieceIndex) { if t.pieceAllDirty(piece) { t.pendAllChunkSpecs(piece) } if !t.wantPieceIndex(piece) { // t.logger.Printf("piece %d incomplete and unwanted", piece) return } // We could drop any connections that we told we have a piece that we // don't here. But there's a test failure, and it seems clients don't care // if you request pieces that you already claim to have. Pruning bad // connections might just remove any connections that aren't treating us // favourably anyway. // for c := range t.conns { // if c.sentHave(piece) { // c.Drop() // } // } for conn := range t.conns { if conn.PeerHasPiece(piece) { conn.updateRequests() } } }
go
func (t *Torrent) onIncompletePiece(piece pieceIndex) { if t.pieceAllDirty(piece) { t.pendAllChunkSpecs(piece) } if !t.wantPieceIndex(piece) { // t.logger.Printf("piece %d incomplete and unwanted", piece) return } // We could drop any connections that we told we have a piece that we // don't here. But there's a test failure, and it seems clients don't care // if you request pieces that you already claim to have. Pruning bad // connections might just remove any connections that aren't treating us // favourably anyway. // for c := range t.conns { // if c.sentHave(piece) { // c.Drop() // } // } for conn := range t.conns { if conn.PeerHasPiece(piece) { conn.updateRequests() } } }
[ "func", "(", "t", "*", "Torrent", ")", "onIncompletePiece", "(", "piece", "pieceIndex", ")", "{", "if", "t", ".", "pieceAllDirty", "(", "piece", ")", "{", "t", ".", "pendAllChunkSpecs", "(", "piece", ")", "\n", "}", "\n", "if", "!", "t", ".", "wantPieceIndex", "(", "piece", ")", "{", "// t.logger.Printf(\"piece %d incomplete and unwanted\", piece)", "return", "\n", "}", "\n", "// We could drop any connections that we told we have a piece that we", "// don't here. But there's a test failure, and it seems clients don't care", "// if you request pieces that you already claim to have. Pruning bad", "// connections might just remove any connections that aren't treating us", "// favourably anyway.", "// for c := range t.conns {", "// \tif c.sentHave(piece) {", "// \t\tc.Drop()", "// \t}", "// }", "for", "conn", ":=", "range", "t", ".", "conns", "{", "if", "conn", ".", "PeerHasPiece", "(", "piece", ")", "{", "conn", ".", "updateRequests", "(", ")", "\n", "}", "\n", "}", "\n", "}" ]
// Called when a piece is found to be not complete.
[ "Called", "when", "a", "piece", "is", "found", "to", "be", "not", "complete", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1580-L1604
train
anacrolix/torrent
torrent.go
reapPieceTouchers
func (t *Torrent) reapPieceTouchers(piece pieceIndex) (ret []*connection) { for c := range t.pieces[piece].dirtiers { delete(c.peerTouchedPieces, piece) ret = append(ret, c) } t.pieces[piece].dirtiers = nil return }
go
func (t *Torrent) reapPieceTouchers(piece pieceIndex) (ret []*connection) { for c := range t.pieces[piece].dirtiers { delete(c.peerTouchedPieces, piece) ret = append(ret, c) } t.pieces[piece].dirtiers = nil return }
[ "func", "(", "t", "*", "Torrent", ")", "reapPieceTouchers", "(", "piece", "pieceIndex", ")", "(", "ret", "[", "]", "*", "connection", ")", "{", "for", "c", ":=", "range", "t", ".", "pieces", "[", "piece", "]", ".", "dirtiers", "{", "delete", "(", "c", ".", "peerTouchedPieces", ",", "piece", ")", "\n", "ret", "=", "append", "(", "ret", ",", "c", ")", "\n", "}", "\n", "t", ".", "pieces", "[", "piece", "]", ".", "dirtiers", "=", "nil", "\n", "return", "\n", "}" ]
// Return the connections that touched a piece, and clear the entries while // doing it.
[ "Return", "the", "connections", "that", "touched", "a", "piece", "and", "clear", "the", "entries", "while", "doing", "it", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1641-L1648
train
anacrolix/torrent
torrent.go
queuePieceCheck
func (t *Torrent) queuePieceCheck(pieceIndex pieceIndex) { piece := &t.pieces[pieceIndex] if piece.queuedForHash() { return } t.piecesQueuedForHash.Add(bitmap.BitIndex(pieceIndex)) t.publishPieceChange(pieceIndex) t.updatePiecePriority(pieceIndex) go t.verifyPiece(pieceIndex) }
go
func (t *Torrent) queuePieceCheck(pieceIndex pieceIndex) { piece := &t.pieces[pieceIndex] if piece.queuedForHash() { return } t.piecesQueuedForHash.Add(bitmap.BitIndex(pieceIndex)) t.publishPieceChange(pieceIndex) t.updatePiecePriority(pieceIndex) go t.verifyPiece(pieceIndex) }
[ "func", "(", "t", "*", "Torrent", ")", "queuePieceCheck", "(", "pieceIndex", "pieceIndex", ")", "{", "piece", ":=", "&", "t", ".", "pieces", "[", "pieceIndex", "]", "\n", "if", "piece", ".", "queuedForHash", "(", ")", "{", "return", "\n", "}", "\n", "t", ".", "piecesQueuedForHash", ".", "Add", "(", "bitmap", ".", "BitIndex", "(", "pieceIndex", ")", ")", "\n", "t", ".", "publishPieceChange", "(", "pieceIndex", ")", "\n", "t", ".", "updatePiecePriority", "(", "pieceIndex", ")", "\n", "go", "t", ".", "verifyPiece", "(", "pieceIndex", ")", "\n", "}" ]
// Currently doesn't really queue, but should in the future.
[ "Currently", "doesn", "t", "really", "queue", "but", "should", "in", "the", "future", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1658-L1667
train
anacrolix/torrent
torrent.go
initiateConn
func (t *Torrent) initiateConn(peer Peer) { if peer.Id == t.cl.peerID { return } if t.cl.badPeerIPPort(peer.IP, peer.Port) { return } addr := IpPort{peer.IP, uint16(peer.Port)} if t.addrActive(addr.String()) { return } t.halfOpen[addr.String()] = peer go t.cl.outgoingConnection(t, addr, peer.Source) }
go
func (t *Torrent) initiateConn(peer Peer) { if peer.Id == t.cl.peerID { return } if t.cl.badPeerIPPort(peer.IP, peer.Port) { return } addr := IpPort{peer.IP, uint16(peer.Port)} if t.addrActive(addr.String()) { return } t.halfOpen[addr.String()] = peer go t.cl.outgoingConnection(t, addr, peer.Source) }
[ "func", "(", "t", "*", "Torrent", ")", "initiateConn", "(", "peer", "Peer", ")", "{", "if", "peer", ".", "Id", "==", "t", ".", "cl", ".", "peerID", "{", "return", "\n", "}", "\n", "if", "t", ".", "cl", ".", "badPeerIPPort", "(", "peer", ".", "IP", ",", "peer", ".", "Port", ")", "{", "return", "\n", "}", "\n", "addr", ":=", "IpPort", "{", "peer", ".", "IP", ",", "uint16", "(", "peer", ".", "Port", ")", "}", "\n", "if", "t", ".", "addrActive", "(", "addr", ".", "String", "(", ")", ")", "{", "return", "\n", "}", "\n", "t", ".", "halfOpen", "[", "addr", ".", "String", "(", ")", "]", "=", "peer", "\n", "go", "t", ".", "cl", ".", "outgoingConnection", "(", "t", ",", "addr", ",", "peer", ".", "Source", ")", "\n", "}" ]
// Start the process of connecting to the given peer for the given torrent if // appropriate.
[ "Start", "the", "process", "of", "connecting", "to", "the", "given", "peer", "for", "the", "given", "torrent", "if", "appropriate", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1677-L1690
train
anacrolix/torrent
torrent.go
allStats
func (t *Torrent) allStats(f func(*ConnStats)) { f(&t.stats) f(&t.cl.stats) }
go
func (t *Torrent) allStats(f func(*ConnStats)) { f(&t.stats) f(&t.cl.stats) }
[ "func", "(", "t", "*", "Torrent", ")", "allStats", "(", "f", "func", "(", "*", "ConnStats", ")", ")", "{", "f", "(", "&", "t", ".", "stats", ")", "\n", "f", "(", "&", "t", ".", "cl", ".", "stats", ")", "\n", "}" ]
// All stats that include this Torrent. Useful when we want to increment // ConnStats but not for every connection.
[ "All", "stats", "that", "include", "this", "Torrent", ".", "Useful", "when", "we", "want", "to", "increment", "ConnStats", "but", "not", "for", "every", "connection", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/torrent.go#L1706-L1709
train
anacrolix/torrent
metainfo/info.go
BuildFromFilePath
func (info *Info) BuildFromFilePath(root string) (err error) { info.Name = filepath.Base(root) info.Files = nil err = filepath.Walk(root, func(path string, fi os.FileInfo, err error) error { if err != nil { return err } if fi.IsDir() { // Directories are implicit in torrent files. return nil } else if path == root { // The root is a file. info.Length = fi.Size() return nil } relPath, err := filepath.Rel(root, path) if err != nil { return fmt.Errorf("error getting relative path: %s", err) } info.Files = append(info.Files, FileInfo{ Path: strings.Split(relPath, string(filepath.Separator)), Length: fi.Size(), }) return nil }) if err != nil { return } slices.Sort(info.Files, func(l, r FileInfo) bool { return strings.Join(l.Path, "/") < strings.Join(r.Path, "/") }) err = info.GeneratePieces(func(fi FileInfo) (io.ReadCloser, error) { return os.Open(filepath.Join(root, strings.Join(fi.Path, string(filepath.Separator)))) }) if err != nil { err = fmt.Errorf("error generating pieces: %s", err) } return }
go
func (info *Info) BuildFromFilePath(root string) (err error) { info.Name = filepath.Base(root) info.Files = nil err = filepath.Walk(root, func(path string, fi os.FileInfo, err error) error { if err != nil { return err } if fi.IsDir() { // Directories are implicit in torrent files. return nil } else if path == root { // The root is a file. info.Length = fi.Size() return nil } relPath, err := filepath.Rel(root, path) if err != nil { return fmt.Errorf("error getting relative path: %s", err) } info.Files = append(info.Files, FileInfo{ Path: strings.Split(relPath, string(filepath.Separator)), Length: fi.Size(), }) return nil }) if err != nil { return } slices.Sort(info.Files, func(l, r FileInfo) bool { return strings.Join(l.Path, "/") < strings.Join(r.Path, "/") }) err = info.GeneratePieces(func(fi FileInfo) (io.ReadCloser, error) { return os.Open(filepath.Join(root, strings.Join(fi.Path, string(filepath.Separator)))) }) if err != nil { err = fmt.Errorf("error generating pieces: %s", err) } return }
[ "func", "(", "info", "*", "Info", ")", "BuildFromFilePath", "(", "root", "string", ")", "(", "err", "error", ")", "{", "info", ".", "Name", "=", "filepath", ".", "Base", "(", "root", ")", "\n", "info", ".", "Files", "=", "nil", "\n", "err", "=", "filepath", ".", "Walk", "(", "root", ",", "func", "(", "path", "string", ",", "fi", "os", ".", "FileInfo", ",", "err", "error", ")", "error", "{", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "fi", ".", "IsDir", "(", ")", "{", "// Directories are implicit in torrent files.", "return", "nil", "\n", "}", "else", "if", "path", "==", "root", "{", "// The root is a file.", "info", ".", "Length", "=", "fi", ".", "Size", "(", ")", "\n", "return", "nil", "\n", "}", "\n", "relPath", ",", "err", ":=", "filepath", ".", "Rel", "(", "root", ",", "path", ")", "\n", "if", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "info", ".", "Files", "=", "append", "(", "info", ".", "Files", ",", "FileInfo", "{", "Path", ":", "strings", ".", "Split", "(", "relPath", ",", "string", "(", "filepath", ".", "Separator", ")", ")", ",", "Length", ":", "fi", ".", "Size", "(", ")", ",", "}", ")", "\n", "return", "nil", "\n", "}", ")", "\n", "if", "err", "!=", "nil", "{", "return", "\n", "}", "\n", "slices", ".", "Sort", "(", "info", ".", "Files", ",", "func", "(", "l", ",", "r", "FileInfo", ")", "bool", "{", "return", "strings", ".", "Join", "(", "l", ".", "Path", ",", "\"", "\"", ")", "<", "strings", ".", "Join", "(", "r", ".", "Path", ",", "\"", "\"", ")", "\n", "}", ")", "\n", "err", "=", "info", ".", "GeneratePieces", "(", "func", "(", "fi", "FileInfo", ")", "(", "io", ".", "ReadCloser", ",", "error", ")", "{", "return", "os", ".", "Open", "(", "filepath", ".", "Join", "(", "root", ",", "strings", ".", "Join", "(", "fi", ".", "Path", ",", "string", "(", "filepath", ".", "Separator", ")", ")", ")", ")", "\n", "}", ")", "\n", "if", "err", "!=", "nil", "{", "err", "=", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "return", "\n", "}" ]
// This is a helper that sets Files and Pieces from a root path and its // children.
[ "This", "is", "a", "helper", "that", "sets", "Files", "and", "Pieces", "from", "a", "root", "path", "and", "its", "children", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/metainfo/info.go#L29-L67
train
anacrolix/torrent
metainfo/info.go
writeFiles
func (info *Info) writeFiles(w io.Writer, open func(fi FileInfo) (io.ReadCloser, error)) error { for _, fi := range info.UpvertedFiles() { r, err := open(fi) if err != nil { return fmt.Errorf("error opening %v: %s", fi, err) } wn, err := io.CopyN(w, r, fi.Length) r.Close() if wn != fi.Length { return fmt.Errorf("error copying %v: %s", fi, err) } } return nil }
go
func (info *Info) writeFiles(w io.Writer, open func(fi FileInfo) (io.ReadCloser, error)) error { for _, fi := range info.UpvertedFiles() { r, err := open(fi) if err != nil { return fmt.Errorf("error opening %v: %s", fi, err) } wn, err := io.CopyN(w, r, fi.Length) r.Close() if wn != fi.Length { return fmt.Errorf("error copying %v: %s", fi, err) } } return nil }
[ "func", "(", "info", "*", "Info", ")", "writeFiles", "(", "w", "io", ".", "Writer", ",", "open", "func", "(", "fi", "FileInfo", ")", "(", "io", ".", "ReadCloser", ",", "error", ")", ")", "error", "{", "for", "_", ",", "fi", ":=", "range", "info", ".", "UpvertedFiles", "(", ")", "{", "r", ",", "err", ":=", "open", "(", "fi", ")", "\n", "if", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "fi", ",", "err", ")", "\n", "}", "\n", "wn", ",", "err", ":=", "io", ".", "CopyN", "(", "w", ",", "r", ",", "fi", ".", "Length", ")", "\n", "r", ".", "Close", "(", ")", "\n", "if", "wn", "!=", "fi", ".", "Length", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "fi", ",", "err", ")", "\n", "}", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// Concatenates all the files in the torrent into w. open is a function that // gets at the contents of the given file.
[ "Concatenates", "all", "the", "files", "in", "the", "torrent", "into", "w", ".", "open", "is", "a", "function", "that", "gets", "at", "the", "contents", "of", "the", "given", "file", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/metainfo/info.go#L71-L84
train
anacrolix/torrent
metainfo/info.go
UpvertedFiles
func (info *Info) UpvertedFiles() []FileInfo { if len(info.Files) == 0 { return []FileInfo{{ Length: info.Length, // Callers should determine that Info.Name is the basename, and // thus a regular file. Path: nil, }} } return info.Files }
go
func (info *Info) UpvertedFiles() []FileInfo { if len(info.Files) == 0 { return []FileInfo{{ Length: info.Length, // Callers should determine that Info.Name is the basename, and // thus a regular file. Path: nil, }} } return info.Files }
[ "func", "(", "info", "*", "Info", ")", "UpvertedFiles", "(", ")", "[", "]", "FileInfo", "{", "if", "len", "(", "info", ".", "Files", ")", "==", "0", "{", "return", "[", "]", "FileInfo", "{", "{", "Length", ":", "info", ".", "Length", ",", "// Callers should determine that Info.Name is the basename, and", "// thus a regular file.", "Path", ":", "nil", ",", "}", "}", "\n", "}", "\n", "return", "info", ".", "Files", "\n", "}" ]
// The files field, converted up from the old single-file in the parent info // dict if necessary. This is a helper to avoid having to conditionally handle // single and multi-file torrent infos.
[ "The", "files", "field", "converted", "up", "from", "the", "old", "single", "-", "file", "in", "the", "parent", "info", "dict", "if", "necessary", ".", "This", "is", "a", "helper", "to", "avoid", "having", "to", "conditionally", "handle", "single", "and", "multi", "-", "file", "torrent", "infos", "." ]
d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb
https://github.com/anacrolix/torrent/blob/d2400968fc54c3d8a6482d4ca5ff16cc5e5139cb/metainfo/info.go#L142-L152
train
kubernetes-retired/heapster
metrics/util/label_copier.go
makeStoredLabels
func makeStoredLabels(labels []string) map[string]string { storedLabels := make(map[string]string) for _, s := range labels { split := strings.SplitN(s, "=", 2) if len(split) == 1 { storedLabels[split[0]] = split[0] } else { storedLabels[split[1]] = split[0] } } return storedLabels }
go
func makeStoredLabels(labels []string) map[string]string { storedLabels := make(map[string]string) for _, s := range labels { split := strings.SplitN(s, "=", 2) if len(split) == 1 { storedLabels[split[0]] = split[0] } else { storedLabels[split[1]] = split[0] } } return storedLabels }
[ "func", "makeStoredLabels", "(", "labels", "[", "]", "string", ")", "map", "[", "string", "]", "string", "{", "storedLabels", ":=", "make", "(", "map", "[", "string", "]", "string", ")", "\n", "for", "_", ",", "s", ":=", "range", "labels", "{", "split", ":=", "strings", ".", "SplitN", "(", "s", ",", "\"", "\"", ",", "2", ")", "\n", "if", "len", "(", "split", ")", "==", "1", "{", "storedLabels", "[", "split", "[", "0", "]", "]", "=", "split", "[", "0", "]", "\n", "}", "else", "{", "storedLabels", "[", "split", "[", "1", "]", "]", "=", "split", "[", "0", "]", "\n", "}", "\n", "}", "\n", "return", "storedLabels", "\n", "}" ]
// makeStoredLabels converts labels into a map for quicker retrieval. // Incoming labels, if desired, may contain mappings in format "newName=oldName"
[ "makeStoredLabels", "converts", "labels", "into", "a", "map", "for", "quicker", "retrieval", ".", "Incoming", "labels", "if", "desired", "may", "contain", "mappings", "in", "format", "newName", "=", "oldName" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/util/label_copier.go#L56-L67
train
kubernetes-retired/heapster
metrics/util/label_copier.go
makeIgnoredLabels
func makeIgnoredLabels(labels []string) map[string]string { ignoredLabels := make(map[string]string) for _, s := range labels { ignoredLabels[s] = "" } return ignoredLabels }
go
func makeIgnoredLabels(labels []string) map[string]string { ignoredLabels := make(map[string]string) for _, s := range labels { ignoredLabels[s] = "" } return ignoredLabels }
[ "func", "makeIgnoredLabels", "(", "labels", "[", "]", "string", ")", "map", "[", "string", "]", "string", "{", "ignoredLabels", ":=", "make", "(", "map", "[", "string", "]", "string", ")", "\n", "for", "_", ",", "s", ":=", "range", "labels", "{", "ignoredLabels", "[", "s", "]", "=", "\"", "\"", "\n", "}", "\n", "return", "ignoredLabels", "\n", "}" ]
// makeIgnoredLabels converts label slice into a map for later use.
[ "makeIgnoredLabels", "converts", "label", "slice", "into", "a", "map", "for", "later", "use", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/util/label_copier.go#L70-L76
train
kubernetes-retired/heapster
metrics/util/label_copier.go
NewLabelCopier
func NewLabelCopier(separator string, storedLabels, ignoredLabels []string) (*LabelCopier, error) { return &LabelCopier{ labelSeparator: separator, storedLabels: makeStoredLabels(storedLabels), ignoredLabels: makeIgnoredLabels(ignoredLabels), }, nil }
go
func NewLabelCopier(separator string, storedLabels, ignoredLabels []string) (*LabelCopier, error) { return &LabelCopier{ labelSeparator: separator, storedLabels: makeStoredLabels(storedLabels), ignoredLabels: makeIgnoredLabels(ignoredLabels), }, nil }
[ "func", "NewLabelCopier", "(", "separator", "string", ",", "storedLabels", ",", "ignoredLabels", "[", "]", "string", ")", "(", "*", "LabelCopier", ",", "error", ")", "{", "return", "&", "LabelCopier", "{", "labelSeparator", ":", "separator", ",", "storedLabels", ":", "makeStoredLabels", "(", "storedLabels", ")", ",", "ignoredLabels", ":", "makeIgnoredLabels", "(", "ignoredLabels", ")", ",", "}", ",", "nil", "\n", "}" ]
// NewLabelCopier creates a new instance of LabelCopier type
[ "NewLabelCopier", "creates", "a", "new", "instance", "of", "LabelCopier", "type" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/util/label_copier.go#L79-L85
train
kubernetes-retired/heapster
events/sinks/influxdb/influxdb.go
getEventValue
func getEventValue(event *kube_api.Event) (string, error) { // TODO: check whether indenting is required. bytes, err := json.MarshalIndent(event, "", " ") if err != nil { return "", err } return string(bytes), nil }
go
func getEventValue(event *kube_api.Event) (string, error) { // TODO: check whether indenting is required. bytes, err := json.MarshalIndent(event, "", " ") if err != nil { return "", err } return string(bytes), nil }
[ "func", "getEventValue", "(", "event", "*", "kube_api", ".", "Event", ")", "(", "string", ",", "error", ")", "{", "// TODO: check whether indenting is required.", "bytes", ",", "err", ":=", "json", ".", "MarshalIndent", "(", "event", ",", "\"", "\"", ",", "\"", "\"", ")", "\n", "if", "err", "!=", "nil", "{", "return", "\"", "\"", ",", "err", "\n", "}", "\n", "return", "string", "(", "bytes", ")", ",", "nil", "\n", "}" ]
// Generate point value for event
[ "Generate", "point", "value", "for", "event" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/events/sinks/influxdb/influxdb.go#L61-L68
train
kubernetes-retired/heapster
events/sinks/influxdb/influxdb.go
newSink
func newSink(c influxdb_common.InfluxdbConfig) core.EventSink { client, err := influxdb_common.NewClient(c) if err != nil { glog.Errorf("issues while creating an InfluxDB sink: %v, will retry on use", err) } return &influxdbSink{ client: client, // can be nil c: c, } }
go
func newSink(c influxdb_common.InfluxdbConfig) core.EventSink { client, err := influxdb_common.NewClient(c) if err != nil { glog.Errorf("issues while creating an InfluxDB sink: %v, will retry on use", err) } return &influxdbSink{ client: client, // can be nil c: c, } }
[ "func", "newSink", "(", "c", "influxdb_common", ".", "InfluxdbConfig", ")", "core", ".", "EventSink", "{", "client", ",", "err", ":=", "influxdb_common", ".", "NewClient", "(", "c", ")", "\n", "if", "err", "!=", "nil", "{", "glog", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "return", "&", "influxdbSink", "{", "client", ":", "client", ",", "// can be nil", "c", ":", "c", ",", "}", "\n", "}" ]
// Returns a thread-safe implementation of core.EventSink for InfluxDB.
[ "Returns", "a", "thread", "-", "safe", "implementation", "of", "core", ".", "EventSink", "for", "InfluxDB", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/events/sinks/influxdb/influxdb.go#L230-L239
train
kubernetes-retired/heapster
metrics/sinks/opentsdb/driver.go
metricToPoint
func (tsdbSink *openTSDBSink) metricToPoint(name string, value core.MetricValue, timestamp time.Time, labels map[string]string) opentsdbclient.DataPoint { seriesName := strings.Replace(toValidOpenTsdbName(name), "/", "_", -1) if value.MetricType.String() != "" { seriesName = fmt.Sprintf("%s_%s", seriesName, value.MetricType.String()) } datapoint := opentsdbclient.DataPoint{ Metric: seriesName, Tags: make(map[string]string, len(labels)), Timestamp: timestamp.Unix(), } if value.ValueType == core.ValueInt64 { datapoint.Value = value.IntValue } else { datapoint.Value = value.FloatValue } for key, value := range labels { key = toValidOpenTsdbName(key) value = toValidOpenTsdbName(value) if value != "" { datapoint.Tags[key] = value } } tsdbSink.putDefaultTags(&datapoint) return datapoint }
go
func (tsdbSink *openTSDBSink) metricToPoint(name string, value core.MetricValue, timestamp time.Time, labels map[string]string) opentsdbclient.DataPoint { seriesName := strings.Replace(toValidOpenTsdbName(name), "/", "_", -1) if value.MetricType.String() != "" { seriesName = fmt.Sprintf("%s_%s", seriesName, value.MetricType.String()) } datapoint := opentsdbclient.DataPoint{ Metric: seriesName, Tags: make(map[string]string, len(labels)), Timestamp: timestamp.Unix(), } if value.ValueType == core.ValueInt64 { datapoint.Value = value.IntValue } else { datapoint.Value = value.FloatValue } for key, value := range labels { key = toValidOpenTsdbName(key) value = toValidOpenTsdbName(value) if value != "" { datapoint.Tags[key] = value } } tsdbSink.putDefaultTags(&datapoint) return datapoint }
[ "func", "(", "tsdbSink", "*", "openTSDBSink", ")", "metricToPoint", "(", "name", "string", ",", "value", "core", ".", "MetricValue", ",", "timestamp", "time", ".", "Time", ",", "labels", "map", "[", "string", "]", "string", ")", "opentsdbclient", ".", "DataPoint", "{", "seriesName", ":=", "strings", ".", "Replace", "(", "toValidOpenTsdbName", "(", "name", ")", ",", "\"", "\"", ",", "\"", "\"", ",", "-", "1", ")", "\n\n", "if", "value", ".", "MetricType", ".", "String", "(", ")", "!=", "\"", "\"", "{", "seriesName", "=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "seriesName", ",", "value", ".", "MetricType", ".", "String", "(", ")", ")", "\n", "}", "\n\n", "datapoint", ":=", "opentsdbclient", ".", "DataPoint", "{", "Metric", ":", "seriesName", ",", "Tags", ":", "make", "(", "map", "[", "string", "]", "string", ",", "len", "(", "labels", ")", ")", ",", "Timestamp", ":", "timestamp", ".", "Unix", "(", ")", ",", "}", "\n", "if", "value", ".", "ValueType", "==", "core", ".", "ValueInt64", "{", "datapoint", ".", "Value", "=", "value", ".", "IntValue", "\n", "}", "else", "{", "datapoint", ".", "Value", "=", "value", ".", "FloatValue", "\n", "}", "\n\n", "for", "key", ",", "value", ":=", "range", "labels", "{", "key", "=", "toValidOpenTsdbName", "(", "key", ")", "\n", "value", "=", "toValidOpenTsdbName", "(", "value", ")", "\n\n", "if", "value", "!=", "\"", "\"", "{", "datapoint", ".", "Tags", "[", "key", "]", "=", "value", "\n", "}", "\n", "}", "\n\n", "tsdbSink", ".", "putDefaultTags", "(", "&", "datapoint", ")", "\n", "return", "datapoint", "\n", "}" ]
// timeSeriesToPoint transfers the contents holding in the given pointer of sink_api.Timeseries // into the instance of opentsdbclient.DataPoint
[ "timeSeriesToPoint", "transfers", "the", "contents", "holding", "in", "the", "given", "pointer", "of", "sink_api", ".", "Timeseries", "into", "the", "instance", "of", "opentsdbclient", ".", "DataPoint" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/opentsdb/driver.go#L117-L146
train
kubernetes-retired/heapster
metrics/sinks/opentsdb/driver.go
putDefaultTags
func (tsdbSink *openTSDBSink) putDefaultTags(datapoint *opentsdbclient.DataPoint) { datapoint.Tags[clusterNameTagName] = tsdbSink.clusterName }
go
func (tsdbSink *openTSDBSink) putDefaultTags(datapoint *opentsdbclient.DataPoint) { datapoint.Tags[clusterNameTagName] = tsdbSink.clusterName }
[ "func", "(", "tsdbSink", "*", "openTSDBSink", ")", "putDefaultTags", "(", "datapoint", "*", "opentsdbclient", ".", "DataPoint", ")", "{", "datapoint", ".", "Tags", "[", "clusterNameTagName", "]", "=", "tsdbSink", ".", "clusterName", "\n", "}" ]
// putDefaultTags just fills in the default key-value pair for the tags. // OpenTSDB requires at least one non-empty tag otherwise the OpenTSDB will return error and the operation of putting // datapoint will be failed.
[ "putDefaultTags", "just", "fills", "in", "the", "default", "key", "-", "value", "pair", "for", "the", "tags", ".", "OpenTSDB", "requires", "at", "least", "one", "non", "-", "empty", "tag", "otherwise", "the", "OpenTSDB", "will", "return", "error", "and", "the", "operation", "of", "putting", "datapoint", "will", "be", "failed", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/opentsdb/driver.go#L151-L153
train
kubernetes-retired/heapster
metrics/api/v1/api.go
NewApi
func NewApi(runningInKubernetes bool, metricSink *metricsink.MetricSink, historicalSource core.HistoricalSource, disableMetricExport bool) *Api { gkeMetrics := make(map[string]core.MetricDescriptor) gkeLabels := make(map[string]core.LabelDescriptor) for _, val := range core.StandardMetrics { gkeMetrics[val.Name] = val.MetricDescriptor } for _, val := range core.LabeledMetrics { gkeMetrics[val.Name] = val.MetricDescriptor } gkeMetrics[core.MetricCpuLimit.Name] = core.MetricCpuLimit.MetricDescriptor gkeMetrics[core.MetricMemoryLimit.Name] = core.MetricMemoryLimit.MetricDescriptor for _, val := range core.CommonLabels() { gkeLabels[val.Key] = val } for _, val := range core.ContainerLabels() { gkeLabels[val.Key] = val } for _, val := range core.PodLabels() { gkeLabels[val.Key] = val } return &Api{ runningInKubernetes: runningInKubernetes, metricSink: metricSink, historicalSource: historicalSource, gkeMetrics: gkeMetrics, gkeLabels: gkeLabels, disabled: disableMetricExport, } }
go
func NewApi(runningInKubernetes bool, metricSink *metricsink.MetricSink, historicalSource core.HistoricalSource, disableMetricExport bool) *Api { gkeMetrics := make(map[string]core.MetricDescriptor) gkeLabels := make(map[string]core.LabelDescriptor) for _, val := range core.StandardMetrics { gkeMetrics[val.Name] = val.MetricDescriptor } for _, val := range core.LabeledMetrics { gkeMetrics[val.Name] = val.MetricDescriptor } gkeMetrics[core.MetricCpuLimit.Name] = core.MetricCpuLimit.MetricDescriptor gkeMetrics[core.MetricMemoryLimit.Name] = core.MetricMemoryLimit.MetricDescriptor for _, val := range core.CommonLabels() { gkeLabels[val.Key] = val } for _, val := range core.ContainerLabels() { gkeLabels[val.Key] = val } for _, val := range core.PodLabels() { gkeLabels[val.Key] = val } return &Api{ runningInKubernetes: runningInKubernetes, metricSink: metricSink, historicalSource: historicalSource, gkeMetrics: gkeMetrics, gkeLabels: gkeLabels, disabled: disableMetricExport, } }
[ "func", "NewApi", "(", "runningInKubernetes", "bool", ",", "metricSink", "*", "metricsink", ".", "MetricSink", ",", "historicalSource", "core", ".", "HistoricalSource", ",", "disableMetricExport", "bool", ")", "*", "Api", "{", "gkeMetrics", ":=", "make", "(", "map", "[", "string", "]", "core", ".", "MetricDescriptor", ")", "\n", "gkeLabels", ":=", "make", "(", "map", "[", "string", "]", "core", ".", "LabelDescriptor", ")", "\n", "for", "_", ",", "val", ":=", "range", "core", ".", "StandardMetrics", "{", "gkeMetrics", "[", "val", ".", "Name", "]", "=", "val", ".", "MetricDescriptor", "\n", "}", "\n", "for", "_", ",", "val", ":=", "range", "core", ".", "LabeledMetrics", "{", "gkeMetrics", "[", "val", ".", "Name", "]", "=", "val", ".", "MetricDescriptor", "\n", "}", "\n", "gkeMetrics", "[", "core", ".", "MetricCpuLimit", ".", "Name", "]", "=", "core", ".", "MetricCpuLimit", ".", "MetricDescriptor", "\n", "gkeMetrics", "[", "core", ".", "MetricMemoryLimit", ".", "Name", "]", "=", "core", ".", "MetricMemoryLimit", ".", "MetricDescriptor", "\n\n", "for", "_", ",", "val", ":=", "range", "core", ".", "CommonLabels", "(", ")", "{", "gkeLabels", "[", "val", ".", "Key", "]", "=", "val", "\n", "}", "\n", "for", "_", ",", "val", ":=", "range", "core", ".", "ContainerLabels", "(", ")", "{", "gkeLabels", "[", "val", ".", "Key", "]", "=", "val", "\n", "}", "\n", "for", "_", ",", "val", ":=", "range", "core", ".", "PodLabels", "(", ")", "{", "gkeLabels", "[", "val", ".", "Key", "]", "=", "val", "\n", "}", "\n\n", "return", "&", "Api", "{", "runningInKubernetes", ":", "runningInKubernetes", ",", "metricSink", ":", "metricSink", ",", "historicalSource", ":", "historicalSource", ",", "gkeMetrics", ":", "gkeMetrics", ",", "gkeLabels", ":", "gkeLabels", ",", "disabled", ":", "disableMetricExport", ",", "}", "\n", "}" ]
// Create a new Api to serve from the specified cache.
[ "Create", "a", "new", "Api", "to", "serve", "from", "the", "specified", "cache", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/api.go#L42-L72
train
kubernetes-retired/heapster
metrics/api/v1/api.go
Register
func (a *Api) Register(container *restful.Container) { ws := new(restful.WebService) ws.Path("/api/v1/metric-export"). Doc("Exports the latest point for all Heapster metrics"). Produces(restful.MIME_JSON) ws.Route(ws.GET(""). To(a.exportMetrics). Doc("export the latest data point for all metrics"). Operation("exportMetrics"). Writes([]*types.Timeseries{})) container.Add(ws) ws = new(restful.WebService) ws.Path("/api/v1/metric-export-schema"). Doc("Schema for metrics exported by heapster"). Produces(restful.MIME_JSON) ws.Route(ws.GET(""). To(a.exportMetricsSchema). Doc("export the schema for all metrics"). Operation("exportmetricsSchema"). Writes(types.TimeseriesSchema{})) container.Add(ws) if a.metricSink != nil { a.RegisterModel(container) } if a.historicalSource != nil { a.RegisterHistorical(container) } }
go
func (a *Api) Register(container *restful.Container) { ws := new(restful.WebService) ws.Path("/api/v1/metric-export"). Doc("Exports the latest point for all Heapster metrics"). Produces(restful.MIME_JSON) ws.Route(ws.GET(""). To(a.exportMetrics). Doc("export the latest data point for all metrics"). Operation("exportMetrics"). Writes([]*types.Timeseries{})) container.Add(ws) ws = new(restful.WebService) ws.Path("/api/v1/metric-export-schema"). Doc("Schema for metrics exported by heapster"). Produces(restful.MIME_JSON) ws.Route(ws.GET(""). To(a.exportMetricsSchema). Doc("export the schema for all metrics"). Operation("exportmetricsSchema"). Writes(types.TimeseriesSchema{})) container.Add(ws) if a.metricSink != nil { a.RegisterModel(container) } if a.historicalSource != nil { a.RegisterHistorical(container) } }
[ "func", "(", "a", "*", "Api", ")", "Register", "(", "container", "*", "restful", ".", "Container", ")", "{", "ws", ":=", "new", "(", "restful", ".", "WebService", ")", "\n", "ws", ".", "Path", "(", "\"", "\"", ")", ".", "Doc", "(", "\"", "\"", ")", ".", "Produces", "(", "restful", ".", "MIME_JSON", ")", "\n", "ws", ".", "Route", "(", "ws", ".", "GET", "(", "\"", "\"", ")", ".", "To", "(", "a", ".", "exportMetrics", ")", ".", "Doc", "(", "\"", "\"", ")", ".", "Operation", "(", "\"", "\"", ")", ".", "Writes", "(", "[", "]", "*", "types", ".", "Timeseries", "{", "}", ")", ")", "\n", "container", ".", "Add", "(", "ws", ")", "\n", "ws", "=", "new", "(", "restful", ".", "WebService", ")", "\n", "ws", ".", "Path", "(", "\"", "\"", ")", ".", "Doc", "(", "\"", "\"", ")", ".", "Produces", "(", "restful", ".", "MIME_JSON", ")", "\n", "ws", ".", "Route", "(", "ws", ".", "GET", "(", "\"", "\"", ")", ".", "To", "(", "a", ".", "exportMetricsSchema", ")", ".", "Doc", "(", "\"", "\"", ")", ".", "Operation", "(", "\"", "\"", ")", ".", "Writes", "(", "types", ".", "TimeseriesSchema", "{", "}", ")", ")", "\n", "container", ".", "Add", "(", "ws", ")", "\n\n", "if", "a", ".", "metricSink", "!=", "nil", "{", "a", ".", "RegisterModel", "(", "container", ")", "\n", "}", "\n\n", "if", "a", ".", "historicalSource", "!=", "nil", "{", "a", ".", "RegisterHistorical", "(", "container", ")", "\n", "}", "\n", "}" ]
// Register the mainApi on the specified endpoint.
[ "Register", "the", "mainApi", "on", "the", "specified", "endpoint", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/api.go#L75-L104
train
kubernetes-retired/heapster
metrics/sinks/hawkular/driver.go
NewHawkularSink
func NewHawkularSink(u *url.URL) (core.DataSink, error) { sink := &hawkularSink{ uri: u, batchSize: 1000, } if err := sink.init(); err != nil { return nil, err } metrics := make([]core.MetricDescriptor, 0, len(core.AllMetrics)) for _, metric := range core.AllMetrics { metrics = append(metrics, metric.MetricDescriptor) } sink.Register(metrics) return sink, nil }
go
func NewHawkularSink(u *url.URL) (core.DataSink, error) { sink := &hawkularSink{ uri: u, batchSize: 1000, } if err := sink.init(); err != nil { return nil, err } metrics := make([]core.MetricDescriptor, 0, len(core.AllMetrics)) for _, metric := range core.AllMetrics { metrics = append(metrics, metric.MetricDescriptor) } sink.Register(metrics) return sink, nil }
[ "func", "NewHawkularSink", "(", "u", "*", "url", ".", "URL", ")", "(", "core", ".", "DataSink", ",", "error", ")", "{", "sink", ":=", "&", "hawkularSink", "{", "uri", ":", "u", ",", "batchSize", ":", "1000", ",", "}", "\n", "if", "err", ":=", "sink", ".", "init", "(", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "metrics", ":=", "make", "(", "[", "]", "core", ".", "MetricDescriptor", ",", "0", ",", "len", "(", "core", ".", "AllMetrics", ")", ")", "\n", "for", "_", ",", "metric", ":=", "range", "core", ".", "AllMetrics", "{", "metrics", "=", "append", "(", "metrics", ",", "metric", ".", "MetricDescriptor", ")", "\n", "}", "\n", "sink", ".", "Register", "(", "metrics", ")", "\n", "return", "sink", ",", "nil", "\n", "}" ]
// NewHawkularSink Creates and returns a new hawkularSink instance
[ "NewHawkularSink", "Creates", "and", "returns", "a", "new", "hawkularSink", "instance" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/hawkular/driver.go#L178-L193
train
kubernetes-retired/heapster
common/elasticsearch/elasticsearch.go
SaveData
func (esSvc *ElasticSearchService) SaveData(date time.Time, typeName string, sinkData []interface{}) error { if typeName == "" || len(sinkData) == 0 { return nil } indexName := esSvc.Index(date) // Use the IndexExists service to check if a specified index exists. exists, err := esSvc.EsClient.IndexExists(indexName) if err != nil { return err } if !exists { // Create a new index. createIndex, err := esSvc.EsClient.CreateIndex(indexName, mapping) if err != nil { return err } ack := false switch i := createIndex.(type) { case *elastic2.IndicesCreateResult: ack = i.Acknowledged case *elastic5.IndicesCreateResult: ack = i.Acknowledged } if !ack { return errors.New("Failed to acknoledge index creation") } } aliases, err := esSvc.EsClient.GetAliases(indexName) if err != nil { return err } aliasName := esSvc.IndexAlias(typeName) hasAlias := false switch a := aliases.(type) { case *elastic2.AliasesResult: hasAlias = a.Indices[indexName].HasAlias(aliasName) case *elastic5.AliasesResult: hasAlias = a.Indices[indexName].HasAlias(aliasName) } if !hasAlias { createAlias, err := esSvc.EsClient.AddAlias(indexName, esSvc.IndexAlias(typeName)) if err != nil { return err } ack := false switch i := createAlias.(type) { case *elastic2.AliasResult: ack = i.Acknowledged case *elastic5.AliasResult: ack = i.Acknowledged } if !ack { return errors.New("Failed to acknoledge index alias creation") } } for _, data := range sinkData { esSvc.EsClient.AddBulkReq(indexName, typeName, data) } return nil }
go
func (esSvc *ElasticSearchService) SaveData(date time.Time, typeName string, sinkData []interface{}) error { if typeName == "" || len(sinkData) == 0 { return nil } indexName := esSvc.Index(date) // Use the IndexExists service to check if a specified index exists. exists, err := esSvc.EsClient.IndexExists(indexName) if err != nil { return err } if !exists { // Create a new index. createIndex, err := esSvc.EsClient.CreateIndex(indexName, mapping) if err != nil { return err } ack := false switch i := createIndex.(type) { case *elastic2.IndicesCreateResult: ack = i.Acknowledged case *elastic5.IndicesCreateResult: ack = i.Acknowledged } if !ack { return errors.New("Failed to acknoledge index creation") } } aliases, err := esSvc.EsClient.GetAliases(indexName) if err != nil { return err } aliasName := esSvc.IndexAlias(typeName) hasAlias := false switch a := aliases.(type) { case *elastic2.AliasesResult: hasAlias = a.Indices[indexName].HasAlias(aliasName) case *elastic5.AliasesResult: hasAlias = a.Indices[indexName].HasAlias(aliasName) } if !hasAlias { createAlias, err := esSvc.EsClient.AddAlias(indexName, esSvc.IndexAlias(typeName)) if err != nil { return err } ack := false switch i := createAlias.(type) { case *elastic2.AliasResult: ack = i.Acknowledged case *elastic5.AliasResult: ack = i.Acknowledged } if !ack { return errors.New("Failed to acknoledge index alias creation") } } for _, data := range sinkData { esSvc.EsClient.AddBulkReq(indexName, typeName, data) } return nil }
[ "func", "(", "esSvc", "*", "ElasticSearchService", ")", "SaveData", "(", "date", "time", ".", "Time", ",", "typeName", "string", ",", "sinkData", "[", "]", "interface", "{", "}", ")", "error", "{", "if", "typeName", "==", "\"", "\"", "||", "len", "(", "sinkData", ")", "==", "0", "{", "return", "nil", "\n", "}", "\n\n", "indexName", ":=", "esSvc", ".", "Index", "(", "date", ")", "\n\n", "// Use the IndexExists service to check if a specified index exists.", "exists", ",", "err", ":=", "esSvc", ".", "EsClient", ".", "IndexExists", "(", "indexName", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "if", "!", "exists", "{", "// Create a new index.", "createIndex", ",", "err", ":=", "esSvc", ".", "EsClient", ".", "CreateIndex", "(", "indexName", ",", "mapping", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "ack", ":=", "false", "\n", "switch", "i", ":=", "createIndex", ".", "(", "type", ")", "{", "case", "*", "elastic2", ".", "IndicesCreateResult", ":", "ack", "=", "i", ".", "Acknowledged", "\n", "case", "*", "elastic5", ".", "IndicesCreateResult", ":", "ack", "=", "i", ".", "Acknowledged", "\n", "}", "\n", "if", "!", "ack", "{", "return", "errors", ".", "New", "(", "\"", "\"", ")", "\n", "}", "\n", "}", "\n\n", "aliases", ",", "err", ":=", "esSvc", ".", "EsClient", ".", "GetAliases", "(", "indexName", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "aliasName", ":=", "esSvc", ".", "IndexAlias", "(", "typeName", ")", "\n\n", "hasAlias", ":=", "false", "\n", "switch", "a", ":=", "aliases", ".", "(", "type", ")", "{", "case", "*", "elastic2", ".", "AliasesResult", ":", "hasAlias", "=", "a", ".", "Indices", "[", "indexName", "]", ".", "HasAlias", "(", "aliasName", ")", "\n", "case", "*", "elastic5", ".", "AliasesResult", ":", "hasAlias", "=", "a", ".", "Indices", "[", "indexName", "]", ".", "HasAlias", "(", "aliasName", ")", "\n", "}", "\n", "if", "!", "hasAlias", "{", "createAlias", ",", "err", ":=", "esSvc", ".", "EsClient", ".", "AddAlias", "(", "indexName", ",", "esSvc", ".", "IndexAlias", "(", "typeName", ")", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "ack", ":=", "false", "\n", "switch", "i", ":=", "createAlias", ".", "(", "type", ")", "{", "case", "*", "elastic2", ".", "AliasResult", ":", "ack", "=", "i", ".", "Acknowledged", "\n", "case", "*", "elastic5", ".", "AliasResult", ":", "ack", "=", "i", ".", "Acknowledged", "\n", "}", "\n", "if", "!", "ack", "{", "return", "errors", ".", "New", "(", "\"", "\"", ")", "\n", "}", "\n", "}", "\n\n", "for", "_", ",", "data", ":=", "range", "sinkData", "{", "esSvc", ".", "EsClient", ".", "AddBulkReq", "(", "indexName", ",", "typeName", ",", "data", ")", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// SaveDataIntoES save metrics and events to ES by using ES client
[ "SaveDataIntoES", "save", "metrics", "and", "events", "to", "ES", "by", "using", "ES", "client" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/common/elasticsearch/elasticsearch.go#L54-L122
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb.go
newSink
func newSink(c influxdb_common.InfluxdbConfig) core.DataSink { client, err := influxdb_common.NewClient(c) if err != nil { glog.Errorf("issues while creating an InfluxDB sink: %v, will retry on use", err) } return &influxdbSink{ client: client, // can be nil c: c, conChan: make(chan struct{}, c.Concurrency), } }
go
func newSink(c influxdb_common.InfluxdbConfig) core.DataSink { client, err := influxdb_common.NewClient(c) if err != nil { glog.Errorf("issues while creating an InfluxDB sink: %v, will retry on use", err) } return &influxdbSink{ client: client, // can be nil c: c, conChan: make(chan struct{}, c.Concurrency), } }
[ "func", "newSink", "(", "c", "influxdb_common", ".", "InfluxdbConfig", ")", "core", ".", "DataSink", "{", "client", ",", "err", ":=", "influxdb_common", ".", "NewClient", "(", "c", ")", "\n", "if", "err", "!=", "nil", "{", "glog", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "return", "&", "influxdbSink", "{", "client", ":", "client", ",", "// can be nil", "c", ":", "c", ",", "conChan", ":", "make", "(", "chan", "struct", "{", "}", ",", "c", ".", "Concurrency", ")", ",", "}", "\n", "}" ]
// Returns a thread-compatible implementation of influxdb interactions.
[ "Returns", "a", "thread", "-", "compatible", "implementation", "of", "influxdb", "interactions", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb.go#L294-L304
train
kubernetes-retired/heapster
integration/framework.go
ParseRBAC
func (self *realKubeFramework) ParseRBAC(filePath string) (*rbacv1.ClusterRoleBinding, error) { obj, err := self.loadRBACObject(filePath) if err != nil { return nil, err } rbac, ok := obj.(*rbacv1.ClusterRoleBinding) if !ok { return nil, fmt.Errorf("Failed to cast clusterrolebinding: %v", obj) } return rbac, nil }
go
func (self *realKubeFramework) ParseRBAC(filePath string) (*rbacv1.ClusterRoleBinding, error) { obj, err := self.loadRBACObject(filePath) if err != nil { return nil, err } rbac, ok := obj.(*rbacv1.ClusterRoleBinding) if !ok { return nil, fmt.Errorf("Failed to cast clusterrolebinding: %v", obj) } return rbac, nil }
[ "func", "(", "self", "*", "realKubeFramework", ")", "ParseRBAC", "(", "filePath", "string", ")", "(", "*", "rbacv1", ".", "ClusterRoleBinding", ",", "error", ")", "{", "obj", ",", "err", ":=", "self", ".", "loadRBACObject", "(", "filePath", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "rbac", ",", "ok", ":=", "obj", ".", "(", "*", "rbacv1", ".", "ClusterRoleBinding", ")", "\n", "if", "!", "ok", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "obj", ")", "\n", "}", "\n", "return", "rbac", ",", "nil", "\n", "}" ]
// Parses and Returns a RBAC object contained in 'filePath'
[ "Parses", "and", "Returns", "a", "RBAC", "object", "contained", "in", "filePath" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/integration/framework.go#L402-L413
train
kubernetes-retired/heapster
integration/framework.go
CreateRBAC
func (self *realKubeFramework) CreateRBAC(rbac *rbacv1.ClusterRoleBinding) error { _, err := self.kubeClient.RbacV1().ClusterRoleBindings().Create(rbac) return err }
go
func (self *realKubeFramework) CreateRBAC(rbac *rbacv1.ClusterRoleBinding) error { _, err := self.kubeClient.RbacV1().ClusterRoleBindings().Create(rbac) return err }
[ "func", "(", "self", "*", "realKubeFramework", ")", "CreateRBAC", "(", "rbac", "*", "rbacv1", ".", "ClusterRoleBinding", ")", "error", "{", "_", ",", "err", ":=", "self", ".", "kubeClient", ".", "RbacV1", "(", ")", ".", "ClusterRoleBindings", "(", ")", ".", "Create", "(", "rbac", ")", "\n", "return", "err", "\n", "}" ]
// CreateRBAC creates the RBAC object
[ "CreateRBAC", "creates", "the", "RBAC", "object" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/integration/framework.go#L416-L419
train
kubernetes-retired/heapster
integration/framework.go
ParseServiceAccount
func (self *realKubeFramework) ParseServiceAccount(filePath string) (*v1.ServiceAccount, error) { obj, err := self.loadObject(filePath) if err != nil { return nil, err } sa, ok := obj.(*v1.ServiceAccount) if !ok { return nil, fmt.Errorf("Failed to cast serviceaccount: %v", obj) } return sa, nil }
go
func (self *realKubeFramework) ParseServiceAccount(filePath string) (*v1.ServiceAccount, error) { obj, err := self.loadObject(filePath) if err != nil { return nil, err } sa, ok := obj.(*v1.ServiceAccount) if !ok { return nil, fmt.Errorf("Failed to cast serviceaccount: %v", obj) } return sa, nil }
[ "func", "(", "self", "*", "realKubeFramework", ")", "ParseServiceAccount", "(", "filePath", "string", ")", "(", "*", "v1", ".", "ServiceAccount", ",", "error", ")", "{", "obj", ",", "err", ":=", "self", ".", "loadObject", "(", "filePath", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "sa", ",", "ok", ":=", "obj", ".", "(", "*", "v1", ".", "ServiceAccount", ")", "\n", "if", "!", "ok", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "obj", ")", "\n", "}", "\n", "return", "sa", ",", "nil", "\n", "}" ]
// Parses and Returns a ServiceAccount object contained in 'filePath'
[ "Parses", "and", "Returns", "a", "ServiceAccount", "object", "contained", "in", "filePath" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/integration/framework.go#L422-L433
train
kubernetes-retired/heapster
integration/framework.go
CreateServiceAccount
func (self *realKubeFramework) CreateServiceAccount(sa *v1.ServiceAccount) error { _, err := self.kubeClient.CoreV1().ServiceAccounts(sa.Namespace).Create(sa) return err }
go
func (self *realKubeFramework) CreateServiceAccount(sa *v1.ServiceAccount) error { _, err := self.kubeClient.CoreV1().ServiceAccounts(sa.Namespace).Create(sa) return err }
[ "func", "(", "self", "*", "realKubeFramework", ")", "CreateServiceAccount", "(", "sa", "*", "v1", ".", "ServiceAccount", ")", "error", "{", "_", ",", "err", ":=", "self", ".", "kubeClient", ".", "CoreV1", "(", ")", ".", "ServiceAccounts", "(", "sa", ".", "Namespace", ")", ".", "Create", "(", "sa", ")", "\n", "return", "err", "\n", "}" ]
// CreateServiceAccount creates the ServiceAccount object
[ "CreateServiceAccount", "creates", "the", "ServiceAccount", "object" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/integration/framework.go#L436-L439
train
kubernetes-retired/heapster
common/honeycomb/honeycomb.go
SendBatch
func (c *HoneycombClient) SendBatch(batch Batch) error { if len(batch) == 0 { // Nothing to send return nil } errs := []string{} for i := 0; i < len(batch); i += maxBatchSize { offset := i + maxBatchSize if offset > len(batch) { offset = len(batch) } if err := c.sendBatch(batch[i:offset]); err != nil { errs = append(errs, err.Error()) } } if len(errs) > 0 { return errors.New(strings.Join(errs, "\n")) } return nil }
go
func (c *HoneycombClient) SendBatch(batch Batch) error { if len(batch) == 0 { // Nothing to send return nil } errs := []string{} for i := 0; i < len(batch); i += maxBatchSize { offset := i + maxBatchSize if offset > len(batch) { offset = len(batch) } if err := c.sendBatch(batch[i:offset]); err != nil { errs = append(errs, err.Error()) } } if len(errs) > 0 { return errors.New(strings.Join(errs, "\n")) } return nil }
[ "func", "(", "c", "*", "HoneycombClient", ")", "SendBatch", "(", "batch", "Batch", ")", "error", "{", "if", "len", "(", "batch", ")", "==", "0", "{", "// Nothing to send", "return", "nil", "\n", "}", "\n\n", "errs", ":=", "[", "]", "string", "{", "}", "\n", "for", "i", ":=", "0", ";", "i", "<", "len", "(", "batch", ")", ";", "i", "+=", "maxBatchSize", "{", "offset", ":=", "i", "+", "maxBatchSize", "\n", "if", "offset", ">", "len", "(", "batch", ")", "{", "offset", "=", "len", "(", "batch", ")", "\n", "}", "\n", "if", "err", ":=", "c", ".", "sendBatch", "(", "batch", "[", "i", ":", "offset", "]", ")", ";", "err", "!=", "nil", "{", "errs", "=", "append", "(", "errs", ",", "err", ".", "Error", "(", ")", ")", "\n", "}", "\n", "}", "\n\n", "if", "len", "(", "errs", ")", ">", "0", "{", "return", "errors", ".", "New", "(", "strings", ".", "Join", "(", "errs", ",", "\"", "\\n", "\"", ")", ")", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// SendBatch splits the top-level batch into sub-batches if needed. Otherwise, // requests that are too large will be rejected by the Honeycomb API.
[ "SendBatch", "splits", "the", "top", "-", "level", "batch", "into", "sub", "-", "batches", "if", "needed", ".", "Otherwise", "requests", "that", "are", "too", "large", "will", "be", "rejected", "by", "the", "Honeycomb", "API", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/common/honeycomb/honeycomb.go#L108-L130
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
checkSanitizedMetricName
func (sink *influxdbSink) checkSanitizedMetricName(name string) error { if !metricAllowedChars.MatchString(name) { return fmt.Errorf("Invalid metric name %q", name) } return nil }
go
func (sink *influxdbSink) checkSanitizedMetricName(name string) error { if !metricAllowedChars.MatchString(name) { return fmt.Errorf("Invalid metric name %q", name) } return nil }
[ "func", "(", "sink", "*", "influxdbSink", ")", "checkSanitizedMetricName", "(", "name", "string", ")", "error", "{", "if", "!", "metricAllowedChars", ".", "MatchString", "(", "name", ")", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "name", ")", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// checkSanitizedMetricName errors out if invalid characters are found in the metric name, since InfluxDB // does not widely support bound parameters yet, and we need to sanitize our inputs.
[ "checkSanitizedMetricName", "errors", "out", "if", "invalid", "characters", "are", "found", "in", "the", "metric", "name", "since", "InfluxDB", "does", "not", "widely", "support", "bound", "parameters", "yet", "and", "we", "need", "to", "sanitize", "our", "inputs", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L78-L84
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
checkSanitizedMetricLabels
func (sink *influxdbSink) checkSanitizedMetricLabels(labels map[string]string) error { // label names have the same restrictions as metric names, here for k, v := range labels { if !metricAllowedChars.MatchString(k) { return fmt.Errorf("Invalid label name %q", k) } // for metric values, we're somewhat more permissive. We allow any // Printable unicode character, except quotation marks, which are used // to delimit things. if strings.ContainsRune(v, '"') || strings.ContainsRune(v, '\'') { return fmt.Errorf("Invalid label value %q", v) } for _, runeVal := range v { if !unicode.IsPrint(runeVal) { return fmt.Errorf("Invalid label value %q", v) } } } return nil }
go
func (sink *influxdbSink) checkSanitizedMetricLabels(labels map[string]string) error { // label names have the same restrictions as metric names, here for k, v := range labels { if !metricAllowedChars.MatchString(k) { return fmt.Errorf("Invalid label name %q", k) } // for metric values, we're somewhat more permissive. We allow any // Printable unicode character, except quotation marks, which are used // to delimit things. if strings.ContainsRune(v, '"') || strings.ContainsRune(v, '\'') { return fmt.Errorf("Invalid label value %q", v) } for _, runeVal := range v { if !unicode.IsPrint(runeVal) { return fmt.Errorf("Invalid label value %q", v) } } } return nil }
[ "func", "(", "sink", "*", "influxdbSink", ")", "checkSanitizedMetricLabels", "(", "labels", "map", "[", "string", "]", "string", ")", "error", "{", "// label names have the same restrictions as metric names, here", "for", "k", ",", "v", ":=", "range", "labels", "{", "if", "!", "metricAllowedChars", ".", "MatchString", "(", "k", ")", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "k", ")", "\n", "}", "\n\n", "// for metric values, we're somewhat more permissive. We allow any", "// Printable unicode character, except quotation marks, which are used", "// to delimit things.", "if", "strings", ".", "ContainsRune", "(", "v", ",", "'\"'", ")", "||", "strings", ".", "ContainsRune", "(", "v", ",", "'\\''", ")", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "v", ")", "\n", "}", "\n\n", "for", "_", ",", "runeVal", ":=", "range", "v", "{", "if", "!", "unicode", ".", "IsPrint", "(", "runeVal", ")", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "v", ")", "\n", "}", "\n", "}", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// checkSanitizedMetricLabels errors out if invalid characters are found in the label name or label value, since // InfluxDb does not widely support bound parameters yet, and we need to sanitize our inputs.
[ "checkSanitizedMetricLabels", "errors", "out", "if", "invalid", "characters", "are", "found", "in", "the", "label", "name", "or", "label", "value", "since", "InfluxDb", "does", "not", "widely", "support", "bound", "parameters", "yet", "and", "we", "need", "to", "sanitize", "our", "inputs", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L88-L110
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
aggregationFunc
func (sink *influxdbSink) aggregationFunc(aggregationName core.AggregationType, fieldName string) string { switch aggregationName { case core.AggregationTypeAverage: return fmt.Sprintf("MEAN(%q)", fieldName) case core.AggregationTypeMaximum: return fmt.Sprintf("MAX(%q)", fieldName) case core.AggregationTypeMinimum: return fmt.Sprintf("MIN(%q)", fieldName) case core.AggregationTypeMedian: return fmt.Sprintf("MEDIAN(%q)", fieldName) case core.AggregationTypeCount: return fmt.Sprintf("COUNT(%q)", fieldName) case core.AggregationTypePercentile50: return fmt.Sprintf("PERCENTILE(%q, 50)", fieldName) case core.AggregationTypePercentile95: return fmt.Sprintf("PERCENTILE(%q, 95)", fieldName) case core.AggregationTypePercentile99: return fmt.Sprintf("PERCENTILE(%q, 99)", fieldName) } // This should have been checked by the API level, so something's seriously wrong here panic(fmt.Sprintf("Unknown aggregation type %q", aggregationName)) }
go
func (sink *influxdbSink) aggregationFunc(aggregationName core.AggregationType, fieldName string) string { switch aggregationName { case core.AggregationTypeAverage: return fmt.Sprintf("MEAN(%q)", fieldName) case core.AggregationTypeMaximum: return fmt.Sprintf("MAX(%q)", fieldName) case core.AggregationTypeMinimum: return fmt.Sprintf("MIN(%q)", fieldName) case core.AggregationTypeMedian: return fmt.Sprintf("MEDIAN(%q)", fieldName) case core.AggregationTypeCount: return fmt.Sprintf("COUNT(%q)", fieldName) case core.AggregationTypePercentile50: return fmt.Sprintf("PERCENTILE(%q, 50)", fieldName) case core.AggregationTypePercentile95: return fmt.Sprintf("PERCENTILE(%q, 95)", fieldName) case core.AggregationTypePercentile99: return fmt.Sprintf("PERCENTILE(%q, 99)", fieldName) } // This should have been checked by the API level, so something's seriously wrong here panic(fmt.Sprintf("Unknown aggregation type %q", aggregationName)) }
[ "func", "(", "sink", "*", "influxdbSink", ")", "aggregationFunc", "(", "aggregationName", "core", ".", "AggregationType", ",", "fieldName", "string", ")", "string", "{", "switch", "aggregationName", "{", "case", "core", ".", "AggregationTypeAverage", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ")", "\n", "case", "core", ".", "AggregationTypeMaximum", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ")", "\n", "case", "core", ".", "AggregationTypeMinimum", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ")", "\n", "case", "core", ".", "AggregationTypeMedian", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ")", "\n", "case", "core", ".", "AggregationTypeCount", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ")", "\n", "case", "core", ".", "AggregationTypePercentile50", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ")", "\n", "case", "core", ".", "AggregationTypePercentile95", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ")", "\n", "case", "core", ".", "AggregationTypePercentile99", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ")", "\n", "}", "\n\n", "// This should have been checked by the API level, so something's seriously wrong here", "panic", "(", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "aggregationName", ")", ")", "\n", "}" ]
// aggregationFunc converts an aggregation name into the equivalent call to an InfluxQL // aggregation function
[ "aggregationFunc", "converts", "an", "aggregation", "name", "into", "the", "equivalent", "call", "to", "an", "InfluxQL", "aggregation", "function" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L114-L136
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
keyToSelector
func (sink *influxdbSink) keyToSelector(key core.HistoricalKey) string { typeSel := fmt.Sprintf("type = '%s'", key.ObjectType) switch key.ObjectType { case core.MetricSetTypeNode: return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelNodename.Key, key.NodeName) case core.MetricSetTypeSystemContainer: return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelContainerName.Key, key.ContainerName, core.LabelNodename.Key, key.NodeName) case core.MetricSetTypeCluster: return typeSel case core.MetricSetTypeNamespace: return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName) case core.MetricSetTypePod: if key.PodId != "" { return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelPodId.Key, key.PodId) } else { return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName, core.LabelPodName.Key, key.PodName) } case core.MetricSetTypePodContainer: if key.PodId != "" { return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelPodId.Key, key.PodId, core.LabelContainerName.Key, key.ContainerName) } else { return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s' AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName, core.LabelPodName.Key, key.PodName, core.LabelContainerName.Key, key.ContainerName) } } // These are assigned by the API, so it shouldn't be possible to reach this unless things are really broken panic(fmt.Sprintf("Unknown metric type %q", key.ObjectType)) }
go
func (sink *influxdbSink) keyToSelector(key core.HistoricalKey) string { typeSel := fmt.Sprintf("type = '%s'", key.ObjectType) switch key.ObjectType { case core.MetricSetTypeNode: return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelNodename.Key, key.NodeName) case core.MetricSetTypeSystemContainer: return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelContainerName.Key, key.ContainerName, core.LabelNodename.Key, key.NodeName) case core.MetricSetTypeCluster: return typeSel case core.MetricSetTypeNamespace: return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName) case core.MetricSetTypePod: if key.PodId != "" { return fmt.Sprintf("%s AND %s = '%s'", typeSel, core.LabelPodId.Key, key.PodId) } else { return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName, core.LabelPodName.Key, key.PodName) } case core.MetricSetTypePodContainer: if key.PodId != "" { return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s'", typeSel, core.LabelPodId.Key, key.PodId, core.LabelContainerName.Key, key.ContainerName) } else { return fmt.Sprintf("%s AND %s = '%s' AND %s = '%s' AND %s = '%s'", typeSel, core.LabelNamespaceName.Key, key.NamespaceName, core.LabelPodName.Key, key.PodName, core.LabelContainerName.Key, key.ContainerName) } } // These are assigned by the API, so it shouldn't be possible to reach this unless things are really broken panic(fmt.Sprintf("Unknown metric type %q", key.ObjectType)) }
[ "func", "(", "sink", "*", "influxdbSink", ")", "keyToSelector", "(", "key", "core", ".", "HistoricalKey", ")", "string", "{", "typeSel", ":=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "key", ".", "ObjectType", ")", "\n", "switch", "key", ".", "ObjectType", "{", "case", "core", ".", "MetricSetTypeNode", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "typeSel", ",", "core", ".", "LabelNodename", ".", "Key", ",", "key", ".", "NodeName", ")", "\n", "case", "core", ".", "MetricSetTypeSystemContainer", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "typeSel", ",", "core", ".", "LabelContainerName", ".", "Key", ",", "key", ".", "ContainerName", ",", "core", ".", "LabelNodename", ".", "Key", ",", "key", ".", "NodeName", ")", "\n", "case", "core", ".", "MetricSetTypeCluster", ":", "return", "typeSel", "\n", "case", "core", ".", "MetricSetTypeNamespace", ":", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "typeSel", ",", "core", ".", "LabelNamespaceName", ".", "Key", ",", "key", ".", "NamespaceName", ")", "\n", "case", "core", ".", "MetricSetTypePod", ":", "if", "key", ".", "PodId", "!=", "\"", "\"", "{", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "typeSel", ",", "core", ".", "LabelPodId", ".", "Key", ",", "key", ".", "PodId", ")", "\n", "}", "else", "{", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "typeSel", ",", "core", ".", "LabelNamespaceName", ".", "Key", ",", "key", ".", "NamespaceName", ",", "core", ".", "LabelPodName", ".", "Key", ",", "key", ".", "PodName", ")", "\n", "}", "\n", "case", "core", ".", "MetricSetTypePodContainer", ":", "if", "key", ".", "PodId", "!=", "\"", "\"", "{", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "typeSel", ",", "core", ".", "LabelPodId", ".", "Key", ",", "key", ".", "PodId", ",", "core", ".", "LabelContainerName", ".", "Key", ",", "key", ".", "ContainerName", ")", "\n", "}", "else", "{", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "typeSel", ",", "core", ".", "LabelNamespaceName", ".", "Key", ",", "key", ".", "NamespaceName", ",", "core", ".", "LabelPodName", ".", "Key", ",", "key", ".", "PodName", ",", "core", ".", "LabelContainerName", ".", "Key", ",", "key", ".", "ContainerName", ")", "\n", "}", "\n", "}", "\n\n", "// These are assigned by the API, so it shouldn't be possible to reach this unless things are really broken", "panic", "(", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "key", ".", "ObjectType", ")", ")", "\n", "}" ]
// keyToSelector converts a HistoricalKey to an InfluxQL predicate
[ "keyToSelector", "converts", "a", "HistoricalKey", "to", "an", "InfluxQL", "predicate" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L139-L166
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
labelsToPredicate
func (sink *influxdbSink) labelsToPredicate(labels map[string]string) string { if len(labels) == 0 { return "" } parts := make([]string, 0, len(labels)) for k, v := range labels { parts = append(parts, fmt.Sprintf("%q = '%s'", k, v)) } return strings.Join(parts, " AND ") }
go
func (sink *influxdbSink) labelsToPredicate(labels map[string]string) string { if len(labels) == 0 { return "" } parts := make([]string, 0, len(labels)) for k, v := range labels { parts = append(parts, fmt.Sprintf("%q = '%s'", k, v)) } return strings.Join(parts, " AND ") }
[ "func", "(", "sink", "*", "influxdbSink", ")", "labelsToPredicate", "(", "labels", "map", "[", "string", "]", "string", ")", "string", "{", "if", "len", "(", "labels", ")", "==", "0", "{", "return", "\"", "\"", "\n", "}", "\n\n", "parts", ":=", "make", "(", "[", "]", "string", ",", "0", ",", "len", "(", "labels", ")", ")", "\n", "for", "k", ",", "v", ":=", "range", "labels", "{", "parts", "=", "append", "(", "parts", ",", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "k", ",", "v", ")", ")", "\n", "}", "\n\n", "return", "strings", ".", "Join", "(", "parts", ",", "\"", "\"", ")", "\n", "}" ]
// labelsToPredicate composes an InfluxQL predicate based on the given map of labels
[ "labelsToPredicate", "composes", "an", "InfluxQL", "predicate", "based", "on", "the", "given", "map", "of", "labels" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L169-L180
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
composeRawQuery
func (sink *influxdbSink) composeRawQuery(metricName string, labels map[string]string, metricKeys []core.HistoricalKey, start, end time.Time) string { seriesName, fieldName := sink.metricToSeriesAndField(metricName) queries := make([]string, len(metricKeys)) for i, key := range metricKeys { pred := sink.keyToSelector(key) if labels != nil { pred += fmt.Sprintf(" AND %s", sink.labelsToPredicate(labels)) } if !start.IsZero() { pred += fmt.Sprintf(" AND time > '%s'", start.Format(time.RFC3339)) } if !end.IsZero() { pred += fmt.Sprintf(" AND time < '%s'", end.Format(time.RFC3339)) } queries[i] = fmt.Sprintf("SELECT time, %q FROM %q WHERE %s", fieldName, seriesName, pred) } return strings.Join(queries, "; ") }
go
func (sink *influxdbSink) composeRawQuery(metricName string, labels map[string]string, metricKeys []core.HistoricalKey, start, end time.Time) string { seriesName, fieldName := sink.metricToSeriesAndField(metricName) queries := make([]string, len(metricKeys)) for i, key := range metricKeys { pred := sink.keyToSelector(key) if labels != nil { pred += fmt.Sprintf(" AND %s", sink.labelsToPredicate(labels)) } if !start.IsZero() { pred += fmt.Sprintf(" AND time > '%s'", start.Format(time.RFC3339)) } if !end.IsZero() { pred += fmt.Sprintf(" AND time < '%s'", end.Format(time.RFC3339)) } queries[i] = fmt.Sprintf("SELECT time, %q FROM %q WHERE %s", fieldName, seriesName, pred) } return strings.Join(queries, "; ") }
[ "func", "(", "sink", "*", "influxdbSink", ")", "composeRawQuery", "(", "metricName", "string", ",", "labels", "map", "[", "string", "]", "string", ",", "metricKeys", "[", "]", "core", ".", "HistoricalKey", ",", "start", ",", "end", "time", ".", "Time", ")", "string", "{", "seriesName", ",", "fieldName", ":=", "sink", ".", "metricToSeriesAndField", "(", "metricName", ")", "\n\n", "queries", ":=", "make", "(", "[", "]", "string", ",", "len", "(", "metricKeys", ")", ")", "\n", "for", "i", ",", "key", ":=", "range", "metricKeys", "{", "pred", ":=", "sink", ".", "keyToSelector", "(", "key", ")", "\n", "if", "labels", "!=", "nil", "{", "pred", "+=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "sink", ".", "labelsToPredicate", "(", "labels", ")", ")", "\n", "}", "\n", "if", "!", "start", ".", "IsZero", "(", ")", "{", "pred", "+=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "start", ".", "Format", "(", "time", ".", "RFC3339", ")", ")", "\n", "}", "\n", "if", "!", "end", ".", "IsZero", "(", ")", "{", "pred", "+=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "end", ".", "Format", "(", "time", ".", "RFC3339", ")", ")", "\n", "}", "\n", "queries", "[", "i", "]", "=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "fieldName", ",", "seriesName", ",", "pred", ")", "\n", "}", "\n\n", "return", "strings", ".", "Join", "(", "queries", ",", "\"", "\"", ")", "\n", "}" ]
// composeRawQuery creates the InfluxQL query to fetch the given metric values
[ "composeRawQuery", "creates", "the", "InfluxQL", "query", "to", "fetch", "the", "given", "metric", "values" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L198-L217
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
parseRawQueryRow
func (sink *influxdbSink) parseRawQueryRow(rawRow influx_models.Row) ([]core.TimestampedMetricValue, error) { vals := make([]core.TimestampedMetricValue, len(rawRow.Values)) wasInt := make(map[string]bool, 1) for i, rawVal := range rawRow.Values { val := core.TimestampedMetricValue{} if ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil { return nil, fmt.Errorf("Unable to parse timestamp %q in series %q", rawVal[0].(string), rawRow.Name) } else { val.Timestamp = ts } if err := tryParseMetricValue("value", rawVal, &val.MetricValue, 1, wasInt); err != nil { glog.Errorf("Unable to parse field \"value\" in series %q: %v", rawRow.Name, err) return nil, fmt.Errorf("Unable to parse values in series %q", rawRow.Name) } vals[i] = val } if wasInt["value"] { for i := range vals { vals[i].MetricValue.ValueType = core.ValueInt64 } } else { for i := range vals { vals[i].MetricValue.ValueType = core.ValueFloat } } return vals, nil }
go
func (sink *influxdbSink) parseRawQueryRow(rawRow influx_models.Row) ([]core.TimestampedMetricValue, error) { vals := make([]core.TimestampedMetricValue, len(rawRow.Values)) wasInt := make(map[string]bool, 1) for i, rawVal := range rawRow.Values { val := core.TimestampedMetricValue{} if ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil { return nil, fmt.Errorf("Unable to parse timestamp %q in series %q", rawVal[0].(string), rawRow.Name) } else { val.Timestamp = ts } if err := tryParseMetricValue("value", rawVal, &val.MetricValue, 1, wasInt); err != nil { glog.Errorf("Unable to parse field \"value\" in series %q: %v", rawRow.Name, err) return nil, fmt.Errorf("Unable to parse values in series %q", rawRow.Name) } vals[i] = val } if wasInt["value"] { for i := range vals { vals[i].MetricValue.ValueType = core.ValueInt64 } } else { for i := range vals { vals[i].MetricValue.ValueType = core.ValueFloat } } return vals, nil }
[ "func", "(", "sink", "*", "influxdbSink", ")", "parseRawQueryRow", "(", "rawRow", "influx_models", ".", "Row", ")", "(", "[", "]", "core", ".", "TimestampedMetricValue", ",", "error", ")", "{", "vals", ":=", "make", "(", "[", "]", "core", ".", "TimestampedMetricValue", ",", "len", "(", "rawRow", ".", "Values", ")", ")", "\n", "wasInt", ":=", "make", "(", "map", "[", "string", "]", "bool", ",", "1", ")", "\n", "for", "i", ",", "rawVal", ":=", "range", "rawRow", ".", "Values", "{", "val", ":=", "core", ".", "TimestampedMetricValue", "{", "}", "\n\n", "if", "ts", ",", "err", ":=", "time", ".", "Parse", "(", "time", ".", "RFC3339", ",", "rawVal", "[", "0", "]", ".", "(", "string", ")", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "rawVal", "[", "0", "]", ".", "(", "string", ")", ",", "rawRow", ".", "Name", ")", "\n", "}", "else", "{", "val", ".", "Timestamp", "=", "ts", "\n", "}", "\n\n", "if", "err", ":=", "tryParseMetricValue", "(", "\"", "\"", ",", "rawVal", ",", "&", "val", ".", "MetricValue", ",", "1", ",", "wasInt", ")", ";", "err", "!=", "nil", "{", "glog", ".", "Errorf", "(", "\"", "\\\"", "\\\"", "\"", ",", "rawRow", ".", "Name", ",", "err", ")", "\n", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "rawRow", ".", "Name", ")", "\n", "}", "\n\n", "vals", "[", "i", "]", "=", "val", "\n", "}", "\n\n", "if", "wasInt", "[", "\"", "\"", "]", "{", "for", "i", ":=", "range", "vals", "{", "vals", "[", "i", "]", ".", "MetricValue", ".", "ValueType", "=", "core", ".", "ValueInt64", "\n", "}", "\n", "}", "else", "{", "for", "i", ":=", "range", "vals", "{", "vals", "[", "i", "]", ".", "MetricValue", ".", "ValueType", "=", "core", ".", "ValueFloat", "\n", "}", "\n", "}", "\n\n", "return", "vals", ",", "nil", "\n", "}" ]
// parseRawQueryRow parses a set of timestamped metric values from unstructured JSON output into the // appropriate Heapster form
[ "parseRawQueryRow", "parses", "a", "set", "of", "timestamped", "metric", "values", "from", "unstructured", "JSON", "output", "into", "the", "appropriate", "Heapster", "form" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L221-L252
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
composeAggregateQuery
func (sink *influxdbSink) composeAggregateQuery(metricName string, labels map[string]string, aggregations []core.AggregationType, metricKeys []core.HistoricalKey, start, end time.Time, bucketSize time.Duration) string { seriesName, fieldName := sink.metricToSeriesAndField(metricName) var bucketSizeNanoSeconds int64 = 0 if bucketSize != 0 { bucketSizeNanoSeconds = int64(bucketSize.Nanoseconds() / int64(time.Microsecond/time.Nanosecond)) } queries := make([]string, len(metricKeys)) for i, key := range metricKeys { pred := sink.keyToSelector(key) if labels != nil { pred += fmt.Sprintf(" AND %s", sink.labelsToPredicate(labels)) } if !start.IsZero() { pred += fmt.Sprintf(" AND time > '%s'", start.Format(time.RFC3339)) } if !end.IsZero() { pred += fmt.Sprintf(" AND time < '%s'", end.Format(time.RFC3339)) } aggParts := make([]string, len(aggregations)) for i, agg := range aggregations { aggParts[i] = sink.aggregationFunc(agg, fieldName) } queries[i] = fmt.Sprintf("SELECT %s FROM %q WHERE %s", strings.Join(aggParts, ", "), seriesName, pred) if bucketSize != 0 { // group by time requires we have at least one time bound if start.IsZero() && end.IsZero() { queries[i] += fmt.Sprintf(" AND time < now()") } // fill(none) makes sure we skip data points will null values (otherwise we'll get a *bunch* of null // values when we go back beyond the time where we started collecting data). queries[i] += fmt.Sprintf(" GROUP BY time(%vu) fill(none)", bucketSizeNanoSeconds) } } return strings.Join(queries, "; ") }
go
func (sink *influxdbSink) composeAggregateQuery(metricName string, labels map[string]string, aggregations []core.AggregationType, metricKeys []core.HistoricalKey, start, end time.Time, bucketSize time.Duration) string { seriesName, fieldName := sink.metricToSeriesAndField(metricName) var bucketSizeNanoSeconds int64 = 0 if bucketSize != 0 { bucketSizeNanoSeconds = int64(bucketSize.Nanoseconds() / int64(time.Microsecond/time.Nanosecond)) } queries := make([]string, len(metricKeys)) for i, key := range metricKeys { pred := sink.keyToSelector(key) if labels != nil { pred += fmt.Sprintf(" AND %s", sink.labelsToPredicate(labels)) } if !start.IsZero() { pred += fmt.Sprintf(" AND time > '%s'", start.Format(time.RFC3339)) } if !end.IsZero() { pred += fmt.Sprintf(" AND time < '%s'", end.Format(time.RFC3339)) } aggParts := make([]string, len(aggregations)) for i, agg := range aggregations { aggParts[i] = sink.aggregationFunc(agg, fieldName) } queries[i] = fmt.Sprintf("SELECT %s FROM %q WHERE %s", strings.Join(aggParts, ", "), seriesName, pred) if bucketSize != 0 { // group by time requires we have at least one time bound if start.IsZero() && end.IsZero() { queries[i] += fmt.Sprintf(" AND time < now()") } // fill(none) makes sure we skip data points will null values (otherwise we'll get a *bunch* of null // values when we go back beyond the time where we started collecting data). queries[i] += fmt.Sprintf(" GROUP BY time(%vu) fill(none)", bucketSizeNanoSeconds) } } return strings.Join(queries, "; ") }
[ "func", "(", "sink", "*", "influxdbSink", ")", "composeAggregateQuery", "(", "metricName", "string", ",", "labels", "map", "[", "string", "]", "string", ",", "aggregations", "[", "]", "core", ".", "AggregationType", ",", "metricKeys", "[", "]", "core", ".", "HistoricalKey", ",", "start", ",", "end", "time", ".", "Time", ",", "bucketSize", "time", ".", "Duration", ")", "string", "{", "seriesName", ",", "fieldName", ":=", "sink", ".", "metricToSeriesAndField", "(", "metricName", ")", "\n\n", "var", "bucketSizeNanoSeconds", "int64", "=", "0", "\n", "if", "bucketSize", "!=", "0", "{", "bucketSizeNanoSeconds", "=", "int64", "(", "bucketSize", ".", "Nanoseconds", "(", ")", "/", "int64", "(", "time", ".", "Microsecond", "/", "time", ".", "Nanosecond", ")", ")", "\n", "}", "\n\n", "queries", ":=", "make", "(", "[", "]", "string", ",", "len", "(", "metricKeys", ")", ")", "\n", "for", "i", ",", "key", ":=", "range", "metricKeys", "{", "pred", ":=", "sink", ".", "keyToSelector", "(", "key", ")", "\n", "if", "labels", "!=", "nil", "{", "pred", "+=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "sink", ".", "labelsToPredicate", "(", "labels", ")", ")", "\n", "}", "\n", "if", "!", "start", ".", "IsZero", "(", ")", "{", "pred", "+=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "start", ".", "Format", "(", "time", ".", "RFC3339", ")", ")", "\n", "}", "\n", "if", "!", "end", ".", "IsZero", "(", ")", "{", "pred", "+=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "end", ".", "Format", "(", "time", ".", "RFC3339", ")", ")", "\n", "}", "\n\n", "aggParts", ":=", "make", "(", "[", "]", "string", ",", "len", "(", "aggregations", ")", ")", "\n", "for", "i", ",", "agg", ":=", "range", "aggregations", "{", "aggParts", "[", "i", "]", "=", "sink", ".", "aggregationFunc", "(", "agg", ",", "fieldName", ")", "\n", "}", "\n\n", "queries", "[", "i", "]", "=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "strings", ".", "Join", "(", "aggParts", ",", "\"", "\"", ")", ",", "seriesName", ",", "pred", ")", "\n\n", "if", "bucketSize", "!=", "0", "{", "// group by time requires we have at least one time bound", "if", "start", ".", "IsZero", "(", ")", "&&", "end", ".", "IsZero", "(", ")", "{", "queries", "[", "i", "]", "+=", "fmt", ".", "Sprintf", "(", "\"", "\"", ")", "\n", "}", "\n\n", "// fill(none) makes sure we skip data points will null values (otherwise we'll get a *bunch* of null", "// values when we go back beyond the time where we started collecting data).", "queries", "[", "i", "]", "+=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "bucketSizeNanoSeconds", ")", "\n", "}", "\n", "}", "\n\n", "return", "strings", ".", "Join", "(", "queries", ",", "\"", "\"", ")", "\n", "}" ]
// composeAggregateQuery creates the InfluxQL query to fetch the given aggregation values
[ "composeAggregateQuery", "creates", "the", "InfluxQL", "query", "to", "fetch", "the", "given", "aggregation", "values" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L337-L378
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
parseAggregateQueryRow
func (sink *influxdbSink) parseAggregateQueryRow(rawRow influx_models.Row, aggregationLookup map[core.AggregationType]int, bucketSize time.Duration) ([]core.TimestampedAggregationValue, error) { vals := make([]core.TimestampedAggregationValue, len(rawRow.Values)) wasInt := make(map[string]bool, len(aggregationLookup)) for i, rawVal := range rawRow.Values { val := core.TimestampedAggregationValue{ BucketSize: bucketSize, AggregationValue: core.AggregationValue{ Aggregations: map[core.AggregationType]core.MetricValue{}, }, } if ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil { return nil, fmt.Errorf("Unable to parse timestamp %q in series %q", rawVal[0].(string), rawRow.Name) } else { val.Timestamp = ts } // The Influx client decods numeric fields to json.Number (a string), so we have to try decoding to both types of numbers // Count is always a uint64 if countIndex, ok := aggregationLookup[core.AggregationTypeCount]; ok { if err := json.Unmarshal([]byte(rawVal[countIndex].(json.Number).String()), &val.Count); err != nil { glog.Errorf("Unable to parse count value in series %q: %v", rawRow.Name, err) return nil, fmt.Errorf("Unable to parse values in series %q", rawRow.Name) } } // The rest of the aggregation values can be either float or int, so attempt to parse both if err := populateAggregations(rawRow.Name, rawVal, &val, aggregationLookup, wasInt); err != nil { return nil, err } vals[i] = val } // figure out whether each aggregation was full of float values, or int values setAggregationValueTypes(vals, wasInt) return vals, nil }
go
func (sink *influxdbSink) parseAggregateQueryRow(rawRow influx_models.Row, aggregationLookup map[core.AggregationType]int, bucketSize time.Duration) ([]core.TimestampedAggregationValue, error) { vals := make([]core.TimestampedAggregationValue, len(rawRow.Values)) wasInt := make(map[string]bool, len(aggregationLookup)) for i, rawVal := range rawRow.Values { val := core.TimestampedAggregationValue{ BucketSize: bucketSize, AggregationValue: core.AggregationValue{ Aggregations: map[core.AggregationType]core.MetricValue{}, }, } if ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil { return nil, fmt.Errorf("Unable to parse timestamp %q in series %q", rawVal[0].(string), rawRow.Name) } else { val.Timestamp = ts } // The Influx client decods numeric fields to json.Number (a string), so we have to try decoding to both types of numbers // Count is always a uint64 if countIndex, ok := aggregationLookup[core.AggregationTypeCount]; ok { if err := json.Unmarshal([]byte(rawVal[countIndex].(json.Number).String()), &val.Count); err != nil { glog.Errorf("Unable to parse count value in series %q: %v", rawRow.Name, err) return nil, fmt.Errorf("Unable to parse values in series %q", rawRow.Name) } } // The rest of the aggregation values can be either float or int, so attempt to parse both if err := populateAggregations(rawRow.Name, rawVal, &val, aggregationLookup, wasInt); err != nil { return nil, err } vals[i] = val } // figure out whether each aggregation was full of float values, or int values setAggregationValueTypes(vals, wasInt) return vals, nil }
[ "func", "(", "sink", "*", "influxdbSink", ")", "parseAggregateQueryRow", "(", "rawRow", "influx_models", ".", "Row", ",", "aggregationLookup", "map", "[", "core", ".", "AggregationType", "]", "int", ",", "bucketSize", "time", ".", "Duration", ")", "(", "[", "]", "core", ".", "TimestampedAggregationValue", ",", "error", ")", "{", "vals", ":=", "make", "(", "[", "]", "core", ".", "TimestampedAggregationValue", ",", "len", "(", "rawRow", ".", "Values", ")", ")", "\n", "wasInt", ":=", "make", "(", "map", "[", "string", "]", "bool", ",", "len", "(", "aggregationLookup", ")", ")", "\n\n", "for", "i", ",", "rawVal", ":=", "range", "rawRow", ".", "Values", "{", "val", ":=", "core", ".", "TimestampedAggregationValue", "{", "BucketSize", ":", "bucketSize", ",", "AggregationValue", ":", "core", ".", "AggregationValue", "{", "Aggregations", ":", "map", "[", "core", ".", "AggregationType", "]", "core", ".", "MetricValue", "{", "}", ",", "}", ",", "}", "\n\n", "if", "ts", ",", "err", ":=", "time", ".", "Parse", "(", "time", ".", "RFC3339", ",", "rawVal", "[", "0", "]", ".", "(", "string", ")", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "rawVal", "[", "0", "]", ".", "(", "string", ")", ",", "rawRow", ".", "Name", ")", "\n", "}", "else", "{", "val", ".", "Timestamp", "=", "ts", "\n", "}", "\n\n", "// The Influx client decods numeric fields to json.Number (a string), so we have to try decoding to both types of numbers", "// Count is always a uint64", "if", "countIndex", ",", "ok", ":=", "aggregationLookup", "[", "core", ".", "AggregationTypeCount", "]", ";", "ok", "{", "if", "err", ":=", "json", ".", "Unmarshal", "(", "[", "]", "byte", "(", "rawVal", "[", "countIndex", "]", ".", "(", "json", ".", "Number", ")", ".", "String", "(", ")", ")", ",", "&", "val", ".", "Count", ")", ";", "err", "!=", "nil", "{", "glog", ".", "Errorf", "(", "\"", "\"", ",", "rawRow", ".", "Name", ",", "err", ")", "\n", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "rawRow", ".", "Name", ")", "\n", "}", "\n", "}", "\n\n", "// The rest of the aggregation values can be either float or int, so attempt to parse both", "if", "err", ":=", "populateAggregations", "(", "rawRow", ".", "Name", ",", "rawVal", ",", "&", "val", ",", "aggregationLookup", ",", "wasInt", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "vals", "[", "i", "]", "=", "val", "\n", "}", "\n\n", "// figure out whether each aggregation was full of float values, or int values", "setAggregationValueTypes", "(", "vals", ",", "wasInt", ")", "\n\n", "return", "vals", ",", "nil", "\n", "}" ]
// parseRawQueryRow parses a set of timestamped aggregation values from unstructured JSON output into the // appropriate Heapster form
[ "parseRawQueryRow", "parses", "a", "set", "of", "timestamped", "aggregation", "values", "from", "unstructured", "JSON", "output", "into", "the", "appropriate", "Heapster", "form" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L382-L422
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
setAggregationValueIfPresent
func setAggregationValueIfPresent(aggName core.AggregationType, rawVal []interface{}, aggregations *core.AggregationValue, indexLookup map[core.AggregationType]int, wasInt map[string]bool) error { if fieldIndex, ok := indexLookup[aggName]; ok { targetValue := &core.MetricValue{} if err := tryParseMetricValue(string(aggName), rawVal, targetValue, fieldIndex, wasInt); err != nil { return err } aggregations.Aggregations[aggName] = *targetValue } return nil }
go
func setAggregationValueIfPresent(aggName core.AggregationType, rawVal []interface{}, aggregations *core.AggregationValue, indexLookup map[core.AggregationType]int, wasInt map[string]bool) error { if fieldIndex, ok := indexLookup[aggName]; ok { targetValue := &core.MetricValue{} if err := tryParseMetricValue(string(aggName), rawVal, targetValue, fieldIndex, wasInt); err != nil { return err } aggregations.Aggregations[aggName] = *targetValue } return nil }
[ "func", "setAggregationValueIfPresent", "(", "aggName", "core", ".", "AggregationType", ",", "rawVal", "[", "]", "interface", "{", "}", ",", "aggregations", "*", "core", ".", "AggregationValue", ",", "indexLookup", "map", "[", "core", ".", "AggregationType", "]", "int", ",", "wasInt", "map", "[", "string", "]", "bool", ")", "error", "{", "if", "fieldIndex", ",", "ok", ":=", "indexLookup", "[", "aggName", "]", ";", "ok", "{", "targetValue", ":=", "&", "core", ".", "MetricValue", "{", "}", "\n", "if", "err", ":=", "tryParseMetricValue", "(", "string", "(", "aggName", ")", ",", "rawVal", ",", "targetValue", ",", "fieldIndex", ",", "wasInt", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "aggregations", ".", "Aggregations", "[", "aggName", "]", "=", "*", "targetValue", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// setAggregationValueIfPresent checks if the given metric value is present in the list of raw values, and if so, // copies it to the output format
[ "setAggregationValueIfPresent", "checks", "if", "the", "given", "metric", "value", "is", "present", "in", "the", "list", "of", "raw", "values", "and", "if", "so", "copies", "it", "to", "the", "output", "format" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L524-L535
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
tryParseMetricValue
func tryParseMetricValue(aggName string, rawVal []interface{}, targetValue *core.MetricValue, fieldIndex int, wasInt map[string]bool) error { // the Influx client decodes numeric fields to json.Number (a string), so we have to deal with that -- // assume, starting off, that values may be either float or int. Try int until we fail once, and always // try float. At the end, figure out which is which. var rv string if rvN, ok := rawVal[fieldIndex].(json.Number); !ok { return fmt.Errorf("Value %q of metric %q was not a json.Number", rawVal[fieldIndex], aggName) } else { rv = rvN.String() } tryInt := false isInt, triedBefore := wasInt[aggName] tryInt = isInt || !triedBefore if tryInt { if err := json.Unmarshal([]byte(rv), &targetValue.IntValue); err != nil { wasInt[aggName] = false } else { wasInt[aggName] = true } } if err := json.Unmarshal([]byte(rv), &targetValue.FloatValue); err != nil { return err } return nil }
go
func tryParseMetricValue(aggName string, rawVal []interface{}, targetValue *core.MetricValue, fieldIndex int, wasInt map[string]bool) error { // the Influx client decodes numeric fields to json.Number (a string), so we have to deal with that -- // assume, starting off, that values may be either float or int. Try int until we fail once, and always // try float. At the end, figure out which is which. var rv string if rvN, ok := rawVal[fieldIndex].(json.Number); !ok { return fmt.Errorf("Value %q of metric %q was not a json.Number", rawVal[fieldIndex], aggName) } else { rv = rvN.String() } tryInt := false isInt, triedBefore := wasInt[aggName] tryInt = isInt || !triedBefore if tryInt { if err := json.Unmarshal([]byte(rv), &targetValue.IntValue); err != nil { wasInt[aggName] = false } else { wasInt[aggName] = true } } if err := json.Unmarshal([]byte(rv), &targetValue.FloatValue); err != nil { return err } return nil }
[ "func", "tryParseMetricValue", "(", "aggName", "string", ",", "rawVal", "[", "]", "interface", "{", "}", ",", "targetValue", "*", "core", ".", "MetricValue", ",", "fieldIndex", "int", ",", "wasInt", "map", "[", "string", "]", "bool", ")", "error", "{", "// the Influx client decodes numeric fields to json.Number (a string), so we have to deal with that --", "// assume, starting off, that values may be either float or int. Try int until we fail once, and always", "// try float. At the end, figure out which is which.", "var", "rv", "string", "\n", "if", "rvN", ",", "ok", ":=", "rawVal", "[", "fieldIndex", "]", ".", "(", "json", ".", "Number", ")", ";", "!", "ok", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "rawVal", "[", "fieldIndex", "]", ",", "aggName", ")", "\n", "}", "else", "{", "rv", "=", "rvN", ".", "String", "(", ")", "\n", "}", "\n\n", "tryInt", ":=", "false", "\n", "isInt", ",", "triedBefore", ":=", "wasInt", "[", "aggName", "]", "\n", "tryInt", "=", "isInt", "||", "!", "triedBefore", "\n\n", "if", "tryInt", "{", "if", "err", ":=", "json", ".", "Unmarshal", "(", "[", "]", "byte", "(", "rv", ")", ",", "&", "targetValue", ".", "IntValue", ")", ";", "err", "!=", "nil", "{", "wasInt", "[", "aggName", "]", "=", "false", "\n", "}", "else", "{", "wasInt", "[", "aggName", "]", "=", "true", "\n", "}", "\n", "}", "\n\n", "if", "err", ":=", "json", ".", "Unmarshal", "(", "[", "]", "byte", "(", "rv", ")", ",", "&", "targetValue", ".", "FloatValue", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// tryParseMetricValue attempts to parse a raw metric value into the appropriate go type.
[ "tryParseMetricValue", "attempts", "to", "parse", "a", "raw", "metric", "value", "into", "the", "appropriate", "go", "type", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L538-L567
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
GetMetricNames
func (sink *influxdbSink) GetMetricNames(metricKey core.HistoricalKey) ([]string, error) { if err := sink.checkSanitizedKey(&metricKey); err != nil { return nil, err } return sink.stringListQuery(fmt.Sprintf("SHOW MEASUREMENTS WHERE %s", sink.keyToSelector(metricKey)), "Unable to list available metrics") }
go
func (sink *influxdbSink) GetMetricNames(metricKey core.HistoricalKey) ([]string, error) { if err := sink.checkSanitizedKey(&metricKey); err != nil { return nil, err } return sink.stringListQuery(fmt.Sprintf("SHOW MEASUREMENTS WHERE %s", sink.keyToSelector(metricKey)), "Unable to list available metrics") }
[ "func", "(", "sink", "*", "influxdbSink", ")", "GetMetricNames", "(", "metricKey", "core", ".", "HistoricalKey", ")", "(", "[", "]", "string", ",", "error", ")", "{", "if", "err", ":=", "sink", ".", "checkSanitizedKey", "(", "&", "metricKey", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "return", "sink", ".", "stringListQuery", "(", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "sink", ".", "keyToSelector", "(", "metricKey", ")", ")", ",", "\"", "\"", ")", "\n", "}" ]
// GetMetricNames retrieves the available metric names for the given object
[ "GetMetricNames", "retrieves", "the", "available", "metric", "names", "for", "the", "given", "object" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L570-L575
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
GetNodes
func (sink *influxdbSink) GetNodes() ([]string, error) { return sink.stringListQuery(fmt.Sprintf("SHOW TAG VALUES WITH KEY = %s", core.LabelNodename.Key), "Unable to list all nodes") }
go
func (sink *influxdbSink) GetNodes() ([]string, error) { return sink.stringListQuery(fmt.Sprintf("SHOW TAG VALUES WITH KEY = %s", core.LabelNodename.Key), "Unable to list all nodes") }
[ "func", "(", "sink", "*", "influxdbSink", ")", "GetNodes", "(", ")", "(", "[", "]", "string", ",", "error", ")", "{", "return", "sink", ".", "stringListQuery", "(", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "core", ".", "LabelNodename", ".", "Key", ")", ",", "\"", "\"", ")", "\n", "}" ]
// GetNodes retrieves the list of nodes in the cluster
[ "GetNodes", "retrieves", "the", "list", "of", "nodes", "in", "the", "cluster" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L578-L580
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
GetNamespaces
func (sink *influxdbSink) GetNamespaces() ([]string, error) { return sink.stringListQuery(fmt.Sprintf("SHOW TAG VALUES WITH KEY = %s", core.LabelNamespaceName.Key), "Unable to list all namespaces") }
go
func (sink *influxdbSink) GetNamespaces() ([]string, error) { return sink.stringListQuery(fmt.Sprintf("SHOW TAG VALUES WITH KEY = %s", core.LabelNamespaceName.Key), "Unable to list all namespaces") }
[ "func", "(", "sink", "*", "influxdbSink", ")", "GetNamespaces", "(", ")", "(", "[", "]", "string", ",", "error", ")", "{", "return", "sink", ".", "stringListQuery", "(", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "core", ".", "LabelNamespaceName", ".", "Key", ")", ",", "\"", "\"", ")", "\n", "}" ]
// GetNamespaces retrieves the list of namespaces in the cluster
[ "GetNamespaces", "retrieves", "the", "list", "of", "namespaces", "in", "the", "cluster" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L583-L585
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
GetPodsFromNamespace
func (sink *influxdbSink) GetPodsFromNamespace(namespace string) ([]string, error) { if !nameAllowedChars.MatchString(namespace) { return nil, fmt.Errorf("Invalid namespace name %q", namespace) } // This is a bit difficult for the influx query language, so we cheat a bit here -- // we just get all series for the uptime measurement for pods which match our namespace // (any measurement should work here, though) q := fmt.Sprintf("SHOW SERIES FROM %q WHERE %s = '%s' AND type = '%s'", core.MetricUptime.MetricDescriptor.Name, core.LabelNamespaceName.Key, namespace, core.MetricSetTypePod) return sink.stringListQueryCol(q, core.LabelPodName.Key, fmt.Sprintf("Unable to list pods in namespace %q", namespace)) }
go
func (sink *influxdbSink) GetPodsFromNamespace(namespace string) ([]string, error) { if !nameAllowedChars.MatchString(namespace) { return nil, fmt.Errorf("Invalid namespace name %q", namespace) } // This is a bit difficult for the influx query language, so we cheat a bit here -- // we just get all series for the uptime measurement for pods which match our namespace // (any measurement should work here, though) q := fmt.Sprintf("SHOW SERIES FROM %q WHERE %s = '%s' AND type = '%s'", core.MetricUptime.MetricDescriptor.Name, core.LabelNamespaceName.Key, namespace, core.MetricSetTypePod) return sink.stringListQueryCol(q, core.LabelPodName.Key, fmt.Sprintf("Unable to list pods in namespace %q", namespace)) }
[ "func", "(", "sink", "*", "influxdbSink", ")", "GetPodsFromNamespace", "(", "namespace", "string", ")", "(", "[", "]", "string", ",", "error", ")", "{", "if", "!", "nameAllowedChars", ".", "MatchString", "(", "namespace", ")", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "namespace", ")", "\n", "}", "\n", "// This is a bit difficult for the influx query language, so we cheat a bit here --", "// we just get all series for the uptime measurement for pods which match our namespace", "// (any measurement should work here, though)", "q", ":=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "core", ".", "MetricUptime", ".", "MetricDescriptor", ".", "Name", ",", "core", ".", "LabelNamespaceName", ".", "Key", ",", "namespace", ",", "core", ".", "MetricSetTypePod", ")", "\n", "return", "sink", ".", "stringListQueryCol", "(", "q", ",", "core", ".", "LabelPodName", ".", "Key", ",", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "namespace", ")", ")", "\n", "}" ]
// GetPodsFromNamespace retrieves the list of pods in a given namespace
[ "GetPodsFromNamespace", "retrieves", "the", "list", "of", "pods", "in", "a", "given", "namespace" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L588-L597
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
GetSystemContainersFromNode
func (sink *influxdbSink) GetSystemContainersFromNode(node string) ([]string, error) { if !nameAllowedChars.MatchString(node) { return nil, fmt.Errorf("Invalid node name %q", node) } // This is a bit difficult for the influx query language, so we cheat a bit here -- // we just get all series for the uptime measurement for system containers on our node // (any measurement should work here, though) q := fmt.Sprintf("SHOW SERIES FROM %q WHERE %s = '%s' AND type = '%s'", core.MetricUptime.MetricDescriptor.Name, core.LabelNodename.Key, node, core.MetricSetTypeSystemContainer) return sink.stringListQueryCol(q, core.LabelContainerName.Key, fmt.Sprintf("Unable to list system containers on node %q", node)) }
go
func (sink *influxdbSink) GetSystemContainersFromNode(node string) ([]string, error) { if !nameAllowedChars.MatchString(node) { return nil, fmt.Errorf("Invalid node name %q", node) } // This is a bit difficult for the influx query language, so we cheat a bit here -- // we just get all series for the uptime measurement for system containers on our node // (any measurement should work here, though) q := fmt.Sprintf("SHOW SERIES FROM %q WHERE %s = '%s' AND type = '%s'", core.MetricUptime.MetricDescriptor.Name, core.LabelNodename.Key, node, core.MetricSetTypeSystemContainer) return sink.stringListQueryCol(q, core.LabelContainerName.Key, fmt.Sprintf("Unable to list system containers on node %q", node)) }
[ "func", "(", "sink", "*", "influxdbSink", ")", "GetSystemContainersFromNode", "(", "node", "string", ")", "(", "[", "]", "string", ",", "error", ")", "{", "if", "!", "nameAllowedChars", ".", "MatchString", "(", "node", ")", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "node", ")", "\n", "}", "\n", "// This is a bit difficult for the influx query language, so we cheat a bit here --", "// we just get all series for the uptime measurement for system containers on our node", "// (any measurement should work here, though)", "q", ":=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "core", ".", "MetricUptime", ".", "MetricDescriptor", ".", "Name", ",", "core", ".", "LabelNodename", ".", "Key", ",", "node", ",", "core", ".", "MetricSetTypeSystemContainer", ")", "\n", "return", "sink", ".", "stringListQueryCol", "(", "q", ",", "core", ".", "LabelContainerName", ".", "Key", ",", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "node", ")", ")", "\n", "}" ]
// GetSystemContainersFromNode retrieves the list of free containers for a given node
[ "GetSystemContainersFromNode", "retrieves", "the", "list", "of", "free", "containers", "for", "a", "given", "node" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L600-L609
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
stringListQueryCol
func (sink *influxdbSink) stringListQueryCol(q, colName string, errStr string) ([]string, error) { sink.RLock() defer sink.RUnlock() resp, err := sink.runQuery(q) if err != nil { return nil, fmt.Errorf(errStr) } if len(resp[0].Series) < 1 { return nil, fmt.Errorf(errStr) } colInd := -1 for i, col := range resp[0].Series[0].Columns { if col == colName { colInd = i break } } if colInd == -1 { glog.Errorf("%s: results did not contain the %q column", errStr, core.LabelPodName.Key) return nil, fmt.Errorf(errStr) } res := make([]string, len(resp[0].Series[0].Values)) for i, rv := range resp[0].Series[0].Values { res[i] = rv[colInd].(string) } return res, nil }
go
func (sink *influxdbSink) stringListQueryCol(q, colName string, errStr string) ([]string, error) { sink.RLock() defer sink.RUnlock() resp, err := sink.runQuery(q) if err != nil { return nil, fmt.Errorf(errStr) } if len(resp[0].Series) < 1 { return nil, fmt.Errorf(errStr) } colInd := -1 for i, col := range resp[0].Series[0].Columns { if col == colName { colInd = i break } } if colInd == -1 { glog.Errorf("%s: results did not contain the %q column", errStr, core.LabelPodName.Key) return nil, fmt.Errorf(errStr) } res := make([]string, len(resp[0].Series[0].Values)) for i, rv := range resp[0].Series[0].Values { res[i] = rv[colInd].(string) } return res, nil }
[ "func", "(", "sink", "*", "influxdbSink", ")", "stringListQueryCol", "(", "q", ",", "colName", "string", ",", "errStr", "string", ")", "(", "[", "]", "string", ",", "error", ")", "{", "sink", ".", "RLock", "(", ")", "\n", "defer", "sink", ".", "RUnlock", "(", ")", "\n\n", "resp", ",", "err", ":=", "sink", ".", "runQuery", "(", "q", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "errStr", ")", "\n", "}", "\n\n", "if", "len", "(", "resp", "[", "0", "]", ".", "Series", ")", "<", "1", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "errStr", ")", "\n", "}", "\n\n", "colInd", ":=", "-", "1", "\n", "for", "i", ",", "col", ":=", "range", "resp", "[", "0", "]", ".", "Series", "[", "0", "]", ".", "Columns", "{", "if", "col", "==", "colName", "{", "colInd", "=", "i", "\n", "break", "\n", "}", "\n", "}", "\n\n", "if", "colInd", "==", "-", "1", "{", "glog", ".", "Errorf", "(", "\"", "\"", ",", "errStr", ",", "core", ".", "LabelPodName", ".", "Key", ")", "\n", "return", "nil", ",", "fmt", ".", "Errorf", "(", "errStr", ")", "\n", "}", "\n\n", "res", ":=", "make", "(", "[", "]", "string", ",", "len", "(", "resp", "[", "0", "]", ".", "Series", "[", "0", "]", ".", "Values", ")", ")", "\n", "for", "i", ",", "rv", ":=", "range", "resp", "[", "0", "]", ".", "Series", "[", "0", "]", ".", "Values", "{", "res", "[", "i", "]", "=", "rv", "[", "colInd", "]", ".", "(", "string", ")", "\n", "}", "\n", "return", "res", ",", "nil", "\n", "}" ]
// stringListQueryCol runs the given query, and returns all results from the given column as a string list
[ "stringListQueryCol", "runs", "the", "given", "query", "and", "returns", "all", "results", "from", "the", "given", "column", "as", "a", "string", "list" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L612-L643
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
stringListQuery
func (sink *influxdbSink) stringListQuery(q string, errStr string) ([]string, error) { sink.RLock() defer sink.RUnlock() resp, err := sink.runQuery(q) if err != nil { return nil, fmt.Errorf(errStr) } if len(resp[0].Series) < 1 { return nil, fmt.Errorf(errStr) } res := make([]string, len(resp[0].Series[0].Values)) for i, rv := range resp[0].Series[0].Values { res[i] = rv[0].(string) } return res, nil }
go
func (sink *influxdbSink) stringListQuery(q string, errStr string) ([]string, error) { sink.RLock() defer sink.RUnlock() resp, err := sink.runQuery(q) if err != nil { return nil, fmt.Errorf(errStr) } if len(resp[0].Series) < 1 { return nil, fmt.Errorf(errStr) } res := make([]string, len(resp[0].Series[0].Values)) for i, rv := range resp[0].Series[0].Values { res[i] = rv[0].(string) } return res, nil }
[ "func", "(", "sink", "*", "influxdbSink", ")", "stringListQuery", "(", "q", "string", ",", "errStr", "string", ")", "(", "[", "]", "string", ",", "error", ")", "{", "sink", ".", "RLock", "(", ")", "\n", "defer", "sink", ".", "RUnlock", "(", ")", "\n\n", "resp", ",", "err", ":=", "sink", ".", "runQuery", "(", "q", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "errStr", ")", "\n", "}", "\n\n", "if", "len", "(", "resp", "[", "0", "]", ".", "Series", ")", "<", "1", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "errStr", ")", "\n", "}", "\n\n", "res", ":=", "make", "(", "[", "]", "string", ",", "len", "(", "resp", "[", "0", "]", ".", "Series", "[", "0", "]", ".", "Values", ")", ")", "\n", "for", "i", ",", "rv", ":=", "range", "resp", "[", "0", "]", ".", "Series", "[", "0", "]", ".", "Values", "{", "res", "[", "i", "]", "=", "rv", "[", "0", "]", ".", "(", "string", ")", "\n", "}", "\n", "return", "res", ",", "nil", "\n", "}" ]
// stringListQuery runs the given query, and returns all results from the first column as a string list
[ "stringListQuery", "runs", "the", "given", "query", "and", "returns", "all", "results", "from", "the", "first", "column", "as", "a", "string", "list" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L646-L664
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
populateAggregations
func populateAggregations(rawRowName string, rawVal []interface{}, val *core.TimestampedAggregationValue, aggregationLookup map[core.AggregationType]int, wasInt map[string]bool) error { for _, aggregation := range core.MultiTypedAggregations { if err := setAggregationValueIfPresent(aggregation, rawVal, &val.AggregationValue, aggregationLookup, wasInt); err != nil { glog.Errorf("Unable to parse field %q in series %q: %v", aggregation, rawRowName, err) return fmt.Errorf("Unable to parse values in series %q", rawRowName) } } return nil }
go
func populateAggregations(rawRowName string, rawVal []interface{}, val *core.TimestampedAggregationValue, aggregationLookup map[core.AggregationType]int, wasInt map[string]bool) error { for _, aggregation := range core.MultiTypedAggregations { if err := setAggregationValueIfPresent(aggregation, rawVal, &val.AggregationValue, aggregationLookup, wasInt); err != nil { glog.Errorf("Unable to parse field %q in series %q: %v", aggregation, rawRowName, err) return fmt.Errorf("Unable to parse values in series %q", rawRowName) } } return nil }
[ "func", "populateAggregations", "(", "rawRowName", "string", ",", "rawVal", "[", "]", "interface", "{", "}", ",", "val", "*", "core", ".", "TimestampedAggregationValue", ",", "aggregationLookup", "map", "[", "core", ".", "AggregationType", "]", "int", ",", "wasInt", "map", "[", "string", "]", "bool", ")", "error", "{", "for", "_", ",", "aggregation", ":=", "range", "core", ".", "MultiTypedAggregations", "{", "if", "err", ":=", "setAggregationValueIfPresent", "(", "aggregation", ",", "rawVal", ",", "&", "val", ".", "AggregationValue", ",", "aggregationLookup", ",", "wasInt", ")", ";", "err", "!=", "nil", "{", "glog", ".", "Errorf", "(", "\"", "\"", ",", "aggregation", ",", "rawRowName", ",", "err", ")", "\n", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "rawRowName", ")", "\n", "}", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// populateAggregations extracts aggregation values from a given data point
[ "populateAggregations", "extracts", "aggregation", "values", "from", "a", "given", "data", "point" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L700-L709
train
kubernetes-retired/heapster
metrics/sinks/influxdb/influxdb_historical.go
setAggregationValueTypes
func setAggregationValueTypes(vals []core.TimestampedAggregationValue, wasInt map[string]bool) { for _, aggregation := range core.MultiTypedAggregations { if isInt, ok := wasInt[string(aggregation)]; ok && isInt { for i := range vals { val := vals[i].Aggregations[aggregation] val.ValueType = core.ValueInt64 vals[i].Aggregations[aggregation] = val } } else if ok { for i := range vals { val := vals[i].Aggregations[aggregation] val.ValueType = core.ValueFloat vals[i].Aggregations[aggregation] = val } } } }
go
func setAggregationValueTypes(vals []core.TimestampedAggregationValue, wasInt map[string]bool) { for _, aggregation := range core.MultiTypedAggregations { if isInt, ok := wasInt[string(aggregation)]; ok && isInt { for i := range vals { val := vals[i].Aggregations[aggregation] val.ValueType = core.ValueInt64 vals[i].Aggregations[aggregation] = val } } else if ok { for i := range vals { val := vals[i].Aggregations[aggregation] val.ValueType = core.ValueFloat vals[i].Aggregations[aggregation] = val } } } }
[ "func", "setAggregationValueTypes", "(", "vals", "[", "]", "core", ".", "TimestampedAggregationValue", ",", "wasInt", "map", "[", "string", "]", "bool", ")", "{", "for", "_", ",", "aggregation", ":=", "range", "core", ".", "MultiTypedAggregations", "{", "if", "isInt", ",", "ok", ":=", "wasInt", "[", "string", "(", "aggregation", ")", "]", ";", "ok", "&&", "isInt", "{", "for", "i", ":=", "range", "vals", "{", "val", ":=", "vals", "[", "i", "]", ".", "Aggregations", "[", "aggregation", "]", "\n", "val", ".", "ValueType", "=", "core", ".", "ValueInt64", "\n", "vals", "[", "i", "]", ".", "Aggregations", "[", "aggregation", "]", "=", "val", "\n", "}", "\n", "}", "else", "if", "ok", "{", "for", "i", ":=", "range", "vals", "{", "val", ":=", "vals", "[", "i", "]", ".", "Aggregations", "[", "aggregation", "]", "\n", "val", ".", "ValueType", "=", "core", ".", "ValueFloat", "\n", "vals", "[", "i", "]", ".", "Aggregations", "[", "aggregation", "]", "=", "val", "\n", "}", "\n", "}", "\n", "}", "\n", "}" ]
// setAggregationValueTypes inspects a set of aggregation values and figures out whether each aggregation value // returned as a float column, or an int column
[ "setAggregationValueTypes", "inspects", "a", "set", "of", "aggregation", "values", "and", "figures", "out", "whether", "each", "aggregation", "value", "returned", "as", "a", "float", "column", "or", "an", "int", "column" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/influxdb/influxdb_historical.go#L713-L729
train
kubernetes-retired/heapster
metrics/sinks/gcm/gcm.go
register
func (sink *gcmSink) register(metrics []core.Metric) error { sink.Lock() defer sink.Unlock() if sink.registered { return nil } for _, metric := range metrics { metricName := fullMetricName(sink.project, metric.MetricDescriptor.Name) metricType := fullMetricType(metric.MetricDescriptor.Name) if _, err := sink.gcmService.Projects.MetricDescriptors.Delete(metricName).Do(); err != nil { glog.Infof("[GCM] Deleting metric %v failed: %v", metricName, err) } labels := make([]*gcm.LabelDescriptor, 0) // Node autoscaling metrics have special labels. if core.IsNodeAutoscalingMetric(metric.MetricDescriptor.Name) { // All and autoscaling. Do not populate for other filters. if sink.metricFilter != metricsAll && sink.metricFilter != metricsOnlyAutoscaling { continue } for _, l := range core.GcmNodeAutoscalingLabels() { labels = append(labels, &gcm.LabelDescriptor{ Key: l.Key, Description: l.Description, }) } } else { // Only all. if sink.metricFilter != metricsAll { continue } for _, l := range core.GcmLabels() { labels = append(labels, &gcm.LabelDescriptor{ Key: l.Key, Description: l.Description, }) } } var metricKind string switch metric.MetricDescriptor.Type { case core.MetricCumulative: metricKind = "CUMULATIVE" case core.MetricGauge: metricKind = "GAUGE" case core.MetricDelta: metricKind = "DELTA" } var valueType string switch metric.MetricDescriptor.ValueType { case core.ValueInt64: valueType = "INT64" case core.ValueFloat: valueType = "DOUBLE" } desc := &gcm.MetricDescriptor{ Name: metricName, Description: metric.MetricDescriptor.Description, Labels: labels, MetricKind: metricKind, ValueType: valueType, Type: metricType, } if _, err := sink.gcmService.Projects.MetricDescriptors.Create(fullProjectName(sink.project), desc).Do(); err != nil { glog.Errorf("Metric registration of %v failed: %v", desc.Name, err) return err } } sink.registered = true return nil }
go
func (sink *gcmSink) register(metrics []core.Metric) error { sink.Lock() defer sink.Unlock() if sink.registered { return nil } for _, metric := range metrics { metricName := fullMetricName(sink.project, metric.MetricDescriptor.Name) metricType := fullMetricType(metric.MetricDescriptor.Name) if _, err := sink.gcmService.Projects.MetricDescriptors.Delete(metricName).Do(); err != nil { glog.Infof("[GCM] Deleting metric %v failed: %v", metricName, err) } labels := make([]*gcm.LabelDescriptor, 0) // Node autoscaling metrics have special labels. if core.IsNodeAutoscalingMetric(metric.MetricDescriptor.Name) { // All and autoscaling. Do not populate for other filters. if sink.metricFilter != metricsAll && sink.metricFilter != metricsOnlyAutoscaling { continue } for _, l := range core.GcmNodeAutoscalingLabels() { labels = append(labels, &gcm.LabelDescriptor{ Key: l.Key, Description: l.Description, }) } } else { // Only all. if sink.metricFilter != metricsAll { continue } for _, l := range core.GcmLabels() { labels = append(labels, &gcm.LabelDescriptor{ Key: l.Key, Description: l.Description, }) } } var metricKind string switch metric.MetricDescriptor.Type { case core.MetricCumulative: metricKind = "CUMULATIVE" case core.MetricGauge: metricKind = "GAUGE" case core.MetricDelta: metricKind = "DELTA" } var valueType string switch metric.MetricDescriptor.ValueType { case core.ValueInt64: valueType = "INT64" case core.ValueFloat: valueType = "DOUBLE" } desc := &gcm.MetricDescriptor{ Name: metricName, Description: metric.MetricDescriptor.Description, Labels: labels, MetricKind: metricKind, ValueType: valueType, Type: metricType, } if _, err := sink.gcmService.Projects.MetricDescriptors.Create(fullProjectName(sink.project), desc).Do(); err != nil { glog.Errorf("Metric registration of %v failed: %v", desc.Name, err) return err } } sink.registered = true return nil }
[ "func", "(", "sink", "*", "gcmSink", ")", "register", "(", "metrics", "[", "]", "core", ".", "Metric", ")", "error", "{", "sink", ".", "Lock", "(", ")", "\n", "defer", "sink", ".", "Unlock", "(", ")", "\n", "if", "sink", ".", "registered", "{", "return", "nil", "\n", "}", "\n\n", "for", "_", ",", "metric", ":=", "range", "metrics", "{", "metricName", ":=", "fullMetricName", "(", "sink", ".", "project", ",", "metric", ".", "MetricDescriptor", ".", "Name", ")", "\n", "metricType", ":=", "fullMetricType", "(", "metric", ".", "MetricDescriptor", ".", "Name", ")", "\n\n", "if", "_", ",", "err", ":=", "sink", ".", "gcmService", ".", "Projects", ".", "MetricDescriptors", ".", "Delete", "(", "metricName", ")", ".", "Do", "(", ")", ";", "err", "!=", "nil", "{", "glog", ".", "Infof", "(", "\"", "\"", ",", "metricName", ",", "err", ")", "\n", "}", "\n", "labels", ":=", "make", "(", "[", "]", "*", "gcm", ".", "LabelDescriptor", ",", "0", ")", "\n\n", "// Node autoscaling metrics have special labels.", "if", "core", ".", "IsNodeAutoscalingMetric", "(", "metric", ".", "MetricDescriptor", ".", "Name", ")", "{", "// All and autoscaling. Do not populate for other filters.", "if", "sink", ".", "metricFilter", "!=", "metricsAll", "&&", "sink", ".", "metricFilter", "!=", "metricsOnlyAutoscaling", "{", "continue", "\n", "}", "\n\n", "for", "_", ",", "l", ":=", "range", "core", ".", "GcmNodeAutoscalingLabels", "(", ")", "{", "labels", "=", "append", "(", "labels", ",", "&", "gcm", ".", "LabelDescriptor", "{", "Key", ":", "l", ".", "Key", ",", "Description", ":", "l", ".", "Description", ",", "}", ")", "\n", "}", "\n", "}", "else", "{", "// Only all.", "if", "sink", ".", "metricFilter", "!=", "metricsAll", "{", "continue", "\n", "}", "\n\n", "for", "_", ",", "l", ":=", "range", "core", ".", "GcmLabels", "(", ")", "{", "labels", "=", "append", "(", "labels", ",", "&", "gcm", ".", "LabelDescriptor", "{", "Key", ":", "l", ".", "Key", ",", "Description", ":", "l", ".", "Description", ",", "}", ")", "\n", "}", "\n", "}", "\n\n", "var", "metricKind", "string", "\n\n", "switch", "metric", ".", "MetricDescriptor", ".", "Type", "{", "case", "core", ".", "MetricCumulative", ":", "metricKind", "=", "\"", "\"", "\n", "case", "core", ".", "MetricGauge", ":", "metricKind", "=", "\"", "\"", "\n", "case", "core", ".", "MetricDelta", ":", "metricKind", "=", "\"", "\"", "\n", "}", "\n\n", "var", "valueType", "string", "\n\n", "switch", "metric", ".", "MetricDescriptor", ".", "ValueType", "{", "case", "core", ".", "ValueInt64", ":", "valueType", "=", "\"", "\"", "\n", "case", "core", ".", "ValueFloat", ":", "valueType", "=", "\"", "\"", "\n", "}", "\n\n", "desc", ":=", "&", "gcm", ".", "MetricDescriptor", "{", "Name", ":", "metricName", ",", "Description", ":", "metric", ".", "MetricDescriptor", ".", "Description", ",", "Labels", ":", "labels", ",", "MetricKind", ":", "metricKind", ",", "ValueType", ":", "valueType", ",", "Type", ":", "metricType", ",", "}", "\n\n", "if", "_", ",", "err", ":=", "sink", ".", "gcmService", ".", "Projects", ".", "MetricDescriptors", ".", "Create", "(", "fullProjectName", "(", "sink", ".", "project", ")", ",", "desc", ")", ".", "Do", "(", ")", ";", "err", "!=", "nil", "{", "glog", ".", "Errorf", "(", "\"", "\"", ",", "desc", ".", "Name", ",", "err", ")", "\n", "return", "err", "\n", "}", "\n", "}", "\n", "sink", ".", "registered", "=", "true", "\n", "return", "nil", "\n", "}" ]
// Adds the specified metrics or updates them if they already exist.
[ "Adds", "the", "specified", "metrics", "or", "updates", "them", "if", "they", "already", "exist", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/gcm/gcm.go#L217-L297
train
kubernetes-retired/heapster
metrics/sinks/manager.go
ExportData
func (this *sinkManager) ExportData(data *core.DataBatch) { var wg sync.WaitGroup for _, sh := range this.sinkHolders { wg.Add(1) go func(sh sinkHolder, wg *sync.WaitGroup) { defer wg.Done() glog.V(2).Infof("Pushing data to: %s", sh.sink.Name()) select { case sh.dataBatchChannel <- data: glog.V(2).Infof("Data push completed: %s", sh.sink.Name()) // everything ok case <-time.After(this.exportDataTimeout): glog.Warningf("Failed to push data to sink: %s", sh.sink.Name()) } }(sh, &wg) } // Wait for all pushes to complete or timeout. wg.Wait() }
go
func (this *sinkManager) ExportData(data *core.DataBatch) { var wg sync.WaitGroup for _, sh := range this.sinkHolders { wg.Add(1) go func(sh sinkHolder, wg *sync.WaitGroup) { defer wg.Done() glog.V(2).Infof("Pushing data to: %s", sh.sink.Name()) select { case sh.dataBatchChannel <- data: glog.V(2).Infof("Data push completed: %s", sh.sink.Name()) // everything ok case <-time.After(this.exportDataTimeout): glog.Warningf("Failed to push data to sink: %s", sh.sink.Name()) } }(sh, &wg) } // Wait for all pushes to complete or timeout. wg.Wait() }
[ "func", "(", "this", "*", "sinkManager", ")", "ExportData", "(", "data", "*", "core", ".", "DataBatch", ")", "{", "var", "wg", "sync", ".", "WaitGroup", "\n", "for", "_", ",", "sh", ":=", "range", "this", ".", "sinkHolders", "{", "wg", ".", "Add", "(", "1", ")", "\n", "go", "func", "(", "sh", "sinkHolder", ",", "wg", "*", "sync", ".", "WaitGroup", ")", "{", "defer", "wg", ".", "Done", "(", ")", "\n", "glog", ".", "V", "(", "2", ")", ".", "Infof", "(", "\"", "\"", ",", "sh", ".", "sink", ".", "Name", "(", ")", ")", "\n", "select", "{", "case", "sh", ".", "dataBatchChannel", "<-", "data", ":", "glog", ".", "V", "(", "2", ")", ".", "Infof", "(", "\"", "\"", ",", "sh", ".", "sink", ".", "Name", "(", ")", ")", "\n", "// everything ok", "case", "<-", "time", ".", "After", "(", "this", ".", "exportDataTimeout", ")", ":", "glog", ".", "Warningf", "(", "\"", "\"", ",", "sh", ".", "sink", ".", "Name", "(", ")", ")", "\n", "}", "\n", "}", "(", "sh", ",", "&", "wg", ")", "\n", "}", "\n", "// Wait for all pushes to complete or timeout.", "wg", ".", "Wait", "(", ")", "\n", "}" ]
// Guarantees that the export will complete in sinkExportDataTimeout.
[ "Guarantees", "that", "the", "export", "will", "complete", "in", "sinkExportDataTimeout", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/manager.go#L106-L124
train
kubernetes-retired/heapster
metrics/sources/kubelet/kubelet_client.go
GetAllRawContainers
func (self *KubeletClient) GetAllRawContainers(host Host, start, end time.Time) ([]cadvisor.ContainerInfo, error) { url := self.getUrl(host, "/stats/container/") return self.getAllContainers(url, start, end) }
go
func (self *KubeletClient) GetAllRawContainers(host Host, start, end time.Time) ([]cadvisor.ContainerInfo, error) { url := self.getUrl(host, "/stats/container/") return self.getAllContainers(url, start, end) }
[ "func", "(", "self", "*", "KubeletClient", ")", "GetAllRawContainers", "(", "host", "Host", ",", "start", ",", "end", "time", ".", "Time", ")", "(", "[", "]", "cadvisor", ".", "ContainerInfo", ",", "error", ")", "{", "url", ":=", "self", ".", "getUrl", "(", "host", ",", "\"", "\"", ")", "\n\n", "return", "self", ".", "getAllContainers", "(", "url", ",", "start", ",", "end", ")", "\n", "}" ]
// Get stats for all non-Kubernetes containers.
[ "Get", "stats", "for", "all", "non", "-", "Kubernetes", "containers", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/kubelet/kubelet_client.go#L152-L156
train
kubernetes-retired/heapster
metrics/util/metrics/http.go
InstrumentRouteFunc
func InstrumentRouteFunc(handlerName string, routeFunc restful.RouteFunction) restful.RouteFunction { opts := prometheus.SummaryOpts{ Subsystem: "http", ConstLabels: prometheus.Labels{"handler": handlerName}, } reqCnt := prometheus.NewCounterVec( prometheus.CounterOpts{ Subsystem: opts.Subsystem, Name: "requests_total", Help: "Total number of HTTP requests made.", ConstLabels: opts.ConstLabels, }, instLabels, ) opts.Name = "request_duration_microseconds" opts.Help = "The HTTP request latencies in microseconds." reqDur := prometheus.NewSummary(opts) opts.Name = "request_size_bytes" opts.Help = "The HTTP request sizes in bytes." reqSz := prometheus.NewSummary(opts) opts.Name = "response_size_bytes" opts.Help = "The HTTP response sizes in bytes." resSz := prometheus.NewSummary(opts) regReqCnt := prometheus.MustRegisterOrGet(reqCnt).(*prometheus.CounterVec) regReqDur := prometheus.MustRegisterOrGet(reqDur).(prometheus.Summary) regReqSz := prometheus.MustRegisterOrGet(reqSz).(prometheus.Summary) regResSz := prometheus.MustRegisterOrGet(resSz).(prometheus.Summary) return restful.RouteFunction(func(request *restful.Request, response *restful.Response) { now := time.Now() delegate := &responseWriterDelegator{ResponseWriter: response.ResponseWriter} out := make(chan int) urlLen := 0 if request.Request.URL != nil { urlLen = len(request.Request.URL.String()) } go computeApproximateRequestSize(request.Request, out, urlLen) _, cn := response.ResponseWriter.(http.CloseNotifier) _, fl := response.ResponseWriter.(http.Flusher) _, hj := response.ResponseWriter.(http.Hijacker) _, rf := response.ResponseWriter.(io.ReaderFrom) var rw http.ResponseWriter if cn && fl && hj && rf { rw = &fancyResponseWriterDelegator{delegate} } else { rw = delegate } response.ResponseWriter = rw routeFunc(request, response) elapsed := float64(time.Since(now)) / float64(time.Microsecond) method := strings.ToLower(request.Request.Method) code := strconv.Itoa(delegate.status) regReqCnt.WithLabelValues(method, code).Inc() regReqDur.Observe(elapsed) regResSz.Observe(float64(delegate.written)) regReqSz.Observe(float64(<-out)) }) }
go
func InstrumentRouteFunc(handlerName string, routeFunc restful.RouteFunction) restful.RouteFunction { opts := prometheus.SummaryOpts{ Subsystem: "http", ConstLabels: prometheus.Labels{"handler": handlerName}, } reqCnt := prometheus.NewCounterVec( prometheus.CounterOpts{ Subsystem: opts.Subsystem, Name: "requests_total", Help: "Total number of HTTP requests made.", ConstLabels: opts.ConstLabels, }, instLabels, ) opts.Name = "request_duration_microseconds" opts.Help = "The HTTP request latencies in microseconds." reqDur := prometheus.NewSummary(opts) opts.Name = "request_size_bytes" opts.Help = "The HTTP request sizes in bytes." reqSz := prometheus.NewSummary(opts) opts.Name = "response_size_bytes" opts.Help = "The HTTP response sizes in bytes." resSz := prometheus.NewSummary(opts) regReqCnt := prometheus.MustRegisterOrGet(reqCnt).(*prometheus.CounterVec) regReqDur := prometheus.MustRegisterOrGet(reqDur).(prometheus.Summary) regReqSz := prometheus.MustRegisterOrGet(reqSz).(prometheus.Summary) regResSz := prometheus.MustRegisterOrGet(resSz).(prometheus.Summary) return restful.RouteFunction(func(request *restful.Request, response *restful.Response) { now := time.Now() delegate := &responseWriterDelegator{ResponseWriter: response.ResponseWriter} out := make(chan int) urlLen := 0 if request.Request.URL != nil { urlLen = len(request.Request.URL.String()) } go computeApproximateRequestSize(request.Request, out, urlLen) _, cn := response.ResponseWriter.(http.CloseNotifier) _, fl := response.ResponseWriter.(http.Flusher) _, hj := response.ResponseWriter.(http.Hijacker) _, rf := response.ResponseWriter.(io.ReaderFrom) var rw http.ResponseWriter if cn && fl && hj && rf { rw = &fancyResponseWriterDelegator{delegate} } else { rw = delegate } response.ResponseWriter = rw routeFunc(request, response) elapsed := float64(time.Since(now)) / float64(time.Microsecond) method := strings.ToLower(request.Request.Method) code := strconv.Itoa(delegate.status) regReqCnt.WithLabelValues(method, code).Inc() regReqDur.Observe(elapsed) regResSz.Observe(float64(delegate.written)) regReqSz.Observe(float64(<-out)) }) }
[ "func", "InstrumentRouteFunc", "(", "handlerName", "string", ",", "routeFunc", "restful", ".", "RouteFunction", ")", "restful", ".", "RouteFunction", "{", "opts", ":=", "prometheus", ".", "SummaryOpts", "{", "Subsystem", ":", "\"", "\"", ",", "ConstLabels", ":", "prometheus", ".", "Labels", "{", "\"", "\"", ":", "handlerName", "}", ",", "}", "\n\n", "reqCnt", ":=", "prometheus", ".", "NewCounterVec", "(", "prometheus", ".", "CounterOpts", "{", "Subsystem", ":", "opts", ".", "Subsystem", ",", "Name", ":", "\"", "\"", ",", "Help", ":", "\"", "\"", ",", "ConstLabels", ":", "opts", ".", "ConstLabels", ",", "}", ",", "instLabels", ",", ")", "\n\n", "opts", ".", "Name", "=", "\"", "\"", "\n", "opts", ".", "Help", "=", "\"", "\"", "\n", "reqDur", ":=", "prometheus", ".", "NewSummary", "(", "opts", ")", "\n\n", "opts", ".", "Name", "=", "\"", "\"", "\n", "opts", ".", "Help", "=", "\"", "\"", "\n", "reqSz", ":=", "prometheus", ".", "NewSummary", "(", "opts", ")", "\n\n", "opts", ".", "Name", "=", "\"", "\"", "\n", "opts", ".", "Help", "=", "\"", "\"", "\n", "resSz", ":=", "prometheus", ".", "NewSummary", "(", "opts", ")", "\n\n", "regReqCnt", ":=", "prometheus", ".", "MustRegisterOrGet", "(", "reqCnt", ")", ".", "(", "*", "prometheus", ".", "CounterVec", ")", "\n", "regReqDur", ":=", "prometheus", ".", "MustRegisterOrGet", "(", "reqDur", ")", ".", "(", "prometheus", ".", "Summary", ")", "\n", "regReqSz", ":=", "prometheus", ".", "MustRegisterOrGet", "(", "reqSz", ")", ".", "(", "prometheus", ".", "Summary", ")", "\n", "regResSz", ":=", "prometheus", ".", "MustRegisterOrGet", "(", "resSz", ")", ".", "(", "prometheus", ".", "Summary", ")", "\n\n", "return", "restful", ".", "RouteFunction", "(", "func", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "now", ":=", "time", ".", "Now", "(", ")", "\n\n", "delegate", ":=", "&", "responseWriterDelegator", "{", "ResponseWriter", ":", "response", ".", "ResponseWriter", "}", "\n", "out", ":=", "make", "(", "chan", "int", ")", "\n", "urlLen", ":=", "0", "\n", "if", "request", ".", "Request", ".", "URL", "!=", "nil", "{", "urlLen", "=", "len", "(", "request", ".", "Request", ".", "URL", ".", "String", "(", ")", ")", "\n", "}", "\n", "go", "computeApproximateRequestSize", "(", "request", ".", "Request", ",", "out", ",", "urlLen", ")", "\n\n", "_", ",", "cn", ":=", "response", ".", "ResponseWriter", ".", "(", "http", ".", "CloseNotifier", ")", "\n", "_", ",", "fl", ":=", "response", ".", "ResponseWriter", ".", "(", "http", ".", "Flusher", ")", "\n", "_", ",", "hj", ":=", "response", ".", "ResponseWriter", ".", "(", "http", ".", "Hijacker", ")", "\n", "_", ",", "rf", ":=", "response", ".", "ResponseWriter", ".", "(", "io", ".", "ReaderFrom", ")", "\n", "var", "rw", "http", ".", "ResponseWriter", "\n", "if", "cn", "&&", "fl", "&&", "hj", "&&", "rf", "{", "rw", "=", "&", "fancyResponseWriterDelegator", "{", "delegate", "}", "\n", "}", "else", "{", "rw", "=", "delegate", "\n", "}", "\n", "response", ".", "ResponseWriter", "=", "rw", "\n\n", "routeFunc", "(", "request", ",", "response", ")", "\n\n", "elapsed", ":=", "float64", "(", "time", ".", "Since", "(", "now", ")", ")", "/", "float64", "(", "time", ".", "Microsecond", ")", "\n\n", "method", ":=", "strings", ".", "ToLower", "(", "request", ".", "Request", ".", "Method", ")", "\n", "code", ":=", "strconv", ".", "Itoa", "(", "delegate", ".", "status", ")", "\n", "regReqCnt", ".", "WithLabelValues", "(", "method", ",", "code", ")", ".", "Inc", "(", ")", "\n", "regReqDur", ".", "Observe", "(", "elapsed", ")", "\n", "regResSz", ".", "Observe", "(", "float64", "(", "delegate", ".", "written", ")", ")", "\n", "regReqSz", ".", "Observe", "(", "float64", "(", "<-", "out", ")", ")", "\n", "}", ")", "\n", "}" ]
// InstrumentRouteFunc works like Prometheus' InstrumentHandlerFunc but wraps // the go-restful RouteFunction instead of a HandlerFunc
[ "InstrumentRouteFunc", "works", "like", "Prometheus", "InstrumentHandlerFunc", "but", "wraps", "the", "go", "-", "restful", "RouteFunction", "instead", "of", "a", "HandlerFunc" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/util/metrics/http.go#L36-L103
train
kubernetes-retired/heapster
metrics/core/ms_keys.go
PodContainerKey
func PodContainerKey(namespace, podName, containerName string) string { return fmt.Sprintf("namespace:%s/pod:%s/container:%s", namespace, podName, containerName) }
go
func PodContainerKey(namespace, podName, containerName string) string { return fmt.Sprintf("namespace:%s/pod:%s/container:%s", namespace, podName, containerName) }
[ "func", "PodContainerKey", "(", "namespace", ",", "podName", ",", "containerName", "string", ")", "string", "{", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "namespace", ",", "podName", ",", "containerName", ")", "\n", "}" ]
// MetricsSet keys are inside of DataBatch. The structure of the returned string is // an implementation detail and no component should rely on it as it may change // anytime. It only guaranteed that it is unique for the unique combination of // passed parameters.
[ "MetricsSet", "keys", "are", "inside", "of", "DataBatch", ".", "The", "structure", "of", "the", "returned", "string", "is", "an", "implementation", "detail", "and", "no", "component", "should", "rely", "on", "it", "as", "it", "may", "change", "anytime", ".", "It", "only", "guaranteed", "that", "it", "is", "unique", "for", "the", "unique", "combination", "of", "passed", "parameters", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/core/ms_keys.go#L26-L28
train
kubernetes-retired/heapster
metrics/processors/namespace_based_enricher.go
addNamespaceInfo
func (this *NamespaceBasedEnricher) addNamespaceInfo(metricSet *core.MetricSet) { metricSetType, found := metricSet.Labels[core.LabelMetricSetType.Key] if !found { return } if metricSetType != core.MetricSetTypePodContainer && metricSetType != core.MetricSetTypePod && metricSetType != core.MetricSetTypeNamespace { return } namespaceName, found := metricSet.Labels[core.LabelNamespaceName.Key] if !found { return } nsObj, exists, err := this.store.GetByKey(namespaceName) if exists && err == nil { namespace, ok := nsObj.(*kube_api.Namespace) if ok { metricSet.Labels[core.LabelPodNamespaceUID.Key] = string(namespace.UID) } else { glog.Errorf("Wrong namespace store content") } } else if err != nil { glog.Warningf("Failed to get namespace %s: %v", namespaceName, err) } else if !exists { glog.Warningf("Namespace doesn't exist: %s", namespaceName) } }
go
func (this *NamespaceBasedEnricher) addNamespaceInfo(metricSet *core.MetricSet) { metricSetType, found := metricSet.Labels[core.LabelMetricSetType.Key] if !found { return } if metricSetType != core.MetricSetTypePodContainer && metricSetType != core.MetricSetTypePod && metricSetType != core.MetricSetTypeNamespace { return } namespaceName, found := metricSet.Labels[core.LabelNamespaceName.Key] if !found { return } nsObj, exists, err := this.store.GetByKey(namespaceName) if exists && err == nil { namespace, ok := nsObj.(*kube_api.Namespace) if ok { metricSet.Labels[core.LabelPodNamespaceUID.Key] = string(namespace.UID) } else { glog.Errorf("Wrong namespace store content") } } else if err != nil { glog.Warningf("Failed to get namespace %s: %v", namespaceName, err) } else if !exists { glog.Warningf("Namespace doesn't exist: %s", namespaceName) } }
[ "func", "(", "this", "*", "NamespaceBasedEnricher", ")", "addNamespaceInfo", "(", "metricSet", "*", "core", ".", "MetricSet", ")", "{", "metricSetType", ",", "found", ":=", "metricSet", ".", "Labels", "[", "core", ".", "LabelMetricSetType", ".", "Key", "]", "\n", "if", "!", "found", "{", "return", "\n", "}", "\n", "if", "metricSetType", "!=", "core", ".", "MetricSetTypePodContainer", "&&", "metricSetType", "!=", "core", ".", "MetricSetTypePod", "&&", "metricSetType", "!=", "core", ".", "MetricSetTypeNamespace", "{", "return", "\n", "}", "\n\n", "namespaceName", ",", "found", ":=", "metricSet", ".", "Labels", "[", "core", ".", "LabelNamespaceName", ".", "Key", "]", "\n", "if", "!", "found", "{", "return", "\n", "}", "\n\n", "nsObj", ",", "exists", ",", "err", ":=", "this", ".", "store", ".", "GetByKey", "(", "namespaceName", ")", "\n", "if", "exists", "&&", "err", "==", "nil", "{", "namespace", ",", "ok", ":=", "nsObj", ".", "(", "*", "kube_api", ".", "Namespace", ")", "\n", "if", "ok", "{", "metricSet", ".", "Labels", "[", "core", ".", "LabelPodNamespaceUID", ".", "Key", "]", "=", "string", "(", "namespace", ".", "UID", ")", "\n", "}", "else", "{", "glog", ".", "Errorf", "(", "\"", "\"", ")", "\n", "}", "\n", "}", "else", "if", "err", "!=", "nil", "{", "glog", ".", "Warningf", "(", "\"", "\"", ",", "namespaceName", ",", "err", ")", "\n", "}", "else", "if", "!", "exists", "{", "glog", ".", "Warningf", "(", "\"", "\"", ",", "namespaceName", ")", "\n", "}", "\n", "}" ]
// Adds UID to all namespaced elements.
[ "Adds", "UID", "to", "all", "namespaced", "elements", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/processors/namespace_based_enricher.go#L49-L78
train
kubernetes-retired/heapster
metrics/api/v1/model_handlers.go
availableClusterMetrics
func (a *Api) availableClusterMetrics(request *restful.Request, response *restful.Response) { a.processMetricNamesRequest(core.ClusterKey(), response) }
go
func (a *Api) availableClusterMetrics(request *restful.Request, response *restful.Response) { a.processMetricNamesRequest(core.ClusterKey(), response) }
[ "func", "(", "a", "*", "Api", ")", "availableClusterMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "a", ".", "processMetricNamesRequest", "(", "core", ".", "ClusterKey", "(", ")", ",", "response", ")", "\n", "}" ]
// availableMetrics returns a list of available cluster metric names.
[ "availableMetrics", "returns", "a", "list", "of", "available", "cluster", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/model_handlers.go#L283-L285
train
kubernetes-retired/heapster
metrics/api/v1/model_handlers.go
availableNodeMetrics
func (a *Api) availableNodeMetrics(request *restful.Request, response *restful.Response) { a.processMetricNamesRequest(core.NodeKey(request.PathParameter("node-name")), response) }
go
func (a *Api) availableNodeMetrics(request *restful.Request, response *restful.Response) { a.processMetricNamesRequest(core.NodeKey(request.PathParameter("node-name")), response) }
[ "func", "(", "a", "*", "Api", ")", "availableNodeMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "a", ".", "processMetricNamesRequest", "(", "core", ".", "NodeKey", "(", "request", ".", "PathParameter", "(", "\"", "\"", ")", ")", ",", "response", ")", "\n", "}" ]
// availableMetrics returns a list of available node metric names.
[ "availableMetrics", "returns", "a", "list", "of", "available", "node", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/model_handlers.go#L288-L290
train
kubernetes-retired/heapster
metrics/api/v1/model_handlers.go
availableNamespaceMetrics
func (a *Api) availableNamespaceMetrics(request *restful.Request, response *restful.Response) { a.processMetricNamesRequest(core.NamespaceKey(request.PathParameter("namespace-name")), response) }
go
func (a *Api) availableNamespaceMetrics(request *restful.Request, response *restful.Response) { a.processMetricNamesRequest(core.NamespaceKey(request.PathParameter("namespace-name")), response) }
[ "func", "(", "a", "*", "Api", ")", "availableNamespaceMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "a", ".", "processMetricNamesRequest", "(", "core", ".", "NamespaceKey", "(", "request", ".", "PathParameter", "(", "\"", "\"", ")", ")", ",", "response", ")", "\n", "}" ]
// availableMetrics returns a list of available namespace metric names.
[ "availableMetrics", "returns", "a", "list", "of", "available", "namespace", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/model_handlers.go#L293-L295
train
kubernetes-retired/heapster
metrics/api/v1/model_handlers.go
parseTimeParam
func parseTimeParam(queryParam string, defaultValue time.Time) (time.Time, error) { if queryParam != "" { reqStamp, err := time.Parse(time.RFC3339, queryParam) if err != nil { return time.Time{}, fmt.Errorf("timestamp argument cannot be parsed: %s", err) } return reqStamp, nil } return defaultValue, nil }
go
func parseTimeParam(queryParam string, defaultValue time.Time) (time.Time, error) { if queryParam != "" { reqStamp, err := time.Parse(time.RFC3339, queryParam) if err != nil { return time.Time{}, fmt.Errorf("timestamp argument cannot be parsed: %s", err) } return reqStamp, nil } return defaultValue, nil }
[ "func", "parseTimeParam", "(", "queryParam", "string", ",", "defaultValue", "time", ".", "Time", ")", "(", "time", ".", "Time", ",", "error", ")", "{", "if", "queryParam", "!=", "\"", "\"", "{", "reqStamp", ",", "err", ":=", "time", ".", "Parse", "(", "time", ".", "RFC3339", ",", "queryParam", ")", "\n", "if", "err", "!=", "nil", "{", "return", "time", ".", "Time", "{", "}", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "return", "reqStamp", ",", "nil", "\n", "}", "\n", "return", "defaultValue", ",", "nil", "\n", "}" ]
// parseRequestParam parses a time.Time from a named QueryParam, using the RFC3339 format.
[ "parseRequestParam", "parses", "a", "time", ".", "Time", "from", "a", "named", "QueryParam", "using", "the", "RFC3339", "format", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/model_handlers.go#L429-L438
train
kubernetes-retired/heapster
metrics/sinks/riemann/driver.go
appendEvent
func appendEvent(events []riemanngo.Event, sink *RiemannSink, host, name string, value interface{}, labels map[string]string, timestamp int64) []riemanngo.Event { event := riemanngo.Event{ Time: timestamp, Service: name, Host: host, Description: "", Attributes: labels, Metric: value, Ttl: sink.config.Ttl, State: sink.config.State, Tags: sink.config.Tags, } // state everywhere events = append(events, event) if len(events) >= sink.config.BatchSize { err := riemannCommon.SendData(sink.client, events) if err != nil { glog.Warningf("Error sending events to Riemann: %v", err) // client will reconnect later sink.client = nil } events = nil } return events }
go
func appendEvent(events []riemanngo.Event, sink *RiemannSink, host, name string, value interface{}, labels map[string]string, timestamp int64) []riemanngo.Event { event := riemanngo.Event{ Time: timestamp, Service: name, Host: host, Description: "", Attributes: labels, Metric: value, Ttl: sink.config.Ttl, State: sink.config.State, Tags: sink.config.Tags, } // state everywhere events = append(events, event) if len(events) >= sink.config.BatchSize { err := riemannCommon.SendData(sink.client, events) if err != nil { glog.Warningf("Error sending events to Riemann: %v", err) // client will reconnect later sink.client = nil } events = nil } return events }
[ "func", "appendEvent", "(", "events", "[", "]", "riemanngo", ".", "Event", ",", "sink", "*", "RiemannSink", ",", "host", ",", "name", "string", ",", "value", "interface", "{", "}", ",", "labels", "map", "[", "string", "]", "string", ",", "timestamp", "int64", ")", "[", "]", "riemanngo", ".", "Event", "{", "event", ":=", "riemanngo", ".", "Event", "{", "Time", ":", "timestamp", ",", "Service", ":", "name", ",", "Host", ":", "host", ",", "Description", ":", "\"", "\"", ",", "Attributes", ":", "labels", ",", "Metric", ":", "value", ",", "Ttl", ":", "sink", ".", "config", ".", "Ttl", ",", "State", ":", "sink", ".", "config", ".", "State", ",", "Tags", ":", "sink", ".", "config", ".", "Tags", ",", "}", "\n", "// state everywhere", "events", "=", "append", "(", "events", ",", "event", ")", "\n", "if", "len", "(", "events", ")", ">=", "sink", ".", "config", ".", "BatchSize", "{", "err", ":=", "riemannCommon", ".", "SendData", "(", "sink", ".", "client", ",", "events", ")", "\n", "if", "err", "!=", "nil", "{", "glog", ".", "Warningf", "(", "\"", "\"", ",", "err", ")", "\n", "// client will reconnect later", "sink", ".", "client", "=", "nil", "\n", "}", "\n", "events", "=", "nil", "\n", "}", "\n", "return", "events", "\n", "}" ]
// Receives a list of riemanngo.Event, the sink, and parameters. // Creates a new event using the parameters and the sink config, and add it into the Event list. // Can send events if events is full // Return the list.
[ "Receives", "a", "list", "of", "riemanngo", ".", "Event", "the", "sink", "and", "parameters", ".", "Creates", "a", "new", "event", "using", "the", "parameters", "and", "the", "sink", "config", "and", "add", "it", "into", "the", "Event", "list", ".", "Can", "send", "events", "if", "events", "is", "full", "Return", "the", "list", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/riemann/driver.go#L61-L85
train
kubernetes-retired/heapster
metrics/sinks/riemann/driver.go
ExportData
func (sink *RiemannSink) ExportData(dataBatch *core.DataBatch) { sink.Lock() defer sink.Unlock() if sink.client == nil { // the client could be nil here, so we reconnect client, err := riemannCommon.GetRiemannClient(sink.config) if err != nil { glog.Warningf("Riemann sink not connected: %v", err) return } sink.client = client } var events []riemanngo.Event for _, metricSet := range dataBatch.MetricSets { host := metricSet.Labels[core.LabelHostname.Key] for metricName, metricValue := range metricSet.MetricValues { if value := metricValue.GetValue(); value != nil { timestamp := dataBatch.Timestamp.Unix() // creates an event and add it to dataEvent events = appendEvent(events, sink, host, metricName, value, metricSet.Labels, timestamp) } } for _, metric := range metricSet.LabeledMetrics { if value := metric.GetValue(); value != nil { labels := make(map[string]string) for k, v := range metricSet.Labels { labels[k] = v } for k, v := range metric.Labels { labels[k] = v } timestamp := dataBatch.Timestamp.Unix() // creates an event and add it to dataEvent events = appendEvent(events, sink, host, metric.Name, value, labels, timestamp) } } } // Send events to Riemann if events is not empty if len(events) > 0 { err := riemannCommon.SendData(sink.client, events) if err != nil { glog.Warningf("Error sending events to Riemann: %v", err) // client will reconnect later sink.client = nil } } }
go
func (sink *RiemannSink) ExportData(dataBatch *core.DataBatch) { sink.Lock() defer sink.Unlock() if sink.client == nil { // the client could be nil here, so we reconnect client, err := riemannCommon.GetRiemannClient(sink.config) if err != nil { glog.Warningf("Riemann sink not connected: %v", err) return } sink.client = client } var events []riemanngo.Event for _, metricSet := range dataBatch.MetricSets { host := metricSet.Labels[core.LabelHostname.Key] for metricName, metricValue := range metricSet.MetricValues { if value := metricValue.GetValue(); value != nil { timestamp := dataBatch.Timestamp.Unix() // creates an event and add it to dataEvent events = appendEvent(events, sink, host, metricName, value, metricSet.Labels, timestamp) } } for _, metric := range metricSet.LabeledMetrics { if value := metric.GetValue(); value != nil { labels := make(map[string]string) for k, v := range metricSet.Labels { labels[k] = v } for k, v := range metric.Labels { labels[k] = v } timestamp := dataBatch.Timestamp.Unix() // creates an event and add it to dataEvent events = appendEvent(events, sink, host, metric.Name, value, labels, timestamp) } } } // Send events to Riemann if events is not empty if len(events) > 0 { err := riemannCommon.SendData(sink.client, events) if err != nil { glog.Warningf("Error sending events to Riemann: %v", err) // client will reconnect later sink.client = nil } } }
[ "func", "(", "sink", "*", "RiemannSink", ")", "ExportData", "(", "dataBatch", "*", "core", ".", "DataBatch", ")", "{", "sink", ".", "Lock", "(", ")", "\n", "defer", "sink", ".", "Unlock", "(", ")", "\n\n", "if", "sink", ".", "client", "==", "nil", "{", "// the client could be nil here, so we reconnect", "client", ",", "err", ":=", "riemannCommon", ".", "GetRiemannClient", "(", "sink", ".", "config", ")", "\n", "if", "err", "!=", "nil", "{", "glog", ".", "Warningf", "(", "\"", "\"", ",", "err", ")", "\n", "return", "\n", "}", "\n", "sink", ".", "client", "=", "client", "\n", "}", "\n\n", "var", "events", "[", "]", "riemanngo", ".", "Event", "\n\n", "for", "_", ",", "metricSet", ":=", "range", "dataBatch", ".", "MetricSets", "{", "host", ":=", "metricSet", ".", "Labels", "[", "core", ".", "LabelHostname", ".", "Key", "]", "\n", "for", "metricName", ",", "metricValue", ":=", "range", "metricSet", ".", "MetricValues", "{", "if", "value", ":=", "metricValue", ".", "GetValue", "(", ")", ";", "value", "!=", "nil", "{", "timestamp", ":=", "dataBatch", ".", "Timestamp", ".", "Unix", "(", ")", "\n", "// creates an event and add it to dataEvent", "events", "=", "appendEvent", "(", "events", ",", "sink", ",", "host", ",", "metricName", ",", "value", ",", "metricSet", ".", "Labels", ",", "timestamp", ")", "\n", "}", "\n", "}", "\n", "for", "_", ",", "metric", ":=", "range", "metricSet", ".", "LabeledMetrics", "{", "if", "value", ":=", "metric", ".", "GetValue", "(", ")", ";", "value", "!=", "nil", "{", "labels", ":=", "make", "(", "map", "[", "string", "]", "string", ")", "\n", "for", "k", ",", "v", ":=", "range", "metricSet", ".", "Labels", "{", "labels", "[", "k", "]", "=", "v", "\n", "}", "\n", "for", "k", ",", "v", ":=", "range", "metric", ".", "Labels", "{", "labels", "[", "k", "]", "=", "v", "\n", "}", "\n", "timestamp", ":=", "dataBatch", ".", "Timestamp", ".", "Unix", "(", ")", "\n", "// creates an event and add it to dataEvent", "events", "=", "appendEvent", "(", "events", ",", "sink", ",", "host", ",", "metric", ".", "Name", ",", "value", ",", "labels", ",", "timestamp", ")", "\n", "}", "\n", "}", "\n", "}", "\n", "// Send events to Riemann if events is not empty", "if", "len", "(", "events", ")", ">", "0", "{", "err", ":=", "riemannCommon", ".", "SendData", "(", "sink", ".", "client", ",", "events", ")", "\n", "if", "err", "!=", "nil", "{", "glog", ".", "Warningf", "(", "\"", "\"", ",", "err", ")", "\n", "// client will reconnect later", "sink", ".", "client", "=", "nil", "\n", "}", "\n", "}", "\n", "}" ]
// ExportData Send a collection of Timeseries to Riemann
[ "ExportData", "Send", "a", "collection", "of", "Timeseries", "to", "Riemann" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/riemann/driver.go#L88-L137
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
availableClusterMetrics
func (a *HistoricalApi) availableClusterMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ObjectType: core.MetricSetTypeCluster} a.processMetricNamesRequest(key, response) }
go
func (a *HistoricalApi) availableClusterMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ObjectType: core.MetricSetTypeCluster} a.processMetricNamesRequest(key, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "availableClusterMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeCluster", "}", "\n", "a", ".", "processMetricNamesRequest", "(", "key", ",", "response", ")", "\n", "}" ]
// availableClusterMetrics returns a list of available cluster metric names.
[ "availableClusterMetrics", "returns", "a", "list", "of", "available", "cluster", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L259-L262
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
availableNodeMetrics
func (a *HistoricalApi) availableNodeMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeNode, NodeName: request.PathParameter("node-name"), } a.processMetricNamesRequest(key, response) }
go
func (a *HistoricalApi) availableNodeMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeNode, NodeName: request.PathParameter("node-name"), } a.processMetricNamesRequest(key, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "availableNodeMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeNode", ",", "NodeName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processMetricNamesRequest", "(", "key", ",", "response", ")", "\n", "}" ]
// availableNodeMetrics returns a list of available node metric names.
[ "availableNodeMetrics", "returns", "a", "list", "of", "available", "node", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L265-L271
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
availableNamespaceMetrics
func (a *HistoricalApi) availableNamespaceMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeNamespace, NamespaceName: request.PathParameter("namespace-name"), } a.processMetricNamesRequest(key, response) }
go
func (a *HistoricalApi) availableNamespaceMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeNamespace, NamespaceName: request.PathParameter("namespace-name"), } a.processMetricNamesRequest(key, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "availableNamespaceMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeNamespace", ",", "NamespaceName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processMetricNamesRequest", "(", "key", ",", "response", ")", "\n", "}" ]
// availableNamespaceMetrics returns a list of available namespace metric names.
[ "availableNamespaceMetrics", "returns", "a", "list", "of", "available", "namespace", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L274-L280
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
availablePodMetrics
func (a *HistoricalApi) availablePodMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, NamespaceName: request.PathParameter("namespace-name"), PodName: request.PathParameter("pod-name"), } a.processMetricNamesRequest(key, response) }
go
func (a *HistoricalApi) availablePodMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, NamespaceName: request.PathParameter("namespace-name"), PodName: request.PathParameter("pod-name"), } a.processMetricNamesRequest(key, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "availablePodMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypePod", ",", "NamespaceName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "PodName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processMetricNamesRequest", "(", "key", ",", "response", ")", "\n", "}" ]
// availablePodMetrics returns a list of available pod metric names.
[ "availablePodMetrics", "returns", "a", "list", "of", "available", "pod", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L283-L290
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
availablePodContainerMetrics
func (a *HistoricalApi) availablePodContainerMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePodContainer, NamespaceName: request.PathParameter("namespace-name"), PodName: request.PathParameter("pod-name"), ContainerName: request.PathParameter("container-name"), } a.processMetricNamesRequest(key, response) }
go
func (a *HistoricalApi) availablePodContainerMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePodContainer, NamespaceName: request.PathParameter("namespace-name"), PodName: request.PathParameter("pod-name"), ContainerName: request.PathParameter("container-name"), } a.processMetricNamesRequest(key, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "availablePodContainerMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypePodContainer", ",", "NamespaceName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "PodName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "ContainerName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processMetricNamesRequest", "(", "key", ",", "response", ")", "\n", "}" ]
// availablePodContainerMetrics returns a list of available pod container metric names.
[ "availablePodContainerMetrics", "returns", "a", "list", "of", "available", "pod", "container", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L293-L301
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
availableFreeContainerMetrics
func (a *HistoricalApi) availableFreeContainerMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeSystemContainer, NodeName: request.PathParameter("node-name"), ContainerName: request.PathParameter("container-name"), } a.processMetricNamesRequest(key, response) }
go
func (a *HistoricalApi) availableFreeContainerMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeSystemContainer, NodeName: request.PathParameter("node-name"), ContainerName: request.PathParameter("container-name"), } a.processMetricNamesRequest(key, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "availableFreeContainerMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeSystemContainer", ",", "NodeName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "ContainerName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processMetricNamesRequest", "(", "key", ",", "response", ")", "\n", "}" ]
// availableFreeContainerMetrics returns a list of available pod metric names.
[ "availableFreeContainerMetrics", "returns", "a", "list", "of", "available", "pod", "metric", "names", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L304-L311
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
nodeList
func (a *HistoricalApi) nodeList(request *restful.Request, response *restful.Response) { if resp, err := a.historicalSource.GetNodes(); err != nil { response.WriteError(http.StatusInternalServerError, err) } else { response.WriteEntity(resp) } }
go
func (a *HistoricalApi) nodeList(request *restful.Request, response *restful.Response) { if resp, err := a.historicalSource.GetNodes(); err != nil { response.WriteError(http.StatusInternalServerError, err) } else { response.WriteEntity(resp) } }
[ "func", "(", "a", "*", "HistoricalApi", ")", "nodeList", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "if", "resp", ",", "err", ":=", "a", ".", "historicalSource", ".", "GetNodes", "(", ")", ";", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusInternalServerError", ",", "err", ")", "\n", "}", "else", "{", "response", ".", "WriteEntity", "(", "resp", ")", "\n", "}", "\n", "}" ]
// nodeList lists all nodes for which we have metrics
[ "nodeList", "lists", "all", "nodes", "for", "which", "we", "have", "metrics" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L314-L320
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
namespaceList
func (a *HistoricalApi) namespaceList(request *restful.Request, response *restful.Response) { if resp, err := a.historicalSource.GetNamespaces(); err != nil { response.WriteError(http.StatusInternalServerError, err) } else { response.WriteEntity(resp) } }
go
func (a *HistoricalApi) namespaceList(request *restful.Request, response *restful.Response) { if resp, err := a.historicalSource.GetNamespaces(); err != nil { response.WriteError(http.StatusInternalServerError, err) } else { response.WriteEntity(resp) } }
[ "func", "(", "a", "*", "HistoricalApi", ")", "namespaceList", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "if", "resp", ",", "err", ":=", "a", ".", "historicalSource", ".", "GetNamespaces", "(", ")", ";", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusInternalServerError", ",", "err", ")", "\n", "}", "else", "{", "response", ".", "WriteEntity", "(", "resp", ")", "\n", "}", "\n", "}" ]
// namespaceList lists all namespaces for which we have metrics
[ "namespaceList", "lists", "all", "namespaces", "for", "which", "we", "have", "metrics" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L323-L329
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
namespacePodList
func (a *HistoricalApi) namespacePodList(request *restful.Request, response *restful.Response) { if resp, err := a.historicalSource.GetPodsFromNamespace(request.PathParameter("namespace-name")); err != nil { response.WriteError(http.StatusInternalServerError, err) } else { response.WriteEntity(resp) } }
go
func (a *HistoricalApi) namespacePodList(request *restful.Request, response *restful.Response) { if resp, err := a.historicalSource.GetPodsFromNamespace(request.PathParameter("namespace-name")); err != nil { response.WriteError(http.StatusInternalServerError, err) } else { response.WriteEntity(resp) } }
[ "func", "(", "a", "*", "HistoricalApi", ")", "namespacePodList", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "if", "resp", ",", "err", ":=", "a", ".", "historicalSource", ".", "GetPodsFromNamespace", "(", "request", ".", "PathParameter", "(", "\"", "\"", ")", ")", ";", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusInternalServerError", ",", "err", ")", "\n", "}", "else", "{", "response", ".", "WriteEntity", "(", "resp", ")", "\n", "}", "\n", "}" ]
// namespacePodList lists all pods for which we have metrics in a particular namespace
[ "namespacePodList", "lists", "all", "pods", "for", "which", "we", "have", "metrics", "in", "a", "particular", "namespace" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L332-L338
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
freeContainerMetrics
func (a *HistoricalApi) freeContainerMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeSystemContainer, NodeName: request.PathParameter("node-name"), ContainerName: request.PathParameter("container-name"), } a.processMetricRequest(key, request, response) }
go
func (a *HistoricalApi) freeContainerMetrics(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeSystemContainer, NodeName: request.PathParameter("node-name"), ContainerName: request.PathParameter("container-name"), } a.processMetricRequest(key, request, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "freeContainerMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeSystemContainer", ",", "NodeName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "ContainerName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processMetricRequest", "(", "key", ",", "request", ",", "response", ")", "\n", "}" ]
// freeContainerMetrics returns a metric timeseries for a metric of the Container entity. // freeContainerMetrics addresses only free containers.
[ "freeContainerMetrics", "returns", "a", "metric", "timeseries", "for", "a", "metric", "of", "the", "Container", "entity", ".", "freeContainerMetrics", "addresses", "only", "free", "containers", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L393-L400
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
podListMetrics
func (a *HistoricalApi) podListMetrics(request *restful.Request, response *restful.Response) { start, end, err := getStartEndTimeHistorical(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } keys := []core.HistoricalKey{} if request.PathParameter("pod-id-list") != "" { for _, podId := range strings.Split(request.PathParameter("pod-id-list"), ",") { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, PodId: podId, } keys = append(keys, key) } } else { for _, podName := range strings.Split(request.PathParameter("pod-list"), ",") { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, NamespaceName: request.PathParameter("namespace-name"), PodName: podName, } keys = append(keys, key) } } labels, err := getLabels(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } metricName := request.PathParameter("metric-name") convertedMetricName := convertMetricName(metricName) var metrics map[core.HistoricalKey][]core.TimestampedMetricValue if labels != nil { metrics, err = a.historicalSource.GetLabeledMetric(convertedMetricName, labels, keys, start, end) } else { metrics, err = a.historicalSource.GetMetric(convertedMetricName, keys, start, end) } if err != nil { response.WriteError(http.StatusInternalServerError, err) return } result := types.MetricResultList{ Items: make([]types.MetricResult, 0, len(keys)), } for _, key := range keys { result.Items = append(result.Items, exportTimestampedMetricValue(metrics[key])) } response.PrettyPrint(false) response.WriteEntity(result) }
go
func (a *HistoricalApi) podListMetrics(request *restful.Request, response *restful.Response) { start, end, err := getStartEndTimeHistorical(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } keys := []core.HistoricalKey{} if request.PathParameter("pod-id-list") != "" { for _, podId := range strings.Split(request.PathParameter("pod-id-list"), ",") { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, PodId: podId, } keys = append(keys, key) } } else { for _, podName := range strings.Split(request.PathParameter("pod-list"), ",") { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, NamespaceName: request.PathParameter("namespace-name"), PodName: podName, } keys = append(keys, key) } } labels, err := getLabels(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } metricName := request.PathParameter("metric-name") convertedMetricName := convertMetricName(metricName) var metrics map[core.HistoricalKey][]core.TimestampedMetricValue if labels != nil { metrics, err = a.historicalSource.GetLabeledMetric(convertedMetricName, labels, keys, start, end) } else { metrics, err = a.historicalSource.GetMetric(convertedMetricName, keys, start, end) } if err != nil { response.WriteError(http.StatusInternalServerError, err) return } result := types.MetricResultList{ Items: make([]types.MetricResult, 0, len(keys)), } for _, key := range keys { result.Items = append(result.Items, exportTimestampedMetricValue(metrics[key])) } response.PrettyPrint(false) response.WriteEntity(result) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "podListMetrics", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "start", ",", "end", ",", "err", ":=", "getStartEndTimeHistorical", "(", "request", ")", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusBadRequest", ",", "err", ")", "\n", "return", "\n", "}", "\n\n", "keys", ":=", "[", "]", "core", ".", "HistoricalKey", "{", "}", "\n", "if", "request", ".", "PathParameter", "(", "\"", "\"", ")", "!=", "\"", "\"", "{", "for", "_", ",", "podId", ":=", "range", "strings", ".", "Split", "(", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "\"", "\"", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypePod", ",", "PodId", ":", "podId", ",", "}", "\n", "keys", "=", "append", "(", "keys", ",", "key", ")", "\n", "}", "\n", "}", "else", "{", "for", "_", ",", "podName", ":=", "range", "strings", ".", "Split", "(", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "\"", "\"", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypePod", ",", "NamespaceName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "PodName", ":", "podName", ",", "}", "\n", "keys", "=", "append", "(", "keys", ",", "key", ")", "\n", "}", "\n", "}", "\n\n", "labels", ",", "err", ":=", "getLabels", "(", "request", ")", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusBadRequest", ",", "err", ")", "\n", "return", "\n", "}", "\n\n", "metricName", ":=", "request", ".", "PathParameter", "(", "\"", "\"", ")", "\n", "convertedMetricName", ":=", "convertMetricName", "(", "metricName", ")", "\n\n", "var", "metrics", "map", "[", "core", ".", "HistoricalKey", "]", "[", "]", "core", ".", "TimestampedMetricValue", "\n", "if", "labels", "!=", "nil", "{", "metrics", ",", "err", "=", "a", ".", "historicalSource", ".", "GetLabeledMetric", "(", "convertedMetricName", ",", "labels", ",", "keys", ",", "start", ",", "end", ")", "\n", "}", "else", "{", "metrics", ",", "err", "=", "a", ".", "historicalSource", ".", "GetMetric", "(", "convertedMetricName", ",", "keys", ",", "start", ",", "end", ")", "\n", "}", "\n\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusInternalServerError", ",", "err", ")", "\n", "return", "\n", "}", "\n\n", "result", ":=", "types", ".", "MetricResultList", "{", "Items", ":", "make", "(", "[", "]", "types", ".", "MetricResult", ",", "0", ",", "len", "(", "keys", ")", ")", ",", "}", "\n", "for", "_", ",", "key", ":=", "range", "keys", "{", "result", ".", "Items", "=", "append", "(", "result", ".", "Items", ",", "exportTimestampedMetricValue", "(", "metrics", "[", "key", "]", ")", ")", "\n", "}", "\n", "response", ".", "PrettyPrint", "(", "false", ")", "\n", "response", ".", "WriteEntity", "(", "result", ")", "\n", "}" ]
// podListMetrics returns a list of metric timeseries for each for the listed nodes
[ "podListMetrics", "returns", "a", "list", "of", "metric", "timeseries", "for", "each", "for", "the", "listed", "nodes" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L403-L459
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
clusterAggregations
func (a *HistoricalApi) clusterAggregations(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ObjectType: core.MetricSetTypeCluster} a.processAggregationRequest(key, request, response) }
go
func (a *HistoricalApi) clusterAggregations(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ObjectType: core.MetricSetTypeCluster} a.processAggregationRequest(key, request, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "clusterAggregations", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeCluster", "}", "\n", "a", ".", "processAggregationRequest", "(", "key", ",", "request", ",", "response", ")", "\n", "}" ]
// clusterAggregations returns a metric timeseries for a metric of the Cluster entity.
[ "clusterAggregations", "returns", "a", "metric", "timeseries", "for", "a", "metric", "of", "the", "Cluster", "entity", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L482-L485
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
nodeAggregations
func (a *HistoricalApi) nodeAggregations(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeNode, NodeName: request.PathParameter("node-name"), } a.processAggregationRequest(key, request, response) }
go
func (a *HistoricalApi) nodeAggregations(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeNode, NodeName: request.PathParameter("node-name"), } a.processAggregationRequest(key, request, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "nodeAggregations", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeNode", ",", "NodeName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processAggregationRequest", "(", "key", ",", "request", ",", "response", ")", "\n", "}" ]
// nodeAggregations returns a metric timeseries for a metric of the Node entity.
[ "nodeAggregations", "returns", "a", "metric", "timeseries", "for", "a", "metric", "of", "the", "Node", "entity", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L488-L494
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
namespaceAggregations
func (a *HistoricalApi) namespaceAggregations(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeNamespace, NamespaceName: request.PathParameter("namespace-name"), } a.processAggregationRequest(key, request, response) }
go
func (a *HistoricalApi) namespaceAggregations(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeNamespace, NamespaceName: request.PathParameter("namespace-name"), } a.processAggregationRequest(key, request, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "namespaceAggregations", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeNamespace", ",", "NamespaceName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processAggregationRequest", "(", "key", ",", "request", ",", "response", ")", "\n", "}" ]
// namespaceAggregations returns a metric timeseries for a metric of the Namespace entity.
[ "namespaceAggregations", "returns", "a", "metric", "timeseries", "for", "a", "metric", "of", "the", "Namespace", "entity", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L497-L503
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
podContainerAggregations
func (a *HistoricalApi) podContainerAggregations(request *restful.Request, response *restful.Response) { var key core.HistoricalKey if request.PathParameter("pod-id") != "" { key = core.HistoricalKey{ ObjectType: core.MetricSetTypePodContainer, PodId: request.PathParameter("pod-id"), ContainerName: request.PathParameter("container-name"), } } else { key = core.HistoricalKey{ ObjectType: core.MetricSetTypePodContainer, NamespaceName: request.PathParameter("namespace-name"), PodName: request.PathParameter("pod-name"), ContainerName: request.PathParameter("container-name"), } } a.processAggregationRequest(key, request, response) }
go
func (a *HistoricalApi) podContainerAggregations(request *restful.Request, response *restful.Response) { var key core.HistoricalKey if request.PathParameter("pod-id") != "" { key = core.HistoricalKey{ ObjectType: core.MetricSetTypePodContainer, PodId: request.PathParameter("pod-id"), ContainerName: request.PathParameter("container-name"), } } else { key = core.HistoricalKey{ ObjectType: core.MetricSetTypePodContainer, NamespaceName: request.PathParameter("namespace-name"), PodName: request.PathParameter("pod-name"), ContainerName: request.PathParameter("container-name"), } } a.processAggregationRequest(key, request, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "podContainerAggregations", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "var", "key", "core", ".", "HistoricalKey", "\n", "if", "request", ".", "PathParameter", "(", "\"", "\"", ")", "!=", "\"", "\"", "{", "key", "=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypePodContainer", ",", "PodId", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "ContainerName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "}", "else", "{", "key", "=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypePodContainer", ",", "NamespaceName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "PodName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "ContainerName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "}", "\n", "a", ".", "processAggregationRequest", "(", "key", ",", "request", ",", "response", ")", "\n", "}" ]
// podContainerAggregations returns a metric timeseries for a metric of a Pod Container entity.
[ "podContainerAggregations", "returns", "a", "metric", "timeseries", "for", "a", "metric", "of", "a", "Pod", "Container", "entity", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L524-L541
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
freeContainerAggregations
func (a *HistoricalApi) freeContainerAggregations(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeSystemContainer, NodeName: request.PathParameter("node-name"), ContainerName: request.PathParameter("container-name"), } a.processAggregationRequest(key, request, response) }
go
func (a *HistoricalApi) freeContainerAggregations(request *restful.Request, response *restful.Response) { key := core.HistoricalKey{ ObjectType: core.MetricSetTypeSystemContainer, NodeName: request.PathParameter("node-name"), ContainerName: request.PathParameter("container-name"), } a.processAggregationRequest(key, request, response) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "freeContainerAggregations", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypeSystemContainer", ",", "NodeName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "ContainerName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "}", "\n", "a", ".", "processAggregationRequest", "(", "key", ",", "request", ",", "response", ")", "\n", "}" ]
// freeContainerAggregations returns a metric timeseries for a metric of the Container entity. // freeContainerAggregations addresses only free containers.
[ "freeContainerAggregations", "returns", "a", "metric", "timeseries", "for", "a", "metric", "of", "the", "Container", "entity", ".", "freeContainerAggregations", "addresses", "only", "free", "containers", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L545-L552
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
podListAggregations
func (a *HistoricalApi) podListAggregations(request *restful.Request, response *restful.Response) { start, end, err := getStartEndTimeHistorical(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } bucketSize, err := getBucketSize(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } aggregations, err := getAggregations(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } labels, err := getLabels(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } keys := []core.HistoricalKey{} if request.PathParameter("pod-id-list") != "" { for _, podId := range strings.Split(request.PathParameter("pod-id-list"), ",") { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, PodId: podId, } keys = append(keys, key) } } else { for _, podName := range strings.Split(request.PathParameter("pod-list"), ",") { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, NamespaceName: request.PathParameter("namespace-name"), PodName: podName, } keys = append(keys, key) } } metricName := request.PathParameter("metric-name") convertedMetricName := convertMetricName(metricName) var metrics map[core.HistoricalKey][]core.TimestampedAggregationValue if labels != nil { metrics, err = a.historicalSource.GetLabeledAggregation(convertedMetricName, labels, aggregations, keys, start, end, bucketSize) } else { metrics, err = a.historicalSource.GetAggregation(convertedMetricName, aggregations, keys, start, end, bucketSize) } if err != nil { response.WriteError(http.StatusInternalServerError, err) return } result := types.MetricAggregationResultList{ Items: make([]types.MetricAggregationResult, 0, len(keys)), } for _, key := range keys { result.Items = append(result.Items, exportTimestampedAggregationValue(metrics[key])) } response.PrettyPrint(false) response.WriteEntity(result) }
go
func (a *HistoricalApi) podListAggregations(request *restful.Request, response *restful.Response) { start, end, err := getStartEndTimeHistorical(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } bucketSize, err := getBucketSize(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } aggregations, err := getAggregations(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } labels, err := getLabels(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } keys := []core.HistoricalKey{} if request.PathParameter("pod-id-list") != "" { for _, podId := range strings.Split(request.PathParameter("pod-id-list"), ",") { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, PodId: podId, } keys = append(keys, key) } } else { for _, podName := range strings.Split(request.PathParameter("pod-list"), ",") { key := core.HistoricalKey{ ObjectType: core.MetricSetTypePod, NamespaceName: request.PathParameter("namespace-name"), PodName: podName, } keys = append(keys, key) } } metricName := request.PathParameter("metric-name") convertedMetricName := convertMetricName(metricName) var metrics map[core.HistoricalKey][]core.TimestampedAggregationValue if labels != nil { metrics, err = a.historicalSource.GetLabeledAggregation(convertedMetricName, labels, aggregations, keys, start, end, bucketSize) } else { metrics, err = a.historicalSource.GetAggregation(convertedMetricName, aggregations, keys, start, end, bucketSize) } if err != nil { response.WriteError(http.StatusInternalServerError, err) return } result := types.MetricAggregationResultList{ Items: make([]types.MetricAggregationResult, 0, len(keys)), } for _, key := range keys { result.Items = append(result.Items, exportTimestampedAggregationValue(metrics[key])) } response.PrettyPrint(false) response.WriteEntity(result) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "podListAggregations", "(", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "start", ",", "end", ",", "err", ":=", "getStartEndTimeHistorical", "(", "request", ")", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusBadRequest", ",", "err", ")", "\n", "return", "\n", "}", "\n", "bucketSize", ",", "err", ":=", "getBucketSize", "(", "request", ")", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusBadRequest", ",", "err", ")", "\n", "return", "\n", "}", "\n", "aggregations", ",", "err", ":=", "getAggregations", "(", "request", ")", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusBadRequest", ",", "err", ")", "\n", "return", "\n", "}", "\n", "labels", ",", "err", ":=", "getLabels", "(", "request", ")", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusBadRequest", ",", "err", ")", "\n", "return", "\n", "}", "\n", "keys", ":=", "[", "]", "core", ".", "HistoricalKey", "{", "}", "\n", "if", "request", ".", "PathParameter", "(", "\"", "\"", ")", "!=", "\"", "\"", "{", "for", "_", ",", "podId", ":=", "range", "strings", ".", "Split", "(", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "\"", "\"", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypePod", ",", "PodId", ":", "podId", ",", "}", "\n", "keys", "=", "append", "(", "keys", ",", "key", ")", "\n", "}", "\n", "}", "else", "{", "for", "_", ",", "podName", ":=", "range", "strings", ".", "Split", "(", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "\"", "\"", ")", "{", "key", ":=", "core", ".", "HistoricalKey", "{", "ObjectType", ":", "core", ".", "MetricSetTypePod", ",", "NamespaceName", ":", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "PodName", ":", "podName", ",", "}", "\n", "keys", "=", "append", "(", "keys", ",", "key", ")", "\n", "}", "\n", "}", "\n", "metricName", ":=", "request", ".", "PathParameter", "(", "\"", "\"", ")", "\n", "convertedMetricName", ":=", "convertMetricName", "(", "metricName", ")", "\n", "var", "metrics", "map", "[", "core", ".", "HistoricalKey", "]", "[", "]", "core", ".", "TimestampedAggregationValue", "\n", "if", "labels", "!=", "nil", "{", "metrics", ",", "err", "=", "a", ".", "historicalSource", ".", "GetLabeledAggregation", "(", "convertedMetricName", ",", "labels", ",", "aggregations", ",", "keys", ",", "start", ",", "end", ",", "bucketSize", ")", "\n", "}", "else", "{", "metrics", ",", "err", "=", "a", ".", "historicalSource", ".", "GetAggregation", "(", "convertedMetricName", ",", "aggregations", ",", "keys", ",", "start", ",", "end", ",", "bucketSize", ")", "\n", "}", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusInternalServerError", ",", "err", ")", "\n", "return", "\n", "}", "\n\n", "result", ":=", "types", ".", "MetricAggregationResultList", "{", "Items", ":", "make", "(", "[", "]", "types", ".", "MetricAggregationResult", ",", "0", ",", "len", "(", "keys", ")", ")", ",", "}", "\n", "for", "_", ",", "key", ":=", "range", "keys", "{", "result", ".", "Items", "=", "append", "(", "result", ".", "Items", ",", "exportTimestampedAggregationValue", "(", "metrics", "[", "key", "]", ")", ")", "\n", "}", "\n", "response", ".", "PrettyPrint", "(", "false", ")", "\n", "response", ".", "WriteEntity", "(", "result", ")", "\n", "}" ]
// podListAggregations returns a list of metric timeseries for the specified pods.
[ "podListAggregations", "returns", "a", "list", "of", "metric", "timeseries", "for", "the", "specified", "pods", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L555-L616
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
processMetricRequest
func (a *HistoricalApi) processMetricRequest(key core.HistoricalKey, request *restful.Request, response *restful.Response) { start, end, err := getStartEndTimeHistorical(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } labels, err := getLabels(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } metricName := request.PathParameter("metric-name") convertedMetricName := convertMetricName(metricName) var metrics map[core.HistoricalKey][]core.TimestampedMetricValue if labels != nil { metrics, err = a.historicalSource.GetLabeledMetric(convertedMetricName, labels, []core.HistoricalKey{key}, start, end) } else { metrics, err = a.historicalSource.GetMetric(convertedMetricName, []core.HistoricalKey{key}, start, end) } if err != nil { response.WriteError(http.StatusInternalServerError, err) return } converted := exportTimestampedMetricValue(metrics[key]) response.WriteEntity(converted) }
go
func (a *HistoricalApi) processMetricRequest(key core.HistoricalKey, request *restful.Request, response *restful.Response) { start, end, err := getStartEndTimeHistorical(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } labels, err := getLabels(request) if err != nil { response.WriteError(http.StatusBadRequest, err) return } metricName := request.PathParameter("metric-name") convertedMetricName := convertMetricName(metricName) var metrics map[core.HistoricalKey][]core.TimestampedMetricValue if labels != nil { metrics, err = a.historicalSource.GetLabeledMetric(convertedMetricName, labels, []core.HistoricalKey{key}, start, end) } else { metrics, err = a.historicalSource.GetMetric(convertedMetricName, []core.HistoricalKey{key}, start, end) } if err != nil { response.WriteError(http.StatusInternalServerError, err) return } converted := exportTimestampedMetricValue(metrics[key]) response.WriteEntity(converted) }
[ "func", "(", "a", "*", "HistoricalApi", ")", "processMetricRequest", "(", "key", "core", ".", "HistoricalKey", ",", "request", "*", "restful", ".", "Request", ",", "response", "*", "restful", ".", "Response", ")", "{", "start", ",", "end", ",", "err", ":=", "getStartEndTimeHistorical", "(", "request", ")", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusBadRequest", ",", "err", ")", "\n", "return", "\n", "}", "\n", "labels", ",", "err", ":=", "getLabels", "(", "request", ")", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusBadRequest", ",", "err", ")", "\n", "return", "\n", "}", "\n", "metricName", ":=", "request", ".", "PathParameter", "(", "\"", "\"", ")", "\n", "convertedMetricName", ":=", "convertMetricName", "(", "metricName", ")", "\n\n", "var", "metrics", "map", "[", "core", ".", "HistoricalKey", "]", "[", "]", "core", ".", "TimestampedMetricValue", "\n", "if", "labels", "!=", "nil", "{", "metrics", ",", "err", "=", "a", ".", "historicalSource", ".", "GetLabeledMetric", "(", "convertedMetricName", ",", "labels", ",", "[", "]", "core", ".", "HistoricalKey", "{", "key", "}", ",", "start", ",", "end", ")", "\n", "}", "else", "{", "metrics", ",", "err", "=", "a", ".", "historicalSource", ".", "GetMetric", "(", "convertedMetricName", ",", "[", "]", "core", ".", "HistoricalKey", "{", "key", "}", ",", "start", ",", "end", ")", "\n", "}", "\n", "if", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusInternalServerError", ",", "err", ")", "\n", "return", "\n", "}", "\n\n", "converted", ":=", "exportTimestampedMetricValue", "(", "metrics", "[", "key", "]", ")", "\n", "response", ".", "WriteEntity", "(", "converted", ")", "\n", "}" ]
// processMetricRequest retrieves a metric for the object at the requested key.
[ "processMetricRequest", "retrieves", "a", "metric", "for", "the", "object", "at", "the", "requested", "key", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L619-L646
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
processMetricNamesRequest
func (a *HistoricalApi) processMetricNamesRequest(key core.HistoricalKey, response *restful.Response) { if resp, err := a.historicalSource.GetMetricNames(key); err != nil { response.WriteError(http.StatusInternalServerError, err) } else { response.WriteEntity(resp) } }
go
func (a *HistoricalApi) processMetricNamesRequest(key core.HistoricalKey, response *restful.Response) { if resp, err := a.historicalSource.GetMetricNames(key); err != nil { response.WriteError(http.StatusInternalServerError, err) } else { response.WriteEntity(resp) } }
[ "func", "(", "a", "*", "HistoricalApi", ")", "processMetricNamesRequest", "(", "key", "core", ".", "HistoricalKey", ",", "response", "*", "restful", ".", "Response", ")", "{", "if", "resp", ",", "err", ":=", "a", ".", "historicalSource", ".", "GetMetricNames", "(", "key", ")", ";", "err", "!=", "nil", "{", "response", ".", "WriteError", "(", "http", ".", "StatusInternalServerError", ",", "err", ")", "\n", "}", "else", "{", "response", ".", "WriteEntity", "(", "resp", ")", "\n", "}", "\n", "}" ]
// processMetricNamesRequest retrieves the available metrics for the object at the specified key.
[ "processMetricNamesRequest", "retrieves", "the", "available", "metrics", "for", "the", "object", "at", "the", "specified", "key", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L649-L655
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
getBucketSize
func getBucketSize(request *restful.Request) (time.Duration, error) { rawSize := request.QueryParameter("bucket") if rawSize == "" { return 0, nil } if len(rawSize) < 2 { return 0, fmt.Errorf("unable to parse bucket size: %q is too short to be a duration", rawSize) } var multiplier time.Duration var num string switch rawSize[len(rawSize)-1] { case 's': // could be s or ms if len(rawSize) < 3 || rawSize[len(rawSize)-2] != 'm' { multiplier = time.Second num = rawSize[:len(rawSize)-1] } else { multiplier = time.Millisecond num = rawSize[:len(rawSize)-2] } case 'h': multiplier = time.Hour num = rawSize[:len(rawSize)-1] case 'd': multiplier = 24 * time.Hour num = rawSize[:len(rawSize)-1] case 'm': multiplier = time.Minute num = rawSize[:len(rawSize)-1] default: return 0, fmt.Errorf("unable to parse bucket size: %q has no known duration suffix", rawSize) } parsedNum, err := strconv.ParseUint(num, 10, 64) if err != nil { return 0, err } return time.Duration(parsedNum) * multiplier, nil }
go
func getBucketSize(request *restful.Request) (time.Duration, error) { rawSize := request.QueryParameter("bucket") if rawSize == "" { return 0, nil } if len(rawSize) < 2 { return 0, fmt.Errorf("unable to parse bucket size: %q is too short to be a duration", rawSize) } var multiplier time.Duration var num string switch rawSize[len(rawSize)-1] { case 's': // could be s or ms if len(rawSize) < 3 || rawSize[len(rawSize)-2] != 'm' { multiplier = time.Second num = rawSize[:len(rawSize)-1] } else { multiplier = time.Millisecond num = rawSize[:len(rawSize)-2] } case 'h': multiplier = time.Hour num = rawSize[:len(rawSize)-1] case 'd': multiplier = 24 * time.Hour num = rawSize[:len(rawSize)-1] case 'm': multiplier = time.Minute num = rawSize[:len(rawSize)-1] default: return 0, fmt.Errorf("unable to parse bucket size: %q has no known duration suffix", rawSize) } parsedNum, err := strconv.ParseUint(num, 10, 64) if err != nil { return 0, err } return time.Duration(parsedNum) * multiplier, nil }
[ "func", "getBucketSize", "(", "request", "*", "restful", ".", "Request", ")", "(", "time", ".", "Duration", ",", "error", ")", "{", "rawSize", ":=", "request", ".", "QueryParameter", "(", "\"", "\"", ")", "\n", "if", "rawSize", "==", "\"", "\"", "{", "return", "0", ",", "nil", "\n", "}", "\n\n", "if", "len", "(", "rawSize", ")", "<", "2", "{", "return", "0", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "rawSize", ")", "\n", "}", "\n", "var", "multiplier", "time", ".", "Duration", "\n", "var", "num", "string", "\n\n", "switch", "rawSize", "[", "len", "(", "rawSize", ")", "-", "1", "]", "{", "case", "'s'", ":", "// could be s or ms", "if", "len", "(", "rawSize", ")", "<", "3", "||", "rawSize", "[", "len", "(", "rawSize", ")", "-", "2", "]", "!=", "'m'", "{", "multiplier", "=", "time", ".", "Second", "\n", "num", "=", "rawSize", "[", ":", "len", "(", "rawSize", ")", "-", "1", "]", "\n", "}", "else", "{", "multiplier", "=", "time", ".", "Millisecond", "\n", "num", "=", "rawSize", "[", ":", "len", "(", "rawSize", ")", "-", "2", "]", "\n", "}", "\n", "case", "'h'", ":", "multiplier", "=", "time", ".", "Hour", "\n", "num", "=", "rawSize", "[", ":", "len", "(", "rawSize", ")", "-", "1", "]", "\n", "case", "'d'", ":", "multiplier", "=", "24", "*", "time", ".", "Hour", "\n", "num", "=", "rawSize", "[", ":", "len", "(", "rawSize", ")", "-", "1", "]", "\n", "case", "'m'", ":", "multiplier", "=", "time", ".", "Minute", "\n", "num", "=", "rawSize", "[", ":", "len", "(", "rawSize", ")", "-", "1", "]", "\n", "default", ":", "return", "0", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "rawSize", ")", "\n", "}", "\n\n", "parsedNum", ",", "err", ":=", "strconv", ".", "ParseUint", "(", "num", ",", "10", ",", "64", ")", "\n", "if", "err", "!=", "nil", "{", "return", "0", ",", "err", "\n", "}", "\n\n", "return", "time", ".", "Duration", "(", "parsedNum", ")", "*", "multiplier", ",", "nil", "\n", "}" ]
// getBucketSize parses the bucket size specifier into a
[ "getBucketSize", "parses", "the", "bucket", "size", "specifier", "into", "a" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L698-L739
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
getAggregations
func getAggregations(request *restful.Request) ([]core.AggregationType, error) { aggregationsRaw := strings.Split(request.PathParameter("aggregations"), ",") if len(aggregationsRaw) == 0 { return nil, fmt.Errorf("No aggregations specified") } aggregations := make([]core.AggregationType, len(aggregationsRaw)) for ind, aggNameRaw := range aggregationsRaw { aggName := core.AggregationType(aggNameRaw) if _, ok := core.AllAggregations[aggName]; !ok { return nil, fmt.Errorf("Unknown aggregation %q", aggName) } aggregations[ind] = aggName } return aggregations, nil }
go
func getAggregations(request *restful.Request) ([]core.AggregationType, error) { aggregationsRaw := strings.Split(request.PathParameter("aggregations"), ",") if len(aggregationsRaw) == 0 { return nil, fmt.Errorf("No aggregations specified") } aggregations := make([]core.AggregationType, len(aggregationsRaw)) for ind, aggNameRaw := range aggregationsRaw { aggName := core.AggregationType(aggNameRaw) if _, ok := core.AllAggregations[aggName]; !ok { return nil, fmt.Errorf("Unknown aggregation %q", aggName) } aggregations[ind] = aggName } return aggregations, nil }
[ "func", "getAggregations", "(", "request", "*", "restful", ".", "Request", ")", "(", "[", "]", "core", ".", "AggregationType", ",", "error", ")", "{", "aggregationsRaw", ":=", "strings", ".", "Split", "(", "request", ".", "PathParameter", "(", "\"", "\"", ")", ",", "\"", "\"", ")", "\n", "if", "len", "(", "aggregationsRaw", ")", "==", "0", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ")", "\n", "}", "\n\n", "aggregations", ":=", "make", "(", "[", "]", "core", ".", "AggregationType", ",", "len", "(", "aggregationsRaw", ")", ")", "\n\n", "for", "ind", ",", "aggNameRaw", ":=", "range", "aggregationsRaw", "{", "aggName", ":=", "core", ".", "AggregationType", "(", "aggNameRaw", ")", "\n", "if", "_", ",", "ok", ":=", "core", ".", "AllAggregations", "[", "aggName", "]", ";", "!", "ok", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "aggName", ")", "\n", "}", "\n", "aggregations", "[", "ind", "]", "=", "aggName", "\n", "}", "\n\n", "return", "aggregations", ",", "nil", "\n", "}" ]
// getAggregations extracts and validates the list of requested aggregations
[ "getAggregations", "extracts", "and", "validates", "the", "list", "of", "requested", "aggregations" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L742-L759
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
exportMetricValue
func exportMetricValue(value *core.MetricValue) *types.MetricValue { if value == nil { return nil } if value.ValueType == core.ValueInt64 { return &types.MetricValue{ IntValue: &value.IntValue, } } else { floatVal := float64(value.FloatValue) return &types.MetricValue{ FloatValue: &floatVal, } } }
go
func exportMetricValue(value *core.MetricValue) *types.MetricValue { if value == nil { return nil } if value.ValueType == core.ValueInt64 { return &types.MetricValue{ IntValue: &value.IntValue, } } else { floatVal := float64(value.FloatValue) return &types.MetricValue{ FloatValue: &floatVal, } } }
[ "func", "exportMetricValue", "(", "value", "*", "core", ".", "MetricValue", ")", "*", "types", ".", "MetricValue", "{", "if", "value", "==", "nil", "{", "return", "nil", "\n", "}", "\n\n", "if", "value", ".", "ValueType", "==", "core", ".", "ValueInt64", "{", "return", "&", "types", ".", "MetricValue", "{", "IntValue", ":", "&", "value", ".", "IntValue", ",", "}", "\n", "}", "else", "{", "floatVal", ":=", "float64", "(", "value", ".", "FloatValue", ")", "\n", "return", "&", "types", ".", "MetricValue", "{", "FloatValue", ":", "&", "floatVal", ",", "}", "\n", "}", "\n", "}" ]
// exportMetricValue converts a core.MetricValue into an API MetricValue
[ "exportMetricValue", "converts", "a", "core", ".", "MetricValue", "into", "an", "API", "MetricValue" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L762-L777
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
extractMetricValue
func extractMetricValue(aggregations *core.AggregationValue, aggName core.AggregationType) *types.MetricValue { if inputVal, ok := aggregations.Aggregations[aggName]; ok { return exportMetricValue(&inputVal) } else { return nil } }
go
func extractMetricValue(aggregations *core.AggregationValue, aggName core.AggregationType) *types.MetricValue { if inputVal, ok := aggregations.Aggregations[aggName]; ok { return exportMetricValue(&inputVal) } else { return nil } }
[ "func", "extractMetricValue", "(", "aggregations", "*", "core", ".", "AggregationValue", ",", "aggName", "core", ".", "AggregationType", ")", "*", "types", ".", "MetricValue", "{", "if", "inputVal", ",", "ok", ":=", "aggregations", ".", "Aggregations", "[", "aggName", "]", ";", "ok", "{", "return", "exportMetricValue", "(", "&", "inputVal", ")", "\n", "}", "else", "{", "return", "nil", "\n", "}", "\n", "}" ]
// extractMetricValue checks to see if the given metric was present in the results, and if so, // returns it in API form
[ "extractMetricValue", "checks", "to", "see", "if", "the", "given", "metric", "was", "present", "in", "the", "results", "and", "if", "so", "returns", "it", "in", "API", "form" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L781-L787
train
kubernetes-retired/heapster
metrics/api/v1/historical_handlers.go
exportTimestampedAggregationValue
func exportTimestampedAggregationValue(values []core.TimestampedAggregationValue) types.MetricAggregationResult { result := types.MetricAggregationResult{ Buckets: make([]types.MetricAggregationBucket, 0, len(values)), BucketSize: 0, } for _, value := range values { // just use the largest bucket size, since all bucket sizes should be uniform // (except for the last one, which may be smaller) if result.BucketSize < value.BucketSize { result.BucketSize = value.BucketSize } bucket := types.MetricAggregationBucket{ Timestamp: value.Timestamp, Count: value.Count, Average: extractMetricValue(&value.AggregationValue, core.AggregationTypeAverage), Maximum: extractMetricValue(&value.AggregationValue, core.AggregationTypeMaximum), Minimum: extractMetricValue(&value.AggregationValue, core.AggregationTypeMinimum), Median: extractMetricValue(&value.AggregationValue, core.AggregationTypeMedian), Percentiles: make(map[string]types.MetricValue, 3), } if val, ok := value.Aggregations[core.AggregationTypePercentile50]; ok { bucket.Percentiles["50"] = *exportMetricValue(&val) } if val, ok := value.Aggregations[core.AggregationTypePercentile95]; ok { bucket.Percentiles["95"] = *exportMetricValue(&val) } if val, ok := value.Aggregations[core.AggregationTypePercentile99]; ok { bucket.Percentiles["99"] = *exportMetricValue(&val) } result.Buckets = append(result.Buckets, bucket) } return result }
go
func exportTimestampedAggregationValue(values []core.TimestampedAggregationValue) types.MetricAggregationResult { result := types.MetricAggregationResult{ Buckets: make([]types.MetricAggregationBucket, 0, len(values)), BucketSize: 0, } for _, value := range values { // just use the largest bucket size, since all bucket sizes should be uniform // (except for the last one, which may be smaller) if result.BucketSize < value.BucketSize { result.BucketSize = value.BucketSize } bucket := types.MetricAggregationBucket{ Timestamp: value.Timestamp, Count: value.Count, Average: extractMetricValue(&value.AggregationValue, core.AggregationTypeAverage), Maximum: extractMetricValue(&value.AggregationValue, core.AggregationTypeMaximum), Minimum: extractMetricValue(&value.AggregationValue, core.AggregationTypeMinimum), Median: extractMetricValue(&value.AggregationValue, core.AggregationTypeMedian), Percentiles: make(map[string]types.MetricValue, 3), } if val, ok := value.Aggregations[core.AggregationTypePercentile50]; ok { bucket.Percentiles["50"] = *exportMetricValue(&val) } if val, ok := value.Aggregations[core.AggregationTypePercentile95]; ok { bucket.Percentiles["95"] = *exportMetricValue(&val) } if val, ok := value.Aggregations[core.AggregationTypePercentile99]; ok { bucket.Percentiles["99"] = *exportMetricValue(&val) } result.Buckets = append(result.Buckets, bucket) } return result }
[ "func", "exportTimestampedAggregationValue", "(", "values", "[", "]", "core", ".", "TimestampedAggregationValue", ")", "types", ".", "MetricAggregationResult", "{", "result", ":=", "types", ".", "MetricAggregationResult", "{", "Buckets", ":", "make", "(", "[", "]", "types", ".", "MetricAggregationBucket", ",", "0", ",", "len", "(", "values", ")", ")", ",", "BucketSize", ":", "0", ",", "}", "\n", "for", "_", ",", "value", ":=", "range", "values", "{", "// just use the largest bucket size, since all bucket sizes should be uniform", "// (except for the last one, which may be smaller)", "if", "result", ".", "BucketSize", "<", "value", ".", "BucketSize", "{", "result", ".", "BucketSize", "=", "value", ".", "BucketSize", "\n", "}", "\n\n", "bucket", ":=", "types", ".", "MetricAggregationBucket", "{", "Timestamp", ":", "value", ".", "Timestamp", ",", "Count", ":", "value", ".", "Count", ",", "Average", ":", "extractMetricValue", "(", "&", "value", ".", "AggregationValue", ",", "core", ".", "AggregationTypeAverage", ")", ",", "Maximum", ":", "extractMetricValue", "(", "&", "value", ".", "AggregationValue", ",", "core", ".", "AggregationTypeMaximum", ")", ",", "Minimum", ":", "extractMetricValue", "(", "&", "value", ".", "AggregationValue", ",", "core", ".", "AggregationTypeMinimum", ")", ",", "Median", ":", "extractMetricValue", "(", "&", "value", ".", "AggregationValue", ",", "core", ".", "AggregationTypeMedian", ")", ",", "Percentiles", ":", "make", "(", "map", "[", "string", "]", "types", ".", "MetricValue", ",", "3", ")", ",", "}", "\n\n", "if", "val", ",", "ok", ":=", "value", ".", "Aggregations", "[", "core", ".", "AggregationTypePercentile50", "]", ";", "ok", "{", "bucket", ".", "Percentiles", "[", "\"", "\"", "]", "=", "*", "exportMetricValue", "(", "&", "val", ")", "\n", "}", "\n", "if", "val", ",", "ok", ":=", "value", ".", "Aggregations", "[", "core", ".", "AggregationTypePercentile95", "]", ";", "ok", "{", "bucket", ".", "Percentiles", "[", "\"", "\"", "]", "=", "*", "exportMetricValue", "(", "&", "val", ")", "\n", "}", "\n", "if", "val", ",", "ok", ":=", "value", ".", "Aggregations", "[", "core", ".", "AggregationTypePercentile99", "]", ";", "ok", "{", "bucket", ".", "Percentiles", "[", "\"", "\"", "]", "=", "*", "exportMetricValue", "(", "&", "val", ")", "\n", "}", "\n\n", "result", ".", "Buckets", "=", "append", "(", "result", ".", "Buckets", ",", "bucket", ")", "\n", "}", "\n", "return", "result", "\n", "}" ]
// exportTimestampedAggregationValue converts a core.TimestampedAggregationValue into an API MetricAggregationResult
[ "exportTimestampedAggregationValue", "converts", "a", "core", ".", "TimestampedAggregationValue", "into", "an", "API", "MetricAggregationResult" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/api/v1/historical_handlers.go#L790-L828
train
kubernetes-retired/heapster
metrics/sources/summary/summary.go
decodeSummary
func (this *summaryMetricsSource) decodeSummary(summary *stats.Summary) map[string]*MetricSet { glog.V(9).Infof("Begin summary decode") result := map[string]*MetricSet{} labels := map[string]string{ LabelNodename.Key: this.node.NodeName, LabelHostname.Key: this.node.HostName, LabelHostID.Key: this.node.HostID, } this.decodeNodeStats(result, labels, &summary.Node) for _, pod := range summary.Pods { this.decodePodStats(result, labels, &pod) } glog.V(9).Infof("End summary decode") return result }
go
func (this *summaryMetricsSource) decodeSummary(summary *stats.Summary) map[string]*MetricSet { glog.V(9).Infof("Begin summary decode") result := map[string]*MetricSet{} labels := map[string]string{ LabelNodename.Key: this.node.NodeName, LabelHostname.Key: this.node.HostName, LabelHostID.Key: this.node.HostID, } this.decodeNodeStats(result, labels, &summary.Node) for _, pod := range summary.Pods { this.decodePodStats(result, labels, &pod) } glog.V(9).Infof("End summary decode") return result }
[ "func", "(", "this", "*", "summaryMetricsSource", ")", "decodeSummary", "(", "summary", "*", "stats", ".", "Summary", ")", "map", "[", "string", "]", "*", "MetricSet", "{", "glog", ".", "V", "(", "9", ")", ".", "Infof", "(", "\"", "\"", ")", "\n", "result", ":=", "map", "[", "string", "]", "*", "MetricSet", "{", "}", "\n\n", "labels", ":=", "map", "[", "string", "]", "string", "{", "LabelNodename", ".", "Key", ":", "this", ".", "node", ".", "NodeName", ",", "LabelHostname", ".", "Key", ":", "this", ".", "node", ".", "HostName", ",", "LabelHostID", ".", "Key", ":", "this", ".", "node", ".", "HostID", ",", "}", "\n\n", "this", ".", "decodeNodeStats", "(", "result", ",", "labels", ",", "&", "summary", ".", "Node", ")", "\n", "for", "_", ",", "pod", ":=", "range", "summary", ".", "Pods", "{", "this", ".", "decodePodStats", "(", "result", ",", "labels", ",", "&", "pod", ")", "\n", "}", "\n\n", "glog", ".", "V", "(", "9", ")", ".", "Infof", "(", "\"", "\"", ")", "\n", "return", "result", "\n", "}" ]
// decodeSummary translates the kubelet statsSummary API into the flattened heapster MetricSet API.
[ "decodeSummary", "translates", "the", "kubelet", "statsSummary", "API", "into", "the", "flattened", "heapster", "MetricSet", "API", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L120-L137
train
kubernetes-retired/heapster
metrics/sources/summary/summary.go
cloneLabels
func (this *summaryMetricsSource) cloneLabels(labels map[string]string) map[string]string { clone := make(map[string]string, len(labels)) for k, v := range labels { clone[k] = v } return clone }
go
func (this *summaryMetricsSource) cloneLabels(labels map[string]string) map[string]string { clone := make(map[string]string, len(labels)) for k, v := range labels { clone[k] = v } return clone }
[ "func", "(", "this", "*", "summaryMetricsSource", ")", "cloneLabels", "(", "labels", "map", "[", "string", "]", "string", ")", "map", "[", "string", "]", "string", "{", "clone", ":=", "make", "(", "map", "[", "string", "]", "string", ",", "len", "(", "labels", ")", ")", "\n", "for", "k", ",", "v", ":=", "range", "labels", "{", "clone", "[", "k", "]", "=", "v", "\n", "}", "\n", "return", "clone", "\n", "}" ]
// Convenience method for labels deep copy.
[ "Convenience", "method", "for", "labels", "deep", "copy", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L140-L146
train
kubernetes-retired/heapster
metrics/sources/summary/summary.go
addIntMetric
func (this *summaryMetricsSource) addIntMetric(metrics *MetricSet, metric *Metric, value *uint64) { if value == nil { glog.V(9).Infof("skipping metric %s because the value was nil", metric.Name) return } val := MetricValue{ ValueType: ValueInt64, MetricType: metric.Type, IntValue: int64(*value), } metrics.MetricValues[metric.Name] = val }
go
func (this *summaryMetricsSource) addIntMetric(metrics *MetricSet, metric *Metric, value *uint64) { if value == nil { glog.V(9).Infof("skipping metric %s because the value was nil", metric.Name) return } val := MetricValue{ ValueType: ValueInt64, MetricType: metric.Type, IntValue: int64(*value), } metrics.MetricValues[metric.Name] = val }
[ "func", "(", "this", "*", "summaryMetricsSource", ")", "addIntMetric", "(", "metrics", "*", "MetricSet", ",", "metric", "*", "Metric", ",", "value", "*", "uint64", ")", "{", "if", "value", "==", "nil", "{", "glog", ".", "V", "(", "9", ")", ".", "Infof", "(", "\"", "\"", ",", "metric", ".", "Name", ")", "\n", "return", "\n", "}", "\n", "val", ":=", "MetricValue", "{", "ValueType", ":", "ValueInt64", ",", "MetricType", ":", "metric", ".", "Type", ",", "IntValue", ":", "int64", "(", "*", "value", ")", ",", "}", "\n", "metrics", ".", "MetricValues", "[", "metric", ".", "Name", "]", "=", "val", "\n", "}" ]
// addIntMetric is a convenience method for adding the metric and value to the metric set.
[ "addIntMetric", "is", "a", "convenience", "method", "for", "adding", "the", "metric", "and", "value", "to", "the", "metric", "set", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L370-L381
train
kubernetes-retired/heapster
metrics/sources/summary/summary.go
addLabeledIntMetric
func (this *summaryMetricsSource) addLabeledIntMetric(metrics *MetricSet, metric *Metric, labels map[string]string, value *uint64) { if value == nil { glog.V(9).Infof("skipping labeled metric %s (%v) because the value was nil", metric.Name, labels) return } val := LabeledMetric{ Name: metric.Name, Labels: labels, MetricValue: MetricValue{ ValueType: ValueInt64, MetricType: metric.Type, IntValue: int64(*value), }, } metrics.LabeledMetrics = append(metrics.LabeledMetrics, val) }
go
func (this *summaryMetricsSource) addLabeledIntMetric(metrics *MetricSet, metric *Metric, labels map[string]string, value *uint64) { if value == nil { glog.V(9).Infof("skipping labeled metric %s (%v) because the value was nil", metric.Name, labels) return } val := LabeledMetric{ Name: metric.Name, Labels: labels, MetricValue: MetricValue{ ValueType: ValueInt64, MetricType: metric.Type, IntValue: int64(*value), }, } metrics.LabeledMetrics = append(metrics.LabeledMetrics, val) }
[ "func", "(", "this", "*", "summaryMetricsSource", ")", "addLabeledIntMetric", "(", "metrics", "*", "MetricSet", ",", "metric", "*", "Metric", ",", "labels", "map", "[", "string", "]", "string", ",", "value", "*", "uint64", ")", "{", "if", "value", "==", "nil", "{", "glog", ".", "V", "(", "9", ")", ".", "Infof", "(", "\"", "\"", ",", "metric", ".", "Name", ",", "labels", ")", "\n", "return", "\n", "}", "\n\n", "val", ":=", "LabeledMetric", "{", "Name", ":", "metric", ".", "Name", ",", "Labels", ":", "labels", ",", "MetricValue", ":", "MetricValue", "{", "ValueType", ":", "ValueInt64", ",", "MetricType", ":", "metric", ".", "Type", ",", "IntValue", ":", "int64", "(", "*", "value", ")", ",", "}", ",", "}", "\n", "metrics", ".", "LabeledMetrics", "=", "append", "(", "metrics", ".", "LabeledMetrics", ",", "val", ")", "\n", "}" ]
// addLabeledIntMetric is a convenience method for adding the labeled metric and value to the metric set.
[ "addLabeledIntMetric", "is", "a", "convenience", "method", "for", "adding", "the", "labeled", "metric", "and", "value", "to", "the", "metric", "set", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L384-L400
train
kubernetes-retired/heapster
metrics/sources/summary/summary.go
getSystemContainerName
func (this *summaryMetricsSource) getSystemContainerName(c *stats.ContainerStats) string { if legacyName, ok := systemNameMap[c.Name]; ok { return legacyName } return c.Name }
go
func (this *summaryMetricsSource) getSystemContainerName(c *stats.ContainerStats) string { if legacyName, ok := systemNameMap[c.Name]; ok { return legacyName } return c.Name }
[ "func", "(", "this", "*", "summaryMetricsSource", ")", "getSystemContainerName", "(", "c", "*", "stats", ".", "ContainerStats", ")", "string", "{", "if", "legacyName", ",", "ok", ":=", "systemNameMap", "[", "c", ".", "Name", "]", ";", "ok", "{", "return", "legacyName", "\n", "}", "\n", "return", "c", ".", "Name", "\n", "}" ]
// Translate system container names to the legacy names for backwards compatibility.
[ "Translate", "system", "container", "names", "to", "the", "legacy", "names", "for", "backwards", "compatibility", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sources/summary/summary.go#L403-L408
train
kubernetes-retired/heapster
common/riemann/riemann.go
GetRiemannClient
func GetRiemannClient(config RiemannConfig) (riemanngo.Client, error) { glog.Infof("Connect Riemann client...") client := riemanngo.NewTcpClient(config.Host) runtime.SetFinalizer(client, func(c riemanngo.Client) { c.Close() }) // 5 seconds timeout err := client.Connect(5) if err != nil { return nil, err } return client, nil }
go
func GetRiemannClient(config RiemannConfig) (riemanngo.Client, error) { glog.Infof("Connect Riemann client...") client := riemanngo.NewTcpClient(config.Host) runtime.SetFinalizer(client, func(c riemanngo.Client) { c.Close() }) // 5 seconds timeout err := client.Connect(5) if err != nil { return nil, err } return client, nil }
[ "func", "GetRiemannClient", "(", "config", "RiemannConfig", ")", "(", "riemanngo", ".", "Client", ",", "error", ")", "{", "glog", ".", "Infof", "(", "\"", "\"", ")", "\n", "client", ":=", "riemanngo", ".", "NewTcpClient", "(", "config", ".", "Host", ")", "\n", "runtime", ".", "SetFinalizer", "(", "client", ",", "func", "(", "c", "riemanngo", ".", "Client", ")", "{", "c", ".", "Close", "(", ")", "}", ")", "\n", "// 5 seconds timeout", "err", ":=", "client", ".", "Connect", "(", "5", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "return", "client", ",", "nil", "\n", "}" ]
// Receives a sink, connect the riemann client.
[ "Receives", "a", "sink", "connect", "the", "riemann", "client", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/common/riemann/riemann.go#L102-L112
train
kubernetes-retired/heapster
common/riemann/riemann.go
SendData
func SendData(client riemanngo.Client, events []riemanngo.Event) error { // do nothing if we are not connected if client == nil { glog.Warningf("Riemann sink not connected") return nil } start := time.Now() _, err := riemanngo.SendEvents(client, &events) end := time.Now() if err == nil { glog.V(4).Infof("Exported %d events to riemann in %s", len(events), end.Sub(start)) return nil } else { glog.Warningf("There were errors sending events to Riemman, forcing reconnection. Error : %+v", err) return err } }
go
func SendData(client riemanngo.Client, events []riemanngo.Event) error { // do nothing if we are not connected if client == nil { glog.Warningf("Riemann sink not connected") return nil } start := time.Now() _, err := riemanngo.SendEvents(client, &events) end := time.Now() if err == nil { glog.V(4).Infof("Exported %d events to riemann in %s", len(events), end.Sub(start)) return nil } else { glog.Warningf("There were errors sending events to Riemman, forcing reconnection. Error : %+v", err) return err } }
[ "func", "SendData", "(", "client", "riemanngo", ".", "Client", ",", "events", "[", "]", "riemanngo", ".", "Event", ")", "error", "{", "// do nothing if we are not connected", "if", "client", "==", "nil", "{", "glog", ".", "Warningf", "(", "\"", "\"", ")", "\n", "return", "nil", "\n", "}", "\n", "start", ":=", "time", ".", "Now", "(", ")", "\n", "_", ",", "err", ":=", "riemanngo", ".", "SendEvents", "(", "client", ",", "&", "events", ")", "\n", "end", ":=", "time", ".", "Now", "(", ")", "\n", "if", "err", "==", "nil", "{", "glog", ".", "V", "(", "4", ")", ".", "Infof", "(", "\"", "\"", ",", "len", "(", "events", ")", ",", "end", ".", "Sub", "(", "start", ")", ")", "\n", "return", "nil", "\n", "}", "else", "{", "glog", ".", "Warningf", "(", "\"", "\"", ",", "err", ")", "\n", "return", "err", "\n", "}", "\n", "}" ]
// Send Events to Riemann using the client from the sink.
[ "Send", "Events", "to", "Riemann", "using", "the", "client", "from", "the", "sink", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/common/riemann/riemann.go#L115-L131
train
kubernetes-retired/heapster
events/sinks/manager.go
ExportEvents
func (this *sinkManager) ExportEvents(data *core.EventBatch) { var wg sync.WaitGroup for _, sh := range this.sinkHolders { wg.Add(1) go func(sh sinkHolder, wg *sync.WaitGroup) { defer wg.Done() glog.V(2).Infof("Pushing events to: %s", sh.sink.Name()) select { case sh.eventBatchChannel <- data: glog.V(2).Infof("Data events completed: %s", sh.sink.Name()) // everything ok case <-time.After(this.exportEventsTimeout): glog.Warningf("Failed to events data to sink: %s", sh.sink.Name()) } }(sh, &wg) } // Wait for all pushes to complete or timeout. wg.Wait() }
go
func (this *sinkManager) ExportEvents(data *core.EventBatch) { var wg sync.WaitGroup for _, sh := range this.sinkHolders { wg.Add(1) go func(sh sinkHolder, wg *sync.WaitGroup) { defer wg.Done() glog.V(2).Infof("Pushing events to: %s", sh.sink.Name()) select { case sh.eventBatchChannel <- data: glog.V(2).Infof("Data events completed: %s", sh.sink.Name()) // everything ok case <-time.After(this.exportEventsTimeout): glog.Warningf("Failed to events data to sink: %s", sh.sink.Name()) } }(sh, &wg) } // Wait for all pushes to complete or timeout. wg.Wait() }
[ "func", "(", "this", "*", "sinkManager", ")", "ExportEvents", "(", "data", "*", "core", ".", "EventBatch", ")", "{", "var", "wg", "sync", ".", "WaitGroup", "\n", "for", "_", ",", "sh", ":=", "range", "this", ".", "sinkHolders", "{", "wg", ".", "Add", "(", "1", ")", "\n", "go", "func", "(", "sh", "sinkHolder", ",", "wg", "*", "sync", ".", "WaitGroup", ")", "{", "defer", "wg", ".", "Done", "(", ")", "\n", "glog", ".", "V", "(", "2", ")", ".", "Infof", "(", "\"", "\"", ",", "sh", ".", "sink", ".", "Name", "(", ")", ")", "\n", "select", "{", "case", "sh", ".", "eventBatchChannel", "<-", "data", ":", "glog", ".", "V", "(", "2", ")", ".", "Infof", "(", "\"", "\"", ",", "sh", ".", "sink", ".", "Name", "(", ")", ")", "\n", "// everything ok", "case", "<-", "time", ".", "After", "(", "this", ".", "exportEventsTimeout", ")", ":", "glog", ".", "Warningf", "(", "\"", "\"", ",", "sh", ".", "sink", ".", "Name", "(", ")", ")", "\n", "}", "\n", "}", "(", "sh", ",", "&", "wg", ")", "\n", "}", "\n", "// Wait for all pushes to complete or timeout.", "wg", ".", "Wait", "(", ")", "\n", "}" ]
// Guarantees that the export will complete in exportEventsTimeout.
[ "Guarantees", "that", "the", "export", "will", "complete", "in", "exportEventsTimeout", "." ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/events/sinks/manager.go#L96-L114
train
kubernetes-retired/heapster
metrics/sinks/hawkular/client.go
cache
func (h *hawkularSink) cache(md *metrics.MetricDefinition) { h.pushToCache(md.ID, hashDefinition(md)) }
go
func (h *hawkularSink) cache(md *metrics.MetricDefinition) { h.pushToCache(md.ID, hashDefinition(md)) }
[ "func", "(", "h", "*", "hawkularSink", ")", "cache", "(", "md", "*", "metrics", ".", "MetricDefinition", ")", "{", "h", ".", "pushToCache", "(", "md", ".", "ID", ",", "hashDefinition", "(", "md", ")", ")", "\n", "}" ]
// cache inserts the item to the cache
[ "cache", "inserts", "the", "item", "to", "the", "cache" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/hawkular/client.go#L50-L52
train
kubernetes-retired/heapster
metrics/sinks/hawkular/client.go
pushToCache
func (h *hawkularSink) pushToCache(key string, hash uint64) { h.regLock.Lock() h.expReg[key] = &expiringItem{ hash: hash, ttl: h.runId, } h.regLock.Unlock() }
go
func (h *hawkularSink) pushToCache(key string, hash uint64) { h.regLock.Lock() h.expReg[key] = &expiringItem{ hash: hash, ttl: h.runId, } h.regLock.Unlock() }
[ "func", "(", "h", "*", "hawkularSink", ")", "pushToCache", "(", "key", "string", ",", "hash", "uint64", ")", "{", "h", ".", "regLock", ".", "Lock", "(", ")", "\n", "h", ".", "expReg", "[", "key", "]", "=", "&", "expiringItem", "{", "hash", ":", "hash", ",", "ttl", ":", "h", ".", "runId", ",", "}", "\n", "h", ".", "regLock", ".", "Unlock", "(", ")", "\n", "}" ]
// toCache inserts the item and updates the TTL in the cache to current time
[ "toCache", "inserts", "the", "item", "and", "updates", "the", "TTL", "in", "the", "cache", "to", "current", "time" ]
e1e83412787b60d8a70088f09a2cb12339b305c3
https://github.com/kubernetes-retired/heapster/blob/e1e83412787b60d8a70088f09a2cb12339b305c3/metrics/sinks/hawkular/client.go#L55-L62
train