Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
X
xanadu
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Analytics
Analytics
Repository
Value Stream
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Commits
Open sidebar
SYNERG
xanadu
Commits
088d0e1a
Commit
088d0e1a
authored
Mar 11, 2020
by
Nilanjan Daw
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
changed metrics format for better readability
parent
f11341bb
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
75 additions
and
60 deletions
+75
-60
dispatch_system/dispatch_manager/index.js
dispatch_system/dispatch_manager/index.js
+23
-15
dispatch_system/dispatch_manager/metrics.js
dispatch_system/dispatch_manager/metrics.js
+52
-45
No files found.
dispatch_system/dispatch_manager/index.js
View file @
088d0e1a
...
...
@@ -324,26 +324,26 @@ function dispatch() {
*/
function
postDeploy
(
message
)
{
logger
.
info
(
"
Deployed Resource:
"
+
JSON
.
stringify
(
message
));
let
id
=
message
.
functionHash
+
message
.
runtime
if
(
message
.
status
==
false
)
{
let
sendQueue
=
db
.
get
(
message
.
functionHash
+
message
.
runtime
)
let
sendQueue
=
db
.
get
(
id
)
// TODO: handle failure
while
(
sendQueue
&&
sendQueue
.
length
!=
0
)
{
let
{
req
,
res
}
=
sendQueue
.
shift
()
res
.
status
(
400
).
json
({
reason
:
message
.
reason
})
}
db
.
delete
(
message
.
functionHash
+
message
.
runtime
)
db
.
delete
(
id
)
return
;
}
if
(
functionToResource
.
has
(
message
.
functionHash
+
message
.
runtime
))
{
let
resourceHeap
=
functionToResource
.
get
(
message
.
functionHash
+
message
.
runtime
)
if
(
functionToResource
.
has
(
id
))
{
let
resourceHeap
=
functionToResource
.
get
(
id
)
heap
.
push
(
resourceHeap
,
{
resource_id
:
message
.
resource_id
,
open_request_count
:
0
},
libSupport
.
compare
)
logger
.
warn
(
"
Horizontally scaling up:
"
+
JSON
.
stringify
(
functionToResource
.
get
(
message
.
functionHash
+
message
.
runtime
)));
JSON
.
stringify
(
functionToResource
.
get
(
id
)));
}
else
{
/**
...
...
@@ -356,9 +356,9 @@ function postDeploy(message) {
resource_id
:
message
.
resource_id
,
open_request_count
:
0
},
libSupport
.
compare
)
functionToResource
.
set
(
message
.
functionHash
+
message
.
runtime
,
resourceHeap
)
functionToResource
.
set
(
id
,
resourceHeap
)
logger
.
warn
(
"
Creating new resource pool
"
+
JSON
.
stringify
(
functionToResource
.
get
(
message
.
functionHash
+
message
.
runtime
)));
+
JSON
.
stringify
(
functionToResource
.
get
(
id
)));
}
...
...
@@ -372,8 +372,8 @@ function postDeploy(message) {
starttime
:
(
Date
.
now
()
-
resource
.
deploy_request_time
)
},
message
.
resource_id
,
resourceMap
)
if
(
db
.
has
(
message
.
functionHash
+
message
.
runtime
))
{
let
sendQueue
=
db
.
get
(
message
.
functionHash
+
message
.
runtime
)
if
(
db
.
has
(
id
))
{
let
sendQueue
=
db
.
get
(
id
)
logger
.
info
(
"
forwarding request via reverse proxy to:
"
+
JSON
.
stringify
(
resource
));
while
(
sendQueue
&&
sendQueue
.
length
!=
0
)
{
let
{
req
,
res
}
=
sendQueue
.
shift
()
...
...
@@ -382,8 +382,12 @@ function postDeploy(message) {
})
}
db
.
delete
(
message
.
functionHash
+
message
.
runtime
)
db
.
delete
(
id
)
}
libSupport
.
metrics
.
collectMetrics
({
type
:
"
scale
"
,
value
:
functionToResource
.
get
(
id
).
length
,
functionHash
:
id
})
}
catch
(
e
)
{
logger
.
error
(
e
.
message
)
}
...
...
@@ -423,8 +427,9 @@ consumer.on('message', function (message) {
// process.exit(0)
}
usedPort
.
delete
(
message
.
port
)
if
(
functionToResource
.
has
(
message
.
functionHash
+
message
.
runtime
))
{
let
resourceArray
=
functionToResource
.
get
(
message
.
functionHash
+
message
.
runtime
)
let
id
=
message
.
functionHash
+
message
.
runtime
if
(
functionToResource
.
has
(
id
))
{
let
resourceArray
=
functionToResource
.
get
(
id
)
for
(
let
i
=
0
;
i
<
resourceArray
.
length
;
i
++
)
if
(
resourceArray
[
i
].
resource_id
===
message
.
resource_id
)
{
resourceArray
.
splice
(
i
,
1
);
...
...
@@ -432,6 +437,9 @@ consumer.on('message', function (message) {
}
heap
.
heapify
(
resourceArray
,
libSupport
.
compare
)
libSupport
.
metrics
.
collectMetrics
({
type
:
"
scale
"
,
value
:
resourceArray
.
length
,
functionHash
:
id
})
libSupport
.
logBroadcast
({
entity_id
:
message
.
entity_id
,
"
reason
"
:
"
terminate
"
,
...
...
@@ -441,7 +449,7 @@ consumer.on('message', function (message) {
.
then
(()
=>
{
resourceMap
.
delete
(
message
.
resource_id
)
if
(
resourceArray
.
length
==
0
)
functionToResource
.
delete
(
message
.
functionHash
+
message
.
runtime
)
functionToResource
.
delete
(
id
)
})
}
...
...
@@ -547,5 +555,5 @@ function periodicMetricBroadcast() {
setInterval
(
libSupport
.
viterbi
,
1000
,
functionBranchTree
)
setInterval
(
autoscalar
,
1000
);
setInterval
(
dispatch
,
1000
);
setInterval
(
periodicMetricBroadcast
,
5000
)
//
setInterval(periodicMetricBroadcast, 5000)
app
.
listen
(
port
,
()
=>
logger
.
info
(
`Server listening on port
${
port
}
!`
))
\ No newline at end of file
dispatch_system/dispatch_manager/metrics.js
View file @
088d0e1a
const
constants
=
require
(
'
.././constants.json
'
);
let
log_channel
=
constants
.
log_channel
,
metrics
=
{
longterm
:
{},
shortterm
:
{}
}
metrics
=
{
}
let
kafka
=
require
(
'
kafka-node
'
),
Producer
=
kafka
.
Producer
,
...
...
@@ -13,66 +13,64 @@ let kafka = require('kafka-node'),
function
collectMetrics
(
metric
)
{
if
(
!
(
metric
.
functionHash
in
metrics
.
shortterm
))
{
metrics
.
shortterm
[
metric
.
functionHash
]
=
{
coldstart
:
0
,
coldstart_total_request
:
0
,
warm_total_request
:
0
,
warmstart
:
0
,
worker_count
:
0
if
(
!
(
metric
.
functionHash
in
metrics
))
{
metrics
[
metric
.
functionHash
]
=
{
shortterm
:
{
coldstart
:
0
,
coldstart_total_request
:
0
,
warm_total_request
:
0
,
warmstart
:
0
,
worker_count
:
0
}
}
}
if
(
metric
.
type
===
'
coldstart
'
)
{
metrics
.
shortterm
[
metric
.
functionHash
]
.
coldstart
+=
metric
.
value
metrics
.
shortterm
[
metric
.
functionHash
]
.
coldstart_total_request
+=
1
metrics
[
metric
.
functionHash
].
shortterm
.
coldstart
+=
metric
.
value
metrics
[
metric
.
functionHash
].
shortterm
.
coldstart_total_request
+=
1
}
else
if
(
metric
.
type
===
'
warmstart
'
)
{
metrics
.
shortterm
[
metric
.
functionHash
]
.
warmstart
+=
metric
.
value
metrics
.
shortterm
[
metric
.
functionHash
]
.
warm_total_request
+=
1
metrics
[
metric
.
functionHash
].
shortterm
.
warmstart
+=
metric
.
value
metrics
[
metric
.
functionHash
].
shortterm
.
warm_total_request
+=
1
}
else
if
(
metric
.
type
===
'
scale
'
)
{
metrics
.
shortterm
[
metric
.
functionHash
].
worker_count
=
metric
.
value
metrics
[
metric
.
functionHash
].
worker_count
=
metric
.
value
}
// console.log(metrics);
}
/**
* FIXME: Some error causing longterm metrics to be flushed.
*/
function
broadcastMetrics
()
{
if
(
Object
.
keys
(
metrics
.
shortterm
).
length
!==
0
)
{
for
(
let
[
functionHash
,
metric
]
of
Object
.
entries
(
metrics
.
shortterm
))
{
if
(
metrics
.
longterm
[
functionHash
]
===
undefined
)
{
metrics
.
longterm
[
functionHash
]
=
{
coldstart
:
0
,
coldstart_total_request
:
0
,
warm_total_request
:
0
,
warmstart
:
0
,
if
(
Object
.
keys
(
metrics
).
length
!==
0
)
{
for
(
let
[
functionHash
,
metric
]
of
Object
.
entries
(
metrics
))
{
if
(
metric
.
longterm
===
undefined
)
{
metric
.
longterm
=
{
coldstart
:
0
,
coldstart_total_request
:
0
,
warm_total_request
:
0
,
warmstart
:
0
}
}
metric
s
.
longterm
[
functionHash
].
coldstart
=
metrics
.
longterm
[
functionHash
]
.
coldstart
*
metric
s
.
longterm
[
functionHash
]
.
coldstart_total_request
+
metric
.
coldstart
metric
.
longterm
.
coldstart
=
metric
.
longterm
.
coldstart
*
metric
.
longterm
.
coldstart_total_request
+
metric
.
shortterm
.
coldstart
metric
s
.
longterm
[
functionHash
].
coldstart_total_request
+=
metric
.
coldstart_total_request
metric
s
.
longterm
[
functionHash
].
coldstart
/=
(
metrics
.
longterm
[
functionHash
]
.
coldstart_total_request
!=
0
)?
metric
s
.
longterm
[
functionHash
]
.
coldstart_total_request
:
1
metric
.
longterm
.
coldstart_total_request
+=
metric
.
shortterm
.
coldstart_total_request
metric
.
longterm
.
coldstart
/=
(
metric
.
longterm
.
coldstart_total_request
!=
0
)?
metric
.
longterm
.
coldstart_total_request
:
1
metric
s
.
longterm
[
functionHash
].
warmstart
=
metrics
.
longterm
[
functionHash
]
.
warmstart
*
metric
s
.
longterm
[
functionHash
]
.
warm_total_request
+
metric
.
warmstart
metric
s
.
longterm
[
functionHash
].
warm_total_request
+=
metric
.
warm_total_request
metric
s
.
longterm
[
functionHash
].
warmstart
/=
(
metrics
.
longterm
[
functionHash
]
.
warm_total_request
!=
0
)?
metric
s
.
longterm
[
functionHash
]
.
warm_total_request
:
1
metric
.
longterm
.
warmstart
=
metric
.
longterm
.
warmstart
*
metric
.
longterm
.
warm_total_request
+
metric
.
shortterm
.
warmstart
metric
.
longterm
.
warm_total_request
+=
metric
.
shortterm
.
warm_total_request
metric
.
longterm
.
warmstart
/=
(
metric
.
longterm
.
warm_total_request
!=
0
)?
metric
.
longterm
.
warm_total_request
:
1
metric
.
coldstart
/=
(
metric
.
coldstart_total_request
!=
0
)?
metric
.
coldstart_total_request
:
1
metric
.
warmstart
/=
(
metric
.
warm_total_request
!=
0
)?
metric
.
warm_total_request
:
1
metric
.
shortterm
.
coldstart
/=
(
metric
.
shortterm
.
coldstart_total_request
!=
0
)?
metric
.
shortterm
.
coldstart_total_request
:
1
metric
.
shortterm
.
warmstart
/=
(
metric
.
shortterm
.
warm_total_request
!=
0
)?
metric
.
shortterm
.
warm_total_request
:
1
metric
.
timestamp
=
Date
.
now
()
}
metrics
.
timestamp
=
Date
.
now
()
console
.
log
(
metrics
);
let
log
=
[{
topic
:
log_channel
,
...
...
@@ -82,7 +80,16 @@ function broadcastMetrics() {
partition
:
0
}]
producer
.
send
(
log
,
()
=>
{
})
metrics
.
shortterm
=
{}
for
(
let
[
functionHash
,
metric
]
of
Object
.
entries
(
metrics
))
{
metric
.
shortterm
=
{
coldstart
:
0
,
coldstart_total_request
:
0
,
warm_total_request
:
0
,
warmstart
:
0
,
worker_count
:
0
}
}
}
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment