mirror of https://github.com/vladmandic/human
update docs
parent
99c3d70ec7
commit
91af84e954
|
@ -91,8 +91,8 @@ After=network.target network-online.target
|
||||||
[Service]
|
[Service]
|
||||||
Type=simple
|
Type=simple
|
||||||
Environment="NODE_ENV=production"
|
Environment="NODE_ENV=production"
|
||||||
ExecStart=/home/vlado/.nvm/versions/node/v15.7.0/bin/node server/serve.js
|
ExecStart=<path-to-node> server/serve.js
|
||||||
WorkingDirectory=/home/vlado/dev/human
|
WorkingDirectory=<your-project-folder>
|
||||||
StandardOutput=inherit
|
StandardOutput=inherit
|
||||||
StandardError=inherit
|
StandardError=inherit
|
||||||
Restart=always
|
Restart=always
|
||||||
|
|
15
Outputs.md
15
Outputs.md
|
@ -11,8 +11,8 @@ result = {
|
||||||
confidence, // <number> returns faceConfidence if exists, otherwise boxConfidence
|
confidence, // <number> returns faceConfidence if exists, otherwise boxConfidence
|
||||||
faceConfidence // <number> confidence in detection box after running mesh
|
faceConfidence // <number> confidence in detection box after running mesh
|
||||||
boxConfidence // <number> confidence in detection box before running mesh
|
boxConfidence // <number> confidence in detection box before running mesh
|
||||||
box, // <array [x, y, width, height]>, normalized to input image size
|
box, // <array [x, y, width, height]>, clamped and normalized to input image size
|
||||||
boxRaw, // <array [x, y, width, height]>, normalized to range of 0..1
|
boxRaw, // <array [x, y, width, height]>, unclamped and normalized to range of 0..1
|
||||||
mesh, // <array of 3D points [x, y, z]> 468 base points & 10 iris points, normalized to input impact size
|
mesh, // <array of 3D points [x, y, z]> 468 base points & 10 iris points, normalized to input impact size
|
||||||
meshRaw, // <array of 3D points [x, y, z]> 468 base points & 10 iris points, normalized to range of 0..1
|
meshRaw, // <array of 3D points [x, y, z]> 468 base points & 10 iris points, normalized to range of 0..1
|
||||||
annotations, // <list of object { landmark: array of points }> 32 base annotated landmarks & 2 iris annotations
|
annotations, // <list of object { landmark: array of points }> 32 base annotated landmarks & 2 iris annotations
|
||||||
|
@ -52,7 +52,8 @@ result = {
|
||||||
[
|
[
|
||||||
{
|
{
|
||||||
confidence, // <number>,
|
confidence, // <number>,
|
||||||
box, // <array [x, y, width, height]>,
|
box, // <array [x, y, width, height]>, clamped and normalized to input image size
|
||||||
|
boxRaw, // <array [x, y, width, height]>, unclamped and normalized to range of 0..1
|
||||||
landmarks, // <array of 3D points [x, y,z]> 21 points
|
landmarks, // <array of 3D points [x, y,z]> 21 points
|
||||||
annotations, // <array of 3D landmarks [ landmark: <array of points> ]> 5 annotated landmakrs
|
annotations, // <array of 3D landmarks [ landmark: <array of points> ]> 5 annotated landmakrs
|
||||||
}
|
}
|
||||||
|
@ -63,10 +64,10 @@ result = {
|
||||||
score, // <number>
|
score, // <number>
|
||||||
class, // <number> class id based on coco labels
|
class, // <number> class id based on coco labels
|
||||||
label, // <string> class label based on coco labels
|
label, // <string> class label based on coco labels
|
||||||
center, // [<number>, <number>] x,y coordinates of the object center
|
center, // <array [x, y]>, clamped and normalized to input image size
|
||||||
box, // [<number>, <number>, <number>, <number>] x1, y1, x2, y2 coordinates of the box around the detected object
|
centerRaw, // <array [x, y]>, unclamped and normalized to range of 0..1
|
||||||
centerRaw, // same as center but normalized to range 0..1
|
box, // <array [x, y, width, height]>, clamped and normalized to input image size
|
||||||
boxRaw, // same as box, but normalized to range 0..1
|
boxRaw, // <array [x, y, width, height]>, unclamped and normalized to range of 0..1
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
gesture: // <array of objects object>
|
gesture: // <array of objects object>
|
||||||
|
|
Loading…
Reference in New Issue