mirror of https://github.com/vladmandic/human
198 lines
6.1 KiB
TypeScript
198 lines
6.1 KiB
TypeScript
/**
|
|
* Type definitions for Human result object
|
|
*/
|
|
import type { Tensor } from './tfjs/types';
|
|
import type { FaceGesture, BodyGesture, HandGesture, IrisGesture } from './gesture/gesture';
|
|
/** generic box as [x, y, width, height] */
|
|
export declare type Box = [number, number, number, number];
|
|
/** generic point as [x, y, z?] */
|
|
export declare type Point = [number, number, number?];
|
|
/** Face results
|
|
* - Combined results of face detector, face mesh, age, gender, emotion, embedding, iris models
|
|
* - Some values may be null if specific model is not enabled
|
|
*/
|
|
export interface FaceResult {
|
|
/** face id */
|
|
id: number;
|
|
/** overall face score */
|
|
score: number;
|
|
/** detection score */
|
|
boxScore: number;
|
|
/** mesh score */
|
|
faceScore: number;
|
|
/** detected face box */
|
|
box: Box;
|
|
/** detected face box normalized to 0..1 */
|
|
boxRaw: Box;
|
|
/** detected face mesh */
|
|
mesh: Array<Point>;
|
|
/** detected face mesh normalized to 0..1 */
|
|
meshRaw: Array<Point>;
|
|
/** mesh keypoints combined into annotated results */
|
|
annotations: Record<string, Point[]>;
|
|
/** detected age */
|
|
age?: number;
|
|
/** detected gender */
|
|
gender?: string;
|
|
/** gender detection score */
|
|
genderScore?: number;
|
|
/** detected emotions */
|
|
emotion?: Array<{
|
|
score: number;
|
|
emotion: string;
|
|
}>;
|
|
/** face descriptor */
|
|
embedding?: Array<number>;
|
|
/** face iris distance from camera */
|
|
iris?: number;
|
|
/** face anti-spoofing result confidence */
|
|
real?: number;
|
|
/** face rotation details */
|
|
rotation?: {
|
|
angle: {
|
|
roll: number;
|
|
yaw: number;
|
|
pitch: number;
|
|
};
|
|
matrix: [number, number, number, number, number, number, number, number, number];
|
|
gaze: {
|
|
bearing: number;
|
|
strength: number;
|
|
};
|
|
};
|
|
/** detected face as tensor that can be used in further pipelines */
|
|
tensor?: Tensor;
|
|
}
|
|
export interface BodyKeypoint {
|
|
/** body part name */
|
|
part: string;
|
|
/** body part position */
|
|
position: Point;
|
|
/** body part position normalized to 0..1 */
|
|
positionRaw: Point;
|
|
/** body part detection score */
|
|
score: number;
|
|
}
|
|
/** Body results */
|
|
export interface BodyResult {
|
|
/** body id */
|
|
id: number;
|
|
/** body detection score */
|
|
score: number;
|
|
/** detected body box */
|
|
box: Box;
|
|
/** detected body box normalized to 0..1 */
|
|
boxRaw: Box;
|
|
/** detected body keypoints */
|
|
keypoints: Array<BodyKeypoint>;
|
|
/** detected body keypoints combined into annotated parts */
|
|
annotations: Record<string, Array<Point[]>>;
|
|
}
|
|
/** Hand results */
|
|
export interface HandResult {
|
|
/** hand id */
|
|
id: number;
|
|
/** hand overal score */
|
|
score: number;
|
|
/** hand detection score */
|
|
boxScore: number;
|
|
/** hand skelton score */
|
|
fingerScore: number;
|
|
/** detected hand box */
|
|
box: Box;
|
|
/** detected hand box normalized to 0..1 */
|
|
boxRaw: Box;
|
|
/** detected hand keypoints */
|
|
keypoints: Array<Point>;
|
|
/** detected hand class */
|
|
label: string;
|
|
/** detected hand keypoints combined into annotated parts */
|
|
annotations: Record<'index' | 'middle' | 'pinky' | 'ring' | 'thumb' | 'palm', Array<Point>>;
|
|
/** detected hand parts annotated with part gestures */
|
|
landmarks: Record<'index' | 'middle' | 'pinky' | 'ring' | 'thumb', {
|
|
curl: 'none' | 'half' | 'full';
|
|
direction: 'verticalUp' | 'verticalDown' | 'horizontalLeft' | 'horizontalRight' | 'diagonalUpRight' | 'diagonalUpLeft' | 'diagonalDownRight' | 'diagonalDownLeft';
|
|
}>;
|
|
}
|
|
/** Object results */
|
|
export interface ObjectResult {
|
|
/** object id */
|
|
id: number;
|
|
/** object detection score */
|
|
score: number;
|
|
/** detected object class id */
|
|
class: number;
|
|
/** detected object class name */
|
|
label: string;
|
|
/** detected object box */
|
|
box: Box;
|
|
/** detected object box normalized to 0..1 */
|
|
boxRaw: Box;
|
|
}
|
|
/** Gesture combined results
|
|
* @typedef Gesture Type
|
|
* Each result has:
|
|
* - part: part name and number where gesture was detected: `face`, `iris`, `body`, `hand`
|
|
* - gesture: gesture detected
|
|
*/
|
|
export declare type GestureResult = {
|
|
'face': number;
|
|
gesture: FaceGesture;
|
|
} | {
|
|
'iris': number;
|
|
gesture: IrisGesture;
|
|
} | {
|
|
'body': number;
|
|
gesture: BodyGesture;
|
|
} | {
|
|
'hand': number;
|
|
gesture: HandGesture;
|
|
};
|
|
/** Person getter
|
|
* - Triggers combining all individual results into a virtual person object
|
|
*/
|
|
export interface PersonResult {
|
|
/** person id */
|
|
id: number;
|
|
/** face result that belongs to this person */
|
|
face: FaceResult;
|
|
/** body result that belongs to this person */
|
|
body: BodyResult | null;
|
|
/** left and right hand results that belong to this person */
|
|
hands: {
|
|
left: HandResult | null;
|
|
right: HandResult | null;
|
|
};
|
|
/** detected gestures specific to this person */
|
|
gestures: Array<GestureResult>;
|
|
/** box that defines the person */
|
|
box: Box;
|
|
/** box that defines the person normalized to 0..1 */
|
|
boxRaw?: Box;
|
|
}
|
|
/**
|
|
* Result interface definition for **Human** library
|
|
*
|
|
* Contains all possible detection results
|
|
*/
|
|
export interface Result {
|
|
/** {@link FaceResult}: detection & analysis results */
|
|
face: Array<FaceResult>;
|
|
/** {@link BodyResult}: detection & analysis results */
|
|
body: Array<BodyResult>;
|
|
/** {@link HandResult}: detection & analysis results */
|
|
hand: Array<HandResult>;
|
|
/** {@link GestureResult}: detection & analysis results */
|
|
gesture: Array<GestureResult>;
|
|
/** {@link ObjectResult}: detection & analysis results */
|
|
object: Array<ObjectResult>;
|
|
/** global performance object with timing values for each operation */
|
|
performance: Record<string, number>;
|
|
/** optional processed canvas that can be used to draw input on screen */
|
|
canvas?: OffscreenCanvas | HTMLCanvasElement | null | undefined;
|
|
/** timestamp of detection representing the milliseconds elapsed since the UNIX epoch */
|
|
readonly timestamp: number;
|
|
/** getter property that returns unified persons object */
|
|
persons: Array<PersonResult>;
|
|
}
|
|
//# sourceMappingURL=result.d.ts.map
|