Glossary#
Term | Description |
---|---|
Age category (group) | A group of people that is within a specified range of ages. In accordance with the age periodization of the World Health Organization, groups 18โ44 (young adults), 45โ60 (middle-aged adults), 61โ75 (older adults) are distinguished |
Attributes | Age, gender, automatically determined by the system |
Authorization | A security mechanism to determine access levels or user privileges related to system resources |
Avatar | A visual representation of the face that can be used in the user interface |
Best shot | A frame of the video stream, in which the face is captured in the optimal angle for further use in the face recognition system |
Biometric sample (sample) | Analog or digital representation of biometric characteristics prior to biometric feature extraction and descriptor generation |
Body | A LUNA PLATFORM 5 mutable object containing information about the human body |
Candidate | Applicant for similarity with the reference |
Cross-matching | Many-to-many comparison (M:N). In the context of this document, comparing multiple lists |
Descriptor | A binary data set prepared by the system based on analyzed characteristic. It is a composite vector of person's face attributes |
Event card | Information block displaying data about an event |
External ID | External face identifier |
Face card | Information block displaying data about a face |
Face ID | A face identifier that is created in the LUNA PLATFORM 5 system as a result of a detection event and attribute extraction |
Task | A task that is created by a user and runs in the background |
Department | An entity that grants specific users access to selected event sources |
Detection | FaceStream entity that contains the coordinates of face or body and the estimated value of the object that determines the best shot |
Event | Detection recorded by the system with the extraction of attributes by the handler |
Exchangeable image file format (EXIF) | A standard for embedding technical metadata in image files that many camera manufacturers use and many image-processing programs support |
Extraction | A descriptor extraction procedure |
Face recognition | A set of methods for collecting, processing, and storing data of person's face images for identity recognition or identity confirmation using mathematical methods |
Faces | Changeable LUNA PLATFORM 5 objects that contain information about one person |
Handler | Image processing entry points that characterize the image processing procedure and define the LUNA PLATFORM 5 algorithms used for this |
Handling policy | A set of rules (policies) for image processing |
Identification | Search for the most suitable descriptor by comparing the vectors of face features with a list of similar descriptors in the database (one to many) |
List | A set of faces in the LUNA PLATFORM 5 system, combined automatically or manually according to a certain criterion |
Liveness | A software method to confirm the vitality of a person by one or several images in order to prevent spoofing attacks |
LUNA PLATFORM 5 (LP5) | VisionLabs automated facial recognition system designed to process, collect, analyze, store, and compare biometric data obtained from facial images |
Matching | A procedure of matching descriptors for the purpose of comparison |
Physical access control system (PACS) | A set of hardware and software tools aimed at controlling the entrance and exit in order to ensure safety and regulate visits to a particular facility |
Reference | Object (attribute, face, body, face and event external IDs, event track ID, descriptor) that is compared/verified with the candidate. |
Software | A program or set of programs used to control a computer |
Similarity | Probability characteristic in the range from 0 to 1, characterizing the level of similarity of subjects of biometric data |
Spoofing attack | Substitution of a real person for a fake image (for example, a photograph) to deceive the system |
Track | Information about objectโs position (face or body of a person) in a sequence of frames |
Introduction#
This document describes the purpose and functions of the LUNA PLATFORM 5 UI user interface (hereinafter referred to as Interface) version {{LUNA}}.
All information provided in the documentation is for informational purposes only. The use of the product may vary significantly depending on various factors (case, legality of use, compliance with the law and regulatory requirements, etc.) and depends on individual circumstances.
Overview#
LUNA PLATFORM 5 UI is a user interface that provides user interaction with LUNA PLATFORM 5 for operating with events and lists.
LUNA PLATFORM 5 UI allows the user to capture and view events according to a customized policy. For example, when identifying persons using control lists, user can search among events for a certain period of time by various attributes and photographic image of a person.
The main functions of LUNA PLATFORM 5 UI are presented below:
- show detection and object recognition events (faces, bodies);
- show information about the temperature of a person, filter events by temperature;
- search through the archive of events;
- create, view and edit face cards containing information about a personโs face;
- create, view and edit lists;
- identify faces, bodies and uploaded photo images by lists;
- face and body verification;
- verify personโs identity;
- create and configure handling policies;
- verification of compliance of the photo with the requirements of biometric standards;
- create tasks (cross-matching of lists, export of faces, bodies and events, batch processing of photo images, batch import of photo images, batch identification of photo images, batch deleting faces from the list);
- show static data about events;
- evaluate uploaded photo images for Liveness, DeepFake, compliance with the requirements of the ISO/IEC 19794-5:2011 standard, ICAO standard, biometric standards;
- show information about user accounts;
- show information about the status of connected components and systems;
- show information information about available licenses;
- show information information about plugins imported into LUNA PLATFORM 5.
System requirements#
Hardware requirements#
To get started with LUNA PLATFORM 5 UI, make sure you can meet the following hardware requirements.
Resource |
Minimum |
Recommended |
---|---|---|
CPU |
Intel Core i3, 2nd Generation / AMD Athlon X4 860K |
Intel Core i3, 4th Generation and above / AMD Ryzen 3 and above |
RAM |
2 GB |
4 GB and above |
Display resolution |
1024 px (for example, 1024x768), 1920px (for example, 1920x1080) |
- |
Software requirements#
To get started with LUNA PLATFORM 5 UI, make sure you can meet the following software and Internet connection requirements.
Resource |
Recommended |
---|---|
Supported web browser |
Google Chrome (version 109.0 and above); Microsoft Edge (version 109.0 and above); Mozilla Firefox (version 109.0 and above). |
It is recommended to update your browser to the latest version. Check browser updates:
|
Installing and configuring the above software is beyond the scope of this document.
Working with interface#
Authorization in interface#
Create account using a POST request "create account" to the API service, or using the Admin service. When creating the account, you must specify the following data: login (email), password and account type.
The Interface is accessed by logging in to the website at <host:5000/ui>
in a web browser. Authorization form is launched when you log first time into LUNA PLATFORM 5 UI (Figure 1). For authorization in the Interface, enter your credentials (email and password) in the appropriate fields and click the โLoginโ button.
When logged in, the user is taken to the โLast eventsโ section (Figure 2).
Switching the interface theme#
The Interface allows you to customize the color theme. Fot this, click on the icon in the top main menu:
-
to activate night mode or dark theme;
-
to activate day mode or light theme.
Sign out of account#
To log out of your account, click the arrow on the right of the userโs name. Click the โExitโ button (Figure 3).
After clicking on the โExitโ button, the user is moved to the authorization form.
Interface sections#
Switching between which is carried out in the main menu bar and in the drop-down menu (Figure 4).
The main menu consists of the โLast eventsโ, โEvents archiveโ, โSearchโ, "Faces" and โListsโ sections.
The drop-down menu consists of the following sections: โHandling policiesโ, โVerificationโ, โTasksโ, โChecksโ, โUsersโ, โMonitoringโ, โLicensesโ and โPluginsโ.
To expand the drop-down menu, click the arrow on the right of the userโs avatar.
Purpose of the sections of the main menu:
- โLast eventsโ displays the last 30 events, and it is possible to filter events by various parameters.
- โEvents archiveโ displays all events recorded by the Interface and it is possible to filter events by various parameters.
- โSearchโ allows user to search faces, bodies and events by the following parameters:
- by external face ID;
- by face image;
- by body image;
- by Face ID from LP5;
- by event ID from LP5.
- โFacesโ allows users to create, edit, and delete a faces.
- โListsโ allows users to create, edit, and delete lists.
Purpose of the sections of the drop-down menu:
- โHandling policiesโ allows user to create, delete, and edit policies (handlers);
- โVerificationโ allows user to create, delete, edit, and test verifiers. Verifiers are used to quickly compare two faces: by face photo and face ID, external ID, attribute, event, and display the test result;
- โTasksโ allows user to create, delete, and view tasks: cross-matching (comparison of two lists of faces), export of faces or events, batch processing of a photo archive according to a specific policy, batch import of an archive with photo images of faces into the list, and batch identification of an archive with photo images by faces or events.
- โCheckโ allows to check uploaded photo images for Liveness, DeepFake, compliance with the requirements of the ISO/IEC 19794-5:2011 standard, ICAO standard, biometric standards;
- โUsersโ shows the list of user accounts created in LUNA PLATFORM 5.
- โMonitoringโ shows status of the connected services, modules, components, and systems.
- โLicensesโ shows status of the available licenses;
- โPluginsโ shows status of plugins imported into LUNA PLATFORM 5.
Last events section#
The โLast eventsโ section displays detection and object (faces, bodies) recognition events, and records identification events using lists (Figure 5).
The section displays the last 30 events within the settings of handling policy for processing incoming images of the video stream, terminals, REST requests, etc. Receiving and displaying events is performed with minimal delays in near real time.
At the bottom of the screen, there is a โView events archiveโ button which leads to โEvents archiveโ section
The filter icon (1), which is located on the right, hides the block with filtering settings. The page shows the following event data (2):
- "Event image"
- a photo image of the face from the video stream;
- a photo image of the body from the video stream;
- "Top Match"โthe column is shown if the โDisplay top matchโ checkbox is active (3). If no matches are found for a photo from an event, then the column with the top match for this event will remain empty. The "Top match" includes:
- reference photo images of the face and/or body;
- value of similarity of the identified face with the reference in percentage terms and with the color coding of similarity thresholds:
- similarity values below โlowโ will be marked in red;
- similarity values between โlowโ and โmediumโ will be marked in yellow;
- similarity values above โmediumโ will be marked in green.
- "Match type"โthe type of object (face or event), according to which the similarity of the identified face/body with the reference was found;
- "External ID"โexternal identifier of the face, the field is shown if such an ID is available (for "Face" in "Match type");
- "User data"โinformation from the database, linked to a person from the control (for "Face" in "Match type");
- "List"โthe name of the list to which the person is attached (for "Face" in "Match type");
- "Date created"โdate and time of fixing the event (for "Event" in "Match type");
- "Source"โthe name of the source that recorded the event at the time the event was created. Users can change the source name. Then the "Video stream" field will show the new name, and the "Source" field will show the original one (for "Event" in "Match type");
- "Video stream"โthe current name of the source that recorded the event, with a link to the real-time image of the stream (for "Event" in "Match type");
- "Handling policy"โthe name of the handler, according to which the reference photo image of the body was processed (for "Event" in "Match type").
- "Event details" shows the available event data:
- "Date of created"โdate and time of event registration;
- "Source"โthe name of the source that recorded the event at the time the event was created. Users can change the source name. Then the "Video stream" field will show the new name, and the "Source" field will show the original one;
- "Video stream"โthe current name of the source that recorded the event, with a link to the real-time image of the stream;
- "Handling policy"โthe name of the handler, according to which the reference photo images of the face/body were processed;
- "Metadata" [^1]โbutton for uploading arbitrary user data in JSON format, the filed is shown if such data was added to the event (for "Event" in "Match type").
- Face attribute, if found:
- "Gender"โgender based on face image;
- โAge categoryโโthe age of the detected person;
- Deepfake check results, if such check was performed;
- Liveness check results, if such check was performed
- Body attributes, if found:
- "Upper body colors"โan indication of the color of the clothes of the human body upper part;
- "Lower body colors"โindicating the color of the human body upper part;
- "Headwear"โthe presence or absence of a headdress, if it is defined.
- "Backpack"โthe presence or absence of a backpack, if it is defined.
[^1]: All detailed capabilities and limitations of the "Metadata" field are specified in the "Administrator Manual" of LUNA PLATFORM 5 in the paragraph 6.9.4 "Events meta-information".
Last events filtering#
The Interface allows you to filter last events (2 in Figure 5) to find and display necessary events (Figure 6).
User can quickly find an event among the last 30, as well as set a limit for displaying new events on the screen.
When a user clicks on the icon (1 in Figure 5) on the โLast eventsโ page, a menu with settings and filters opens. The number next to the icon shows the number of applied filters. A short description of the elements and parameters of the filter block on the โLast eventsโ page is presented below (Table 1).
Table 1. Filters available to the user to search for last events
Name |
Description |
---|---|
"Sound notification" toggle and "Similarity threshold" parameter |
Allows a user to configure sound alerts about detection of an object that is not below the specified value of the "Similarity threshold" field |
General |
|
Source |
Filter by original source namesโselect one or more sources from the list of available ones |
Video stream |
Filter by current source namesโselect one or more sources from the list of available ones |
Handling policies |
Select one or more handling policies, according to which the face or body in the image was processed; |
Tags |
Selection of one or more tags. For example, the "Temperature" tag, is intended for displaying information about the temperature of the human body, filtering events by temperature. "Temperature":
Color coding of temperature values:
See the LUNA Access documentation for more information on setting of temperature ranges; |
Labels |
|
Label |
Name of the labelโrule by which the comparison occurred. Labels are specified when setting up the handling policy; |
Similarity,% |
Lower and/or upper limits of similarity for displaying faces identified by the lists; |
Face attributes and properties |
|
Gender |
Gender of a person to be detected, determined by the image of a face:
|
Age category |
Age range of a person to be detected, determined by the image of a face:
|
Emotion |
Emotion of a person to be detected:
A combination of several values is possible; |
Mask |
Indication of the presence of a mask:
A combination of several values is possible; |
Liveness |
Liveness status selection:
A combination of several values is possible; |
Deepfake (Deepfake license required) |
Liveness status selection:
A combination of several values is possible; |
Body attributes and properties |
|
Upper body colors |
Top clothing color specification:
A combination of several values is possible; |
Lower body type |
Bottom clothing type specification:
A combination of several values is possible; |
Lower body colors |
Bottom clothing color specification:
A combination of several values is possible; |
Shoes color |
Shoe color specification:
A combination of several values is possible; |
Headwear |
Headdress specification:
A combination of several values is possible; |
Headwear colors |
Headdress color specification:
A combination of several values is possible; |
Backpack |
Backpack presence specification:
A combination of several values is possible; |
Sleeve |
Sleeve length specification:
A combination of several values is possible; |
Gender by body |
Gender of a person to be detected, determined by the image of a body:
A combination of several values is possible; |
Age category by body |
Age range of a person to be detected, determined by the image of a body:
|
Location |
|
City Area District Street House number Longitude(-180โฆ180); Accuracy (0โฆ90); Latitude(-90โฆ90); Accuracy (0โฆ90); |
Event location |
Other |
|
Comma-separated track IDs |
Specifying event track IDs. Copy the track ID from the event details |
Add filter by meta |
Allows you to fill in a set of blocks to create a filter by the "meta" field. The number of meta filters is unlimited. The following blocks are required to be filled in when creating a filter by meta:
|
The user selects one filter or a combination of filters and clicks on the โFilterโ button for the applied filters to be applied.
To reset the applied filters, click on the โResetโ button.
The applied filters will affect the appearance of new events on the screen.
To collapse โFiltersโ, click on the filter icon on the right side of the screen.
Event details#
Click on an arrow button on the face or body from the event image (Figure 5) to open a page with detailed event data (Figure 7).
When an event contains data on the detection of both a face and a body, you can switch between these data on the page with event details. If an event contains detection data for only one object, such as a face, then there will be no detection data for another object.
The the page with event details consists of four blocks. The description of the elements of the page is presented below (Table 2).
Table 2. Elements and parameters of the "Event details" page
Name |
Description |
---|---|
Event information |
Basic information about the event |
Date created |
Date and time of event recording |
Event |
โEvent IDโ โ when clicked on |
Track |
โTrack IDโ โ when clicked on |
Handling policy |
Name of the policy by which the image processing in the video stream is performed. Clicking on the name of the policy opens the form for editing its parameters |
Source |
The name of the source that recorded the event at the time the event was created. Users can change the source name. Then the "Video stream" field will show the new name, and the "Source" field will show the original one |
Video Stream |
The current name of the source that recorded the event, with a link to the real-time image of the stream |
Tags |
Name of tags by which the event is filtered |
Metadata |
Uploading arbitrary user data in JSON format, if available |
Location |
Event location data: โCityโ, โAreaโ, โDistrictโ, โStreetโ, โHouse numberโ, โCoordinates (latitude)โ, โCoordinates (longitude)โ |
Detection |
If equipped: face and/or body detection |
Find similar: events |
Clicking on |
Find similar: faces |
Clicking on For face detection only |
Photo image of a face and/or body from a video stream |
Normalized image. When clicked on:
|
Attributes |
Face attributes:
If Liveness and Deepfake checks were performed:
|
Body attributes:
|
|
Additional properties |
Face properties:
|
Best match: Event or Person |
Similarity value of identified face/body with the face/body from control list/event (in percentage); |
Find similar: events |
Clicking on |
Find similar: faces |
Clicking on For face detection only |
Additional information |
โFaceโ type โ โInformationโ, โListsโ, โExternal IDโ. โEventโ type โ โHandling policyโ, โSourceโ with the ability to go to the handling policy editing page and view the stream from the source in real time |
Photo image of a face and/or body |
Reference photo image of the face or body (sample) or no photo. When clicked on
|
Attributes |
Face attributes:
Body attributes:
|
Additional properties |
Face properties:
|
"Matches" |
List of matches with a detected face and/or body |
Event photo |
โFaceโ typeโan avatar, sample, or no photo image. Similarity value of identified face with the face from control list (in percentage) โEventโ typeโdetection (photo image of a face from a video stream). Similarity value of identified face or body with the face or body from the event (in percentage) |
Type |
|
Date created |
Date and time of the biometric sample of the face or body creation from the identification event |
Label |
Name of the labelโrule by which the comparison occurred |
Additional information |
โFaceโ type โ โInformationโ, โListsโ, โExternal IDโ. โEventโ type โ โHandling policyโ, โSourceโ, โVideo streamโ with the ability to go to the handling policy editing page and view the stream from the source in real time |
|
When clicked, the face or event details opens |
The external ID is used to integrate LUNA PLATFORM 5 UI with external systems, to transfer data to other systems in order to analyze and quickly respond to an event.
Events archive section#
The โEvents archiveโ section is designed to display all events of face and body detection as well as recognition and search for events in archive (Figure 8).
Receiving and displaying new events in the event archive is performed with minimal delays in near real-time.
If there are no filters applied (1), the Interface displays the latest detection and identification events identical to those presented in the โLast eventsโ section from the last 30 days.
The number of events displayed on the page is set by the switch in the lower right corner of the page. There can be 10, 25, 50 or 100 events in total on one page (2).
Upload events according to the specified parameters by clicking on the "Export events" button (3): fill in the fields, click "Save", then go to the "Tasks" section and download the results.
The displayed data is identical to the data in the โLast eventsโ section.
Click on a line to open a page with event details.
Click on a reference photo of a face from the control list to open a page with face details.
Archived events filtering#
The Interface allows you to filter archived events (1 in Figure 8) to find and display necessary events.
With filters (Figure 9) user can quickly find an event among the last, as well as set a limit for displaying new events on the screen.
A short description of the elements and parameters of the block with filters in the "Events archive" section is presented below (Table 3).
Table 3. Filters available to the user to search for archived events:
Name |
Description |
---|---|
General |
|
Date from |
Start of the search period by date and time of the event; |
Date to |
End of the search period by date and time of the event; |
Source |
Filter by original source namesโselect one or more sources from the list of available ones |
Video stream |
Filter by current source namesโselect one or more sources from the list of available ones |
Handling policies |
Handling policy names, according to which the face or body in the image was processed. One or several handling policies can be selected for searching. |
Tags |
Selection of one or more tags. For example, the "Temperature" tag, is intended for displaying information about the temperature of the human body, filtering events by temperature. "Temperature":
Color coding of temperature values:
See the LUNA Access documentation for more information on setting of temperature ranges. |
Event ID |
Identifiers of detection and attribute extraction events. Values are separated by commas, for the correct search must be specified in full; |
External events ID |
External identifiers of events. Values are separated by commas, for the correct search must be specified in full; |
Labels |
|
Label |
Name of the labelโrule by which the comparison occurred. Labels are specified when setting up the handling policy |
Similarity,% |
Lower and/or upper limits of similarity for displaying faces identified by the lists; |
ID of objects with maximum match result |
The ID of the top similar object (event or face) from matching results (match policy, values are separated by commas, for the correct search must be specified in full); |
Face attributes and properties |
|
Gender |
Gender of a person to be detected:
|
Age category |
Lower and/or upper limits of age of a person to be detected:
|
Emotion |
Emotion of a person to be detected:
A combination of several values is possible; |
Mask |
Indication of the presence of a mask:
A combination of several values is possible; |
Liveness |
Liveness status selection:
A combination of several values is possible; |
Deepfake (Deepfake license required) |
Liveness status selection:
A combination of several values is possible; |
Face IDs from events |
Face IDs of persons that are created in the LUNA PLATFORM 5 system as a result of a detection event and extraction of attributes. Values are separated by commas, for the correct search must be specified in full; |
Body attributes and properties |
|
Upper body colors |
Top clothing color specification:
A combination of several values is possible; |
Lower body type |
Bottom clothing type specification:
A combination of several values is possible; |
Lower body colors |
Bottom clothing color specification:
A combination of several values is possible; |
Shoes color |
Shoe color specification:
A combination of several values is possible; |
Headwear |
Headdress specification:
A combination of several values is possible; |
Headwear colors |
Headdress color specification:
A combination of several values is possible; |
Backpack |
Backpack presence specification:
A combination of several values is possible; |
Sleeve |
Sleeve length specification:
A combination of several values is possible; |
Gender by body |
Gender of a person to be detected, determined by the image of a body:
A combination of several values is possible; |
Age category by body |
Age range of a person to be detected, determined by the image of a body:
|
Location |
|
City Area District Street House number Longitude(-180โฆ180); Accuracy (0โฆ90); Latitude(-90โฆ90); Accuracy (0โฆ90); |
Event location |
Other |
|
Comma-separated track IDs |
Specifying event track IDs. Copy the track ID from the event details |
Add filter by meta |
Allows you to fill in a set of blocks to create a filter by the "meta" field. The number of meta filters is unlimited. The following blocks are required to be filled in when creating a filter by meta:
|
The user selects one filter or a combination of filters and clicks on the โFilterโ button for the applied filters to be applied.
To reset the applied filters, click on the โResetโ button. To collapse Filters, click on the filter icon on the right side of the page.
Search section#
The โSearchโ section is designed to search by photo, event (event ID), and face (face ID: โExternal IDโ, โFace IDโ). This section displays all detection face and body recognition events that match the search conditions (Figure 10).
If departments are created, and users and event sources are attached to these departments, the user can search only for those events that were recorded by the event sources of his department.
The โSearchโ section contains the following blocks:
- Search options:
- โPhotoโ โ search by uploaded photo image:
- field for uploading a photo image;
- โEventโ โ search by registered event in the system:
- โEvent IDโ โ identifier of the event of detection and attribute extraction;
- โFaceโ โ search by registered face in the system:
- โExternal IDโ โ external face identifier;
- โFace IDโ โ face identifier that is created in the LUNA PLATFORM 5 system as a result of a detection event and attribute extraction;
- Searching results:
- โEventsโ:
- โDisplay search imageโ checkbox โ disable if you need to hide the column with the original photo;
- โSearch resultโ;
- โSearch imageโ;
- โDetailsโ;
- โFacesโ:
- โDisplay search imageโ checkbox โ disable if you need to hide the column with the original photo;
- โSearch resultโ;
- โSearch imageโ;
- โDetailsโ;
- Filters.
A short description of the elements and parameters of the filter block in the "Search" section is presented below (Table 4).
Table 4. โFiltersโ block elements description
Name |
Description |
---|---|
Search events |
|
Date from |
Start of the search period by date and time of the event; |
Date to |
End of the search period by date and time of the event; |
Source |
Filter by original source namesโselect one or more sources from the list of available ones |
Video stream |
Filter by current source namesโselect one or more sources from the list of available ones |
Handling policies |
Handling policy names, according to which the face or body in the image was processed. One or several handling policies can be selected for searching. |
Liveness |
Liveness status selection:
A combination of several values is possible; |
Tags |
Selection of one or more tags. For example, the "Temperature" tag, is intended for displaying information about the temperature of the human body, filtering events by temperature. "Temperature":
Color coding of temperature values:
See the LUNA Access documentation for more information on setting of temperature ranges. |
Event ID |
Identifiers of detection and attribute extraction events. Values are separated by commas, for the correct search must be specified in full); |
External events ID |
External identifiers of events. Values are separated by commas, for the correct search must be specified in full); |
Similarity is not less than, % |
The similarity value is not lower than the specified one, in percent; |
Labels |
|
Label |
Name of the labelโrule by which the comparison occurred. Labels are specified when setting up the handling policy |
ID of objects with maximum match result |
The ID of the top similar object (event or face) from matching results (match policy, values are separated by commas, for the correct search must be specified in full); |
Face attributes and properties |
|
Gender |
Gender of a person to be detected:
|
Age category |
Lower and/or upper limits of age of a person to be detected:
|
Emotion |
Emotion of a person to be detected:
A combination of several values is possible; |
Mask |
Indication of the presence of a mask:
A combination of several values is possible; |
Liveness |
Liveness status selection:
A combination of several values is possible; |
Deepfake (Deepfake license required) |
Liveness status selection:
A combination of several values is possible; |
Face IDs from events |
Face IDs of persons that are created in the LUNA PLATFORM 5 system as a result of a detection event and extraction of attributes. Values are separated by commas, for the correct search must be specified in full; |
Body attributes and properties |
|
Upper body colors |
Top clothing color specification:
A combination of several values is possible; |
Lower body type |
Bottom clothing type specification:
A combination of several values is possible; |
Lower body colors |
Bottom clothing color specification:
A combination of several values is possible; |
Shoes color |
Shoe color specification:
A combination of several values is possible; |
Headwear |
Headdress specification:
A combination of several values is possible; |
Headwear colors |
Headdress color specification:
A combination of several values is possible; |
Backpack |
Backpack presence specification:
A combination of several values is possible; |
Sleeve |
Sleeve length specification:
A combination of several values is possible; |
Gender by body |
Gender of a person to be detected, determined by the image of a body:
A combination of several values is possible; |
Age category by body |
Age range of a person to be detected, determined by the image of a body:
|
Location |
|
City Area District Street House number Longitude(-180โฆ180); Accuracy (0โฆ90); Latitude(-90โฆ90); Accuracy (0โฆ90); |
Event location |
Other |
|
Comma-separated track IDs |
Specifying event track IDs. Copy the track ID from the event details |
Add filter by meta |
Allows you to fill in a set of blocks to create a filter by the "meta" field. The number of meta filters is unlimited. The following blocks are required to be filled in when creating a filter by meta:
|
Search faces |
|
Date from |
Beginning of the search period by date and time of face creation |
Date before |
End of search period by date and time of face creation |
External events ID |
External event identifiers. To correctly search for the value indicated separated by commas and in full; |
Similarity is not less than, % |
The similarity value is not lower than the specified value, in percent |
Lists |
Selecting a list in which to search for a face |
User data |
Information about the person from the database (if available) |
To search by face or body image, select the "Photo" section, click on the field to upload an image from your computer, or drag and drop a photo into this field.
Image file requirements:
- *.jpeg, *.png or *.bmp format;
- image size no less than 320x250 and no more than 3840x2160 pixels;
- image may contain one or more people;
- image must have a person's face or body.
When loading a photo image containing many faces and/or bodies, the Interface detects all faces and/or bodies in the image, then displays them to the right of the loaded photo image and displays the number of detected faces and/or bodies (Figure 11). To reset the image, click on .
To find similar faces or events with the face or body, select one face or body by clicking on it on the uploaded photo image. Then, in the filter block, select the necessary search options and click "Filter". To reset the parameter values, click the "Reset" button. The search results will be displayed at the bottom of the page (Figure 12).
The description of the elements of the search results block is presented below (Table 5 and 6).
Table 5. Elements and parameters of the search results, if "Search events" is selected in the block with filters:
Name |
Description |
---|---|
Search result |
The face and/or body of a person from the event that is the most similar to the one selected in the uploaded photo |
Search image |
Normalized photo image of a person's face or body from the uploaded photo. Displayed if the โDisplay search imageโ checkbox is enabled |
Details |
|
|
|
Face attributes, if a face is selected as the search object |
If Liveness and Deepfake checks were performed:
|
Body attributes, if a body is selected as the search object |
|
|
Go to the โEvent detailsโ page |
Table 6. Elements and parameters of the search results, if "Search faces" is selected in the block with filters:
Name |
Description |
---|---|
Search result |
The face of a person from the event that is the most similar to the one selected in the uploaded photo |
Search image |
Normalized photo image of a person's face from the uploaded photo. Displayed if the โDisplay search imageโ checkbox is enabled |
Details |
|
|
Go to the โFace detailsโ page |
Faces section#
The โFacesโ section is intended for viewing, creating and deleting faces (Figure 13).
The โFacesโ section contains the following elements:
- List of faces (1):
- Photo of the face;
- Information, such as the temperature of the person whose face is in the photo;
- Date and time of creation of the face;
- Lists that contain the face;
- Face ID;
- Filters (2);
- Button to open the face creation form. (3)
Click on the filter icon to find events by (Figure 14):
- Listsโselect from the list the list to which the persons belong;
- Date of creationโselect the period during which the faces were created;
- External IDโspecify comma-separated external identifiers. Copy the external ID from the "Face details" page to find all faces with such external ID;
- Face IDโspecify comma-separated face identifiers that are created in the LUNA PLATFORM 5 system as a result of a detection event and extraction of attributes. Copy the face ID from the "Face details" page to find all faces with such IDs;
- Informationโspecify data associated with the person. Copy the information from the "Face details" page to find all faces with such information.
Click on from the list of faces to go to the face details.
Creating face#
To add a new face, click on the โCreate faceโ button in the upper right corner of the section page. The general view of the window for creating a face is presented below (Figure 15).
Enter the required information:
- Field for uploading a photo of the personโavatar (required to be filled out). There may be one face or several ones in the photo. If there is more than one face in the photo, then after uploading you have to select which face to be added in the list/listsโthe selected face will be highlighted with a green frame. You can add more than one person via batch import;
- โInformationโโinformation about the person. For example, the position of the employee whose face you are adding;
- โExternal IDโโexternal identifier of the face. If you need to add a face to an existing face card, copy the External ID from the "Face details" page and paste it into the field. In this case, several faces will be assigned to one external ID. If you need to create a face with a unique external ID, enter your own identifier in the field as a set of letters and/or numbers;
- โListsโโthe name of the list to which the face will be added (multiple lists can be selected);
- โCheck photo image quality for compliance with the ISO/IEC 19794-5:2011 standardโโif enabled, the photo will be added to the list only after passing the ISO/IEC 19794-5:2011 verification.
โbutton for resetting the uploaded photo image.
Image file requirements:
- *.jpeg, *.png or *.bmp format;
- image size no more than 15 MB and no more than 3840x2160 pixels;
- image may contain one or more persons;
- image must have person's face.
Fill in the fields and click the โSaveโ button. A message about the successful face creating will appear on the screen.
Face details#
To open the page with face details click on the arrow button on the face image from the "Top match" column in the "Last events" section or in the event details.
The "Face details" page consists of two blocks (Figure 16).
Descriptions of page elements are presented below (Table 7).
Table 7. Elements and parameters of the face details page
Name |
Description |
---|---|
Photo image of a face |
Avatar is a biometric sample that is created when uploading a photo image to the list (to the LUNA PLATFORM 5 system). When clicked on |
"Find similar faces" button |
When clicked, a search for faces by face ID is performed in a new tab |
Update photo |
Opening the form to upload a new face photo image |
Edit face user data |
Opening a form for editing face data (โInformationโ, โExternal IDโ, โListsโ) |
Delete face |
Removal of face biometric sample, face photo image, and face details |
Date created |
Date and time of creation of the biometric sample |
Information |
User data from the database, linked to a face (upon availability) |
External ID |
External identifier of the face |
Lists (N) |
The list and number of lists to which the person is attached. Clicking on the name opens the list |
Additional information |
โFace IDโ โ when clicked on |
Face attributes |
โAttributesโ: โGenderโ โ gender of a person (male/female); โAge categoryโ โ indication of age category. Hover the cursor over the card to find out the exact age of the person determined from the face image. |
Additional properties |
Face properties:
Indication of attributes and properties of the face/body is specified in the policy settings |
Faces with the same external ID |
Could be empty (โNo external IDโ) |
Photo image of a face |
Avatar, sample or no photo |
Date created |
Date and time of creation of the biometric sample |
Information |
User data from the database, linked to a face (upon availability) |
Lists (N) |
The list and number of lists to which the person is attached. Clicking on a name opens a list |
|
When clicked, face details open |
View all faces with the same external ID |
When clicked, the search for faces by external ID is performed, and a list of all faces, whose external ID matches the one of the reference photo, opens. |
Last events with this face |
Could be empty (โNo events with this faceโ) |
Photo image of a face from a video stream |
Normalized image:
Similarity value of identified face with the face from control list (in percentage) |
Date created |
Date and time of recording the event with a face |
Event source |
The name of the event source that recorded the event with a face When clicked on the name of the event source, a preview of the video stream in real time opens |
Handling policy |
Name of the policy that processed the image in the video stream Clicking on the name of the policy opens the form for editing its parameters |
|
When clicked, the event details opens |
View all events with this face |
When clicked, an events archive page with maximum match result this person opens |
Editing and deleting face#
Click the โUpdate photoโ button to update the photo image on the page with face details. The general view of the photo image update form is shown below (Figure 17).
Image file requirements:
- *.png , *.jpeg, or *.bmp format;
- image size no more than 15 MB and no more than 3840x2160 pixels;
- image may contain one or more people;
- image must have a person's face.
Click the โEdit face user dataโ button to edit the face user data. The general view of the face user data editing form is shown below (Figure 18).
Face user data editing form contains:
- โUser dataโ โ information from the database, linked to a face (upon availability);
- โExternal IDโ โ external identifier of the face;
- โListsโ โ lists to which the face is attached;
- โSaveโ button โ button for saving changes.
If you need to go back to the page with face details during editing, press the Esc key on your keyboard.
Click the โDelete faceโ button to delete the face along with its data. Confirm the action in the pop-up windowโclick the โDeleteโ button or cancel the action using the โCancelโ button (Esc key on the keyboard). After successful removal, a corresponding notification will appear.
Lists section#
The โListsโ section is intended for combining faces according to a certain criterion automatically or manually. For example, you can create a list with staff of the same unit and use it for a biometric identification policy through the list. This will allow you to compare the face detected in the frame with all the faces from the staff list and, if the comparison is successful, save the face to the specified list.
The โListsโ section allows you to create, delete, edit and view lists (Figure 19).
โListsโ section contains the following elements:
- table of lists:
- checkbox โ selection of a list or lists;
- โNameโ โ name of the list;
- โDate createdโ โ date and time when the list was created;
- โDate modifiedโ โ date and time when the list was last modified;
โ button for counting the number of faces in the list (1);
โ button for editing the list name (2);
โ button for detaching all faces from the selected list and deleting the selected list (3);
โ button for deleting all faces in the selected list (4);
โ button for deleting the list with faces (5);
- โAddโ button โ button for creating a list;
- โExport facesโ button โ button for downloading data on faces with certain parameters;
- โBatch importโ button โ button for uploading several photos with faces to the list at once;
- โDelete without facesโ button โ button for removing all faces from the list and deleting the list;
- โDelete with favesโ button โ button for deleting the list with faces in it;
- the number of lists displayed on the page is set by the switch in the lower right corner of the page. There can be 10, 25, 50 or 100 lists in total on one page (5).
In the table with lists, it is possible to sort by the columns โNameโ, โDate createdโ and โDate modifiedโ. To sort a column in the table, click on the column name.
The sorting arrow icon
indicates the current sorting by one of the parameters: alphabetically, ascending, or descending.
List creation#
To create a list, click on the โAddโ button in the lower left corner of the page.
The general view of the form for creating a list is shown below (Figure 20).
Enter a name for the list and click on the โSaveโ button. A message about the successful list creation will appear on the screen as well as the new list will appear in the table of lists.
Adding faces to the list#
To add a face to the list, click on the line with the name of the list to which you want to add the face. The form for editing the list will open (Figure 21).
The number of faces in this list is displayed next to the list name. To add a face to the list, click on the โAddโ button. A form for adding a face will open (Figure 22).
Enter the required information:
- Field for uploading a photo of the personโavatar (required to be filled out). There may be one face or several ones in the photo. If there is more than one face in the photo, then after uploading you have to select which face to be added in the list/listsโthe selected face will be highlighted with a green frame. You can add more than one person via batch import;
- โInformationโโinformation about the person. For example, the position of the employee whose face you are adding;
- โExternal IDโโexternal identifier of the face. If you need to add a face to an existing face card, copy the External ID from the "Face details" page and paste it into the field. In this case, several faces will be assigned to one external ID. If you need to create a face with a unique external ID, enter your own identifier in the field as a set of letters and/or numbers;
- โListsโโthe name of the list to which the face will be added (multiple lists can be selected);
- โCheck photo image quality for compliance with the ISO/IEC 19794-5:2011 standardโโif enabled, the photo will be added to the list only after passing the ISO/IEC 19794-5:2011 verification.
โbutton for resetting the uploaded photo image.
Image file requirements:
- *.jpeg, *.png or *.bmp format;
- image size no more than 15 MB and no more than 3840x2160 pixels;
- image may contain one or more persons;
- image must have person's face.
Fill in the fields and click on the โSaveโ button. A message about the successful face adding will appear on the screen.
The form for list editing allows to search for faces by user data, external ID or creation date in the line for quick search.
The added faces will be displayed in the form for list editing (Figure 23).
To export faces according to the specified parameters, click the "Export faces" button: fill in the fields, click "Save", then go to the "Tasks" section and download the results.
Sort the table with faces by the columns โUser dataโ, โExternal IDโ and โDate createdโ, clicking on the column name. The sorting arrow icon
indicates the current sorting by one of the parameters: alphabetically, ascending, or descending.
The number of faces displayed on the page is set by the switch in the lower right corner of the page. There can be 10, 25, 50 or 100 faces in total on one page.
To edit a face in the list, click on the button in the line with that face.
To detach a face from the list, click on the button in the line with that face.
To delete a face from the list, click on the button in the line with that face. To delete multiple faces from the list, select those faces and click on the โDeleteโ button. In the pop-up window (Figure 24), confirm the actionโclick on the โDeleteโ button or cancel the action by clicking on the โCancelโ button. Once a face or faces were successfully deleted from the list, a corresponding notification appears.
You can also delete more than one faces via creating task for deleting faces from the list;
To go to face details, click on the button in the line with the face on the list editing page.
List editing#
Editing the name of the list is performed by clicking on the button in the line (Figure 19). The general view of the form for editing the list name is shown below (Figure 25).
Change the name of the list and click on the โSaveโ button. A notification about successful list editing appears.
List deleting#
Deleting the list with faces is performed by clicking on the button (Figure 19).
To delete multiple lists, select those lists. Then click on the โDelete with facesโ button, if you need to delete both the list and the faces in it, or click the โDelete without facesโ button if you want to delete only the list. To detach all faces from the selected lists and delete the lists, check the boxes for the names of these lists, and click the button (Figure 26).
In the pop-up window (Figure 27), confirm the action โ click on the โDeleteโ button or cancel the action by clicking on the โCancelโ button. A corresponding notification appears after successful list deletion.
Handling policies section#
The โHandling policiesโ section is intended for creating, deleting, viewing policies, and editing their parameters (Figure 28).
Handling policies (handlers) can be static or dynamic.
If the handler is static, its parameters are specified when creating the handler.
If the handler is dynamic, then you can change its parameters when generating an event. For this, create a generate events
request with a specific content type (see API Reference Manual of the LUNA PLATFORM 5 documentation). In a dynamic handler, administrator can allow users to specify parameters that change with each request. At the same time other technical parameters can be set separately and left hidden from the user. With a static handler, administrator would have to create a new handler for each new task.
โHandling policiesโ section contains the following elements:
- table of policies:
- โDescriptionโโpolicy name;
- โHandling policy IDโโpolicy identifier;
- โHandler typeโโstatic or dynamic policy;
โbutton for editing policy parameters (1);
โbutton for deleting the policy (2);
- โAdd staticโ buttonโbutton for creating a static handling policy;
- โAdd dynamic buttonโbutton for creating a dynamic handling policy;
- the number of policies displayed on the page is set by the switch in the lower right corner of the page. There can be 10, 25, 50 or 100 policies in total on one page (3).
Policy creation#
Static policy creation#
To create a static policy, click on the โAdd staticโ button (Figure 28). A form will open to select how to create the static policy (Figure 29):
- preconfigured typical policy templates (policies 1โ6);
- step by step custom policy (โOtherโ).
To quickly create simple static policies, use one of the typical policy templates.
Six standard templates are available:
- โPolicy 1. Registration of a reference descriptor (with saving to a list)โโallows user to detect a face on the frame, check Liveness, and save the face to a specified list;
- โPolicy 2. Biometric identification of faces (without saving to a list)โโallows user to detect all faces in the frame and compare them with all faces in the specified list;
- โPolicy 3. Saving the faces identified in the list to the databaseโโallows user to detect all faces on the frame, check Liveness, compare detected faces with all faces in the specified list, and if the comparison is successful, save the face to the specified list;
- โPolicy 4. Determination of attributes and properties of a face without identification (gender, age, emotions, etc.)โโallows user to detect all faces in the frame, perform all possible checks, and save the event;
- โPolicy 5. Saving events for unique faces for later countingโโallows user to detect all faces in the frame, check Liveness, compare the detected faces with all faces in the list of unique faces, and if this face is not in the list, save the face to this list of unique faces;
- โPolicy 6. Registration of a reference descriptor with verification of compliance of the photo with the requirements of biometric standardsโallows you to save the reference descriptor in a specific list only for those photos that have been verified in accordance with the biometric standards.
The following checks are missing in the beta version:
- it is not allowed to use retouching and image editing;
- image cropping is allowed;
- compression code: JPEG (0 x 00), PNG (0 x 03).
When user clicks on a line with a standart template (policies 1โ6), a window opens for entering the main parameters of a preconfigured policy (Figure 30).
Fill in all the required parameters and click on the โCreateโ button. A window will open with a message about the successful creation of the policy (Figure 31).
Click anywhere outside the successful static policy generation message to navigate to the โSelect the type of policy you want to createโ form (Figure 29).
To create a unique static policy that requires detailed parameter settings, use the step-by-step custom policy.
When user clicks on the line with a step-by-step custom policy (โOtherโ), a form for step-by-step static policy creation will open (Figure 32).
Fill in all the required parameters and click on the โNextโ button to proceed to the next step. After setting all the parameters, a window with a message about the successful creation of the policy will open.
Dynamic policy creation#
To create a dynamic policy, click on the โAdd dynamicโ button on the page with the list of policies (Figure 28). In the opened window, enter the name of the new dynamic policy and click โSaveโ (Figure 33). If you need to go back to the page with the list of handlers during creating a handler, press the Esc key on your keyboard.
Policy editing#
Static policy editing#
The general view of the static policy editing form is shown below (Figure 34).
Description of the parameters of the static policy editing form is given below (Table 8-18).
Table 8. Parameters of the static policy editing form: general parameters and determined attributes
Parameter |
Description |
Default value |
---|---|---|
General |
||
Policy name |
Specifies the name that will be displayed in the list of other policies. Required field |
- |
Determined attributes |
||
Detect face |
Face detection in photo images When enabled, the "Face descriptor" and "Basic attributes (gender, age)" options become available |
Off |
Face descriptor |
Image processing and creation of a data set in a closed, binary format using a special extraction algorithm. When the attribute is enabled, the options โLabelsโ, โSave descriptor in databaseโ, โSave face to databaseโ, โAttach face to listโ, โSave event in cases where a face was foundโ, and โDisplay event in cases where a face was foundโ become available |
On |
Basic attributes (gender, age) |
Assessment of the basic attributes of a person in the image. On When the attribute is enabled, the โSave ifโ and "Call only in cases" options become available |
|
Head position |
Assessment of the head position (angles of inclination and rotation of the head left/right and up/down). When the attribute is enabled, the options โDiscard face images with head rotation/tilt angle aboveโ become available |
On |
Emotion |
Determination of the dominant emotion (anger, disgust, fear, happiness, neutral, sadness, surprise) |
Off |
Mask |
Assessment of the presence or absence of a medical mask or mouth covering. When the attribute is enabled, the filter โProcess images only if detectedโ becomes available |
Off |
Image quality |
Determination of quality (the presence of overexposure, blurring, underexposure, the presence of glare on the face, uneven lighting) |
On |
Eye direction |
Assessment of the direction of a person's gaze in the image |
Off |
Presence of glasses |
Assessment of the presence of glasses and their type (glasses, sunglasses, no glasses) |
Off |
Eye status |
Evaluating whether a person's eyes are open or closed in the image, as well as determining key points of the irises of the eyes |
Off |
Mouth status |
Closed or occluded mouth detection and smile detection |
Off |
Position of 68 feature points of the face |
Determination of 68 feature points of the face (requires additional time for calculations, it is used to determine emotions, eye direction or Liveness check) |
Off |
EXIF metadata |
Defining image metadata |
Off |
Table 9. Parameters of the static policy editing form: Liveness check
Parameter |
Description |
Default value |
---|---|---|
Perform Liveness check |
Enabling Liveness check |
Off |
Liveness threshold |
Ignoring images with a Liveness score below the specified value. Possible values: from 0 to 1. Available only if the โPerform Liveness checkโ attribute is enabled |
0,5 |
Table 10. Parameters of the static policy editing form: Deepfake check
Parameter |
Description |
Default value |
---|---|---|
Perform Deepfake check [^2] |
Determination of digital manipulations for replace one person's likeness convincingly with that of another |
Off |
Deepfake threshold |
Ignoring images with a Deepfake score below the specified value. Possible values: from 0 to 1, where 1 is a real person, 0 - fake |
0.5 |
Use specified Deepfake mode |
Possible values:
The choice of mode determines what set of neural networks perform photo processing for deepfake checking. For more information about the neural networks used in deepfake verification modes, contact VisionLabs technical support. |
Mode 2 |
[^2]: Deepfake license required. Deepfake check is not performed on normalized (centered and cropped) images after face detection.
Table 11. Parameters of the static policy editing form: image quality check
Parameter |
Description |
Default value |
---|---|---|
Perform face image quality check |
Enabling quality check |
|
Image format |
Must be saved in .jpeg or .png format (correct verification). Possible values:
|
JPEG; PNG JPEG2000; |
Image size in Mb |
This assessment determines the size of the image in bytes. It also compares the estimated value with the specified threshold |
5120: 2097152 |
Image width in pixels |
This assessment determines the width of the image in pixels. It also compares the estimatedvalues with thresholds (according to ISO or custom thresholds) |
180:1920 |
Image height in pixels |
This assessment determines the width of the image in pixels. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
180:1080 |
Image aspect ratio |
This assessment determines the proportional ratio of the image width to height. It also compares the estimated value with the specified threshold |
0.74:0,8 |
Degree of illumination uniformity |
It is possible to evaluate the uniformity of illumination according to the requirements specified in the ICAO standard. It also compares the estimated value with the specified threshold (correct verification) |
0.3:1 |
Degree of image specularity |
Bright light artifacts and flash reflection from glasses are not allowed (indirect verification) |
0.3:1 |
Degree of image blureness |
The pixel colors of front-type photo images must be represented in the 24-bit RGB color space, in which each pixel has 8 bits for each color component: red, green, and blue (indirect verification) |
0.61:1 |
Degree of absence of underexposure in the photo |
An underexposure assessment is available. It also compares the estimated value with the specified threshold |
0.5:1 |
Degree of absence of overexposure in the photo |
Too much exposure assessment is available. It also compares the estimated value with the specified threshold |
0.57:1 |
Face illumination uniformity |
It is possible to evaluate the uniformity of illumination according to requirements specified in ICAO standard. The face should be evenly lit so that there are no shadows or glare on the face image. It also compares the estimated value with the specified threshold (correct verification) |
0.5:1 |
Skin tone dynamic range |
This assessment is a determination of the ratio of the brightness of the lightest and darkest areas of the face according to the requirements specified in the ICAO standard. It also compares the estimated value with the specified threshold (correct verification) |
0.5:1 |
Degree of uniformity of the background |
This assessment determines the degree of background uniformity from 0 to 1, where:
|
0.5:1 |
Degree of lightness of the background |
This rating determines the degree of background brightness from 0 to 1, where:
|
0.5:1 |
Presence of radial distortion (Fisheye effect) |
Possible values:
|
No |
Type of image color based on face |
Possible values: Color; Grayscale; Infraredโnear-infrared |
Color |
Shoulders position |
This assessment determines the position of the shoulders if they are in the frame: Parallel Non-parallel Hidden |
Parallel |
Face width in pixels |
This assessment determines the width of the face in pixels. It also compares the estimated value with the specified threshold |
180:1920 |
Face height in pixels |
This assessment determines the height of the face in pixels. It also compares the estimated value with the specified threshold |
180:1080 |
Face offset from the top edge of the image in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification) |
20:50 |
Face offset from the bottom edge of the image in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification) |
20:50 |
Face offset from the left edge of the image in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification) |
20:50 |
Face offset from the right edge of the image in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification) |
20:50 |
Head yaw angle |
Head rotation should be no more than 5ยฐ from the frontal position (correct verification) |
-5:5 |
Head pitch angle |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification). The tilt of the head should be no more than 5ยฐ from the frontal position (correct verification) |
-5:5 |
Head roll angle |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification). The inclination of the head should be no more than 8ยฐ from the frontal position (correct verification) |
-8:8 |
Gaze yaw angle |
This assessment determines the direction of gaze (yaw) |
-5:5 |
Gaze pitch angle |
This assessment determines the direction of gaze (pitch) |
-5:5 |
Probability of smile presence |
The facial expression must be neutral (indirect verification). |
0:0.5 |
Probability of mouth occlusion |
It is not allowed to cover the face with hair or foreign objects along the entire width, from the eyebrows to the lower lip (indirect verification) |
0:0.5 |
Probability of open mouth presence |
This assessment determines the state of the mouth The mouth is closed (correct verification) |
0:0.5 |
Smile properties |
This assessment determines the state of the mouth The facial expression must be neutral (indirect verification). Possible values: Noneโsmile is not found; Smile with closed mouth; Smile with teeth |
None |
Glasses |
Sun glasses are not allowed (correct verification). Possible values: Sunglasses; Eyeglasses; No glasses |
No glasses |
Left eye status |
Both eyes are open normally for the respective subject (considering behavioral factors and/or medical conditions, correct verification). It is not allowed to cover the face with hair or foreign objects along the entire width, from the eyebrows to the lower lip (indirect verification) Possible values: Open; Closed; Occluded |
Open |
Right eye status |
Both eyes are open normally for the respective subject (considering behavioral factors and/or medical conditions, correct verification). It is not allowed to cover the face with hair or foreign objects along the entire width, from the eyebrows to the lower lip (indirect verification). Possible values: Open; Closed; Occluded |
Open |
Red eyes effect presence |
Possible values: Noโthere is no red-eye effect; Yesโthere is a red-eye effect |
No |
Distance between eye centers in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any),the top point of the forehead area and the chin (correct verification) The distance between the centers of the eyes must be at least 120 pixels or at least 45 pixels in accordance with paragraph 12 of the procedure for placing and updating biometric personal data in a unified biometric system (correct verification) |
90:100 |
Horizontal head size relative to image size |
This assessment determines the horizontal head size relative to the image size. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
0.5:75 |
Vertical head size relative to image size |
This assessment determines the vertical head size relative to the image size. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
0.6:0.9 |
The position of the center point of the face horizontally relative to the image |
This assessment determines the horizontal position of the center point relative to the image. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
0.45:0.55 |
The position of the center point of the face vertically relative to the image |
This assessment determines the vertical position of the center point relative to the image. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
0.3:0.5 |
Eyebrows state |
The facial expression must be neutral (indirect verification). Possible values: Neutral; Raised; Squnting; Frowning |
Neutral |
Headwear type |
Possible values: None; Baseball_cap; Beanie; Peaked_cap; Shawl; Hat with earflaps; Helmet; Hood; Hat; Other |
None |
Presence of natural lighting |
The face should be evenly lit so that there are no shadows or glare on the face image (correct verification) Possible values: Noโthe lighting is unnatural; Yesโthe lighting is natural |
Yes |
Table 12. Parameters of the static policy editing form: face occlusion check
Parameter |
Description |
Default value |
---|---|---|
Perform face occlusion check |
Detect presence of occlusion areas of the face |
Off |
Reject face images with occlusion of the following parts of the face |
Select which areas of the face should not be covered in frames from the video stream for the system to accept the image: face, forehead, eyes, nose, mouth, lower part of the face |
|
Acceptable hair occlusion |
Set the degree of acceptable face occlusion with hair from 0 to 1 |
0.15 |
Acceptable Face occlusion |
Set the degree of acceptable face occlusion from 0 to 1 |
0.07 |
Acceptable forehead occlusion |
Set the degree of acceptable forehead occlusion from 0 to 1 |
0.2 |
Acceptable eye occlusion |
Set the degree of acceptable eye occlusion from 0 to 1 |
0.15 |
Acceptable nose occlusion |
Set the degree of acceptable nose occlusion from 0 to 1 |
0.2 |
Acceptable mouth occlusion |
Set the degree of acceptable mouth occlusion from 0 to 1 |
0.15 |
Acceptable lower Face occlusion |
Set the degree of acceptable lower face occlusion from 0 to 1 |
0.2 |
Table 13. Parameters of the static policy editing form: determined body attributes
Parameter |
*Description * |
Default value |
---|---|---|
Detect body |
Face detection in photo images |
Off |
Body descriptor |
Image processing and creation of a data set in a closed, binary format using a special extraction algorithm. When the attribute is enabled, the โLabelsโ option becomes available |
On |
Body basic attributes |
Gender and age estimation based on body silhouette |
|
Upper body attributes based on body silhouette |
Estimation of headwear, upper body clothing color, and sleeve length |
|
Lower body attributes based on body silhouette |
Estimation of lower body clothing type and shoe color |
|
Accessories |
Estimation of the presence or absence of a backpack |
Table 14. Parameters of the static policy editing form: estimation of people count
Parameter |
Description |
Default value |
---|---|---|
Estimate people count |
Counts the number of people in the frame |
Off |
Determine the coordinates of people's bounding box |
Getting X and Y coordinates of people |
Enabled |
Table 15. Parameters of the static policy editing form: filters
Parameter |
Description |
Default value |
---|---|---|
Filters |
||
Discard images with multiple faces |
Determination of images containing multiple faces. Possible values: Select only one face of the best qualityโprocess an image containing several faces, but detect only a face of the best quality; Do not discardโdetect all faces in the image; Discardโignore an image containing multiple faces |
Do not discard |
Reject descriptors with quality below the specified threshold |
Ignoring low quality images. To use the filter, you must enable the determination of the descriptor in the determined attributes |
0,5 |
Process images only if detected |
Possible values: Missingโthe event is created when there is no overlap of the face by the medical mask (no mask); Occludedโthe event is created in case of detection of face overlapping; Medical maskโthe event is created when a medical mask is detected on the face. Several filter values can be specified. Available only when defining the โMedical maskโ attribute |
- |
Discard face images with head rotation angle (to the left or right, yaw) above |
Ignoring images in which the person's head is turned to the left or right at a too large angle โno information will be extracted when detecting a face and evaluating the angle of head rotation. Available only if the โHead positionโ attribute is enabled |
30 |
Discard face images with head tilt angle (to the left or right, roll) above |
Ignoring images in which a person's head is tilted to the left or right at a too large angleโno information will be extracted during face detection and head tilt evaluating. Available only if the โHead positionโ attribute is enabled |
40 |
Discard face images with head tilt angle (up or down, pitch) above |
Ignoring images in which the personโs head is tilted up or down at a too large angle โno information will be extracted during face detection and head tilt evaluating. Available only if the โHead positionโ attribute is enabled |
30 |
Process images of faces only with Liveness states |
Processing images with Liveness status: Spoofโthe absence of a โliveโ person in the frame; Realโthe presence of a โliveโ person in the frame; Unknown. Available only if the โPerform Liveness checkโ attribute is enabled |
- |
Process images of faces only with Deepfake states |
Processing images with Deepfake status: Fakeโthe absence of a โliveโ person in the frame; Realโthe presence of a โliveโ person in the frame. Available only if the Perform Deepfake checkโ attribute is enabled |
- |
Filter images based on face image quality assessment results |
Filter images according to the parameters set in the "Perform face image quality assessment" setting that comply with ISO/IEC 19794-5:2011 and ICAO Available only when the parameter โPerform face image quality assessment*โ is enabled |
Off |
Table 16. Parameters of the static policy editing form: labels
Parameter |
Description |
Default value |
---|---|---|
Labels |
||
Label name |
Specify the name that will be displayed in the policy settings, including the parameters for creating and saving an image/descriptor/event/face, adding a tag |
- |
Identify among |
Searching for a detected person for identification among those created in the database:
|
Faces |
Search for a descriptor |
Among the events created in the database, search for a descriptor:
Only for "Identify among events" |
Faces |
Perform search by |
|
- |
|
||
Each filled field imposes a search restrictionโthe comparison will be successful only if all the search conditions are met |
||
Location (only for โIdentify among eventsโ) |
โDistrictโ; โAreaโ; โCityโ; โStreetโ; โHouse numberโ; โLongitude (-180โฆ180)โ; โAccuracy (0โฆ90)โ; โLatitude (-90โฆ90)โ; โAccuracy (0โฆ90)โ |
- |
Filter search result by |
โGenderโโspecifies the gender for which the face comparison is performed; โAge categoryโโspecifies the lower and/or upper limits of the age of the face is indicated for comparison; โLivenessโโspecifies Liveness state (Spoof, Real or Unknown) |
- |
Additional search parameters |
โThe maximum number of similar ones in the search resultsโ; โAccuracy thresholdโโa value from 0 to 1 |
- |
Table 17. Parameters of the policy editing form: save parameters
Parameter |
Description |
Default value |
---|---|---|
Save parameters |
||
Save face sample |
If enabled, images are saved unconditionally in the database. For selective saving, you must specify: โSave ifโ:
โ โSave face image in cases where a face was foundโ:
โ Face sample storage timeโspecifies the lifetime of objects, after which the samples will be deleted. Possible values:
|
On |
Save body sample |
Saving the event without creating a face in the LUNA PLATFORM 5 database. If enabled, images are saved unconditionally in the database. For selective saving, you must specify: โSave ifโ:
โ โSave body image in cases where a body was foundโ:
โ Body sample storage timeโspecifies the lifetime of objects, after which the samples will be deleted. Possible values:
|
On |
Save face attributes in database |
Saving face attributes in the LUNA PLATFORM 5 database. If enabled, the unconditional saving of face attributes in the database is performed. For selective saving, specify the parameters (for more information see description of โSave face sample โ parameter). โ โFaceAttributes storage timeโโindicates the time in seconds after which the descriptor will be deleted from the database |
Off - |
Save original image in database |
Saving the original image in the LUNA PLATFORM 5 database. For selective saving, specify the parameters (for more information see description of โSave face sample โ parameter) โ โUse external link as original image URLโ if enabled, the link to the external image is stored in the address of the original image, thus avoiding image duplication in the database. If a biometric sample was sent in the request and it was stored in the Image Store, then the link to it will be indicated in the address of the original image. โ Origin image storage timeโspecifies the lifetime of objects, after which the samples will be deleted. Possible values:
|
Off Off |
Save face in database |
Saving the face detected in the image in the LUNA PLATFORM 5 database with the creation of a face in the database. Saving is possible only when the option โSave descriptor in databaseโ is enabled. If enabled, the unconditional saving of the descriptor in the database is performed For selective saving, specify the parameters (for more information see description of โSave face sample โ parameter) โ "Attach face to list"โadds the saved face to the control list or lists in LUNA PLATFORM 5. Possible only if the option โSave descriptor in the databaseโ is enabled. For selective saving, specify the parameters (for more information see description of โSave face sample โ parameter) |
Off Off |
Save event in database |
Saving the detection/identification event in the LUNA PLATFORM 5 database. If enabled, all events are stored unconditionally in the database. For selective saving,specify the parameters (for more information see description of โSave face sample โ parameter) |
On |
Receive and display an event in the โLast eventsโ section |
Displaying an event in the โLast eventsโ section. For selective displaying of events, specify the parameters (for more information see description of โSave image in databaseโ parameter) |
On |
Table 18. Parameters of the policy editing form: tagging parameters
Parameter |
Description |
---|---|
Tagging parameters |
|
Tag name |
Assigning a tag of the given name when conditions are met. In the absence of parameter specifications, the assignment is unconditional. |
Save if |
โGenderโโthe gender of the face in the image matches the specified; โAge categoryโโthe age of the face in the image matches the specified limits; โLivenessโโspecifies Liveness state (Spoof, Real or Unknown); โDeepfakeโโspecifies Deepfake check (Fake or Real) |
Add a tag for each case where a face was found |
โLabelsโโthe list of labels, specifies the names of labels; โWith precisionโโthe lower and/or upper limit of the satisfaction of the comparison result with the parameters specified in the comparison (from 0 to 1) |
Table 19. Parameters of the policy editing form: callbacks
Callbacks allows you to send generated events (notifications) to the third-party system at the specified URL. A mechanism for notifications is based on the principles of HTTP webhooks. They provide asynchronous interaction between systems, allowing external services to react to the emergence of events.
Parameter |
Description |
Default value |
---|---|---|
Add callback |
||
Type |
Protocol type when creating a notification |
HTTP |
URL |
Address of the external system where the notification will be sent |
- |
Authorization type |
Selecting the type of authorization into an external system and setting up authorization data. The basic type of authorization requires specifying login and password to enter an external system |
Basic |
Timeout (seconds) |
Maximum time to wait for a request to complete |
60 |
Request body format |
Data interchange format: JSON or MessagePack |
application/json |
HTTP Headers |
HTTP Request Headers |
- |
Call only in cases where |
Conditions for sending notification Activated when determination of basic attributes (gender, age) is enabled, see the Table 9 โ Gender:
โ Age category:
Activated when Liveness check is enabled, see the Table 10 โ Liveness:
Activated when Deepfake check is enabled, see the Table 11 โ Deepfake:
|
|
Call only in cases where a person or body has been found |
|
Adding a label#
To create a label, click on in policy editing form (Figure 35).
If you need to identify faces among other faces in the label, then select โFacesโ for โIdentify amongโ field in the window for the parameter adding (Figure 36). If you need to identify among events, select โEventsโ for โIdentify amongโ field (Figure 37).
Fill in all the required parameters and click on the โAddโ button at the bottom of the form.
Label editing#
Editing of the label is performed by clicking on the button in the line with the label name.
A general view of the form for editing the label is shown below (Figure 38).
Edit the parameter values and click on the โChangeโ button.
Label deleting#
Deletion of the label is performed by clicking on the button in the line with the label name.
Tag adding#
To create a tag, click on in policy editing form (Figure 39).
A general view of the form for creating a tag is shown below (Figure 40).
Fill in all the required parameters and click on the โAddโ button at the bottom of the form.
Tag editing#
Tag editing is performed by clicking on the button in the line with the tag name.
A general view of the tag editing form is shown below (Figure 41).
Edit the values of the tag parameters and click on the โChangeโ button.
Tag deleting#
Deletion of the tag is performed by clicking on the button in the line with the tag name.
After finishing editing the policy, click on the โSaveโ button in the upper right corner (Figure 38).
Dynamic policy editing#
To edit a dynamic policy, first click on the button on the page with a list of policies (1 in Figure 32. Then in the editing form change the name of the policy and click โSaveโ (Figure 42).
Policy deleting#
Deleting a policy is performed by clicking on the button in the line (2 in the Figure 32).
Confirm the action in the pop-up windowโclick on the โDeleteโ button or cancel the action by clicking on the โCancelโ button (Figure 43). After successful deletion, a notification will appear.
"Verification" section#
The โVerificationโ section is intended for creating, deleting, testing verifiers, and editing their parameters. Verifiers are used to quickly compare two faces by face photo image and Face ID, external ID, attribute, event, and display the result of the test (Figure 44).
โVerificationโ section contains the following elements:
- table of verifiers:
- โDescriptionโโverifier name;
- โVerifier IDโโverifier identifier;
โbutton for testing the verifier (1);
โbutton for editing verifier parameters (2);
โbutton for deleting the verifier (3);
- โAddโ buttonโbutton for creating a verifier;
- the number of verifiers displayed on the page is set by the switch in the lower right corner of the page. There can be 10, 25, 50 or 100 verifiers in total on one page (4).
Verifier creation#
To create a verifier, click on the โAddโ button (Figure 44). A form for step by step verifier creation will open (Figure 45).
Fill in all the required parameters and click on the โNextโ button to proceed to the next step. The description of the parameters is presented below, in the "Verifier editing" section.
After setting all the parameters, a window with a message about the successful verifier creation appears (Figure 46). Click anywhere outside the successful verifier creation message to navigate to the โVerificationโ section.
Verifier testing#
Testing a verifier is performed by clicking on the button in the line (1 in Figure 44).
The general view of the โVerifier testingโ form is shown below (Figure 47).
โVerifier testingโ form contains the following blocks:
- โSearch byโโsearch options:
- โFaceโโsearch by registered face in the system:
- โFace IDโโface identifier that is created in the LUNA PLATFORM 5 system as a result of a detection event and attribute extraction;;
- โExternal IDโโsearch by external face identifier:
- โExternal IDโโexternal face identifier;
- โAttributeโโsearch by face attribute:
- โAttribute IDโโattribute (descriptor) identifier;
- โEventโโsearch by registered event in the system:
- โEvent IDโโidentifier of the event of detection and attribute extraction;
- Photo imageโsearch by uploaded photo image:
- field for uploading a photo image;
- Searching results:
- โPhotoโโsample of detected face (candidate);
- โSimilarity, %โโsimilarity value, in percent;
- โStatusโโverification result:
โsuccessful verification;
โunsuccessful verification;
- โLink to the referenceโ
โgo to the page the reference face;
โbutton for downloading the result of the verification (Figure 48).
To test the verifier by face, in the โSearch byโ block enter the Face ID and select photo image, click on or โSelect fileโ, and specify the path to the image file.
Image file requirements:
- *.jpeg, *.png or *.bmp format;
- image size no less than 320x250 and no more than 3840x2160 pixels;
- image may contain one or more people;
- image must have a person's face.
When loading a photo image containing many faces, the Service verifies all faces in the image.
To reset the image, click on .
Verifier editing#
Editing of verifier parameters is performed by clicking the button in the line (2 in the Figure 44). The general view of the verifier editing form is shown below (Figure 49).
Description of the parameters of the verifier editing form is given below (Table 19-24).
Table 19. Parameters of the verifier editing form: general parameters and determined attributes
Parameter |
Description |
Default value |
---|---|---|
General |
||
Verifier name |
Specifies the name that will be displayed in the list of verifiers |
- |
Similarity threshold |
Specifies a similarity score, which will consider that the reference and the candidate contain the face of the same person |
0.93 |
Determined attributes |
||
Basic attributes (gender, age) |
Assessment of the basic attributes of a person in the image. On When the attribute is enabled, the โSave ifโ and "Call only in cases" options become available |
|
Head position |
Assessment of the head position (angles of inclination and rotation of the head left/right and up/down). When the attribute is enabled, the options โDiscard face images with head rotation/tilt angle aboveโ become available |
On |
Emotion |
Determination of the dominant emotion (anger, disgust, fear, happiness, neutral, sadness, surprise) |
Off |
Mask |
Assessment of the presence or absence of a medical mask or mouth covering. When the attribute is enabled, the filter โProcess images only if detectedโ becomes available |
Off |
Image quality |
Determination of quality (the presence of overexposure, blurring, underexposure, the presence of glare on the face, uneven lighting) |
On |
Eye direction |
Assessment of the direction of a person's gaze in the image |
Off |
Presence of glasses |
Assessment of the presence of glasses and their type (glasses, sunglasses, no glasses) |
Off |
Eye status |
Evaluating whether a person's eyes are open or closed in the image, as well as determining key points of the irises of the eyes |
Off |
Mouth status |
Closed or occluded mouth detection and smile detection |
Off |
Position of 68 feature points of the face |
Determination of 68 feature points of the face (requires additional time for calculations, it is used to determine emotions, eye direction or Liveness check) |
Off |
EXIF metadata |
Defining image metadata |
Off |
Table 20. Parameters of the verifier editing form: Liveness check
Parameter |
Description |
Default value |
---|---|---|
Perform Liveness check |
Enabling Liveness check |
Off |
Liveness threshold |
Ignoring images with a Liveness score below the specified value. Possible values: from 0 to 1. Available only if the โPerform Liveness checkโ attribute is enabled |
0,5 |
Table 21. Parameters of the verifier editing form: Deepfake check
Parameter |
Description |
Default value |
---|---|---|
Perform Deepfake check [^2] |
Determination of digital manipulations for replace one person's likeness convincingly with that of another |
Off |
Deepfake threshold |
Ignoring images with a Deepfake score below the specified value. Possible values: from 0 to 1, where 1 is a real person, 0 - fake |
0.5 |
Use specified Deepfake mode |
Possible values:
The choice of mode determines what set of neural networks perform photo processing for deepfake checking. For more information about the neural networks used in deepfake verification modes, contact VisionLabs technical support. |
Mode 2 |
[^2]: Deepfake license required. Deepfake check is not performed on normalized (centered and cropped) images after face detection.
Table 22. Parameters of the verifier editing form: image quality check
Parameter |
Description |
Default value |
---|---|---|
Perform face image quality check |
Enabling quality check |
|
Image format |
Must be saved in .jpeg or .png format (correct verification). Possible values:
|
JPEG; PNG JPEG2000; |
Image size in Mb |
This assessment determines the size of the image in bytes. It also compares the estimated value with the specified threshold |
5120: 2097152 |
Image width in pixels |
This assessment determines the width of the image in pixels. It also compares the estimatedvalues with thresholds (according to ISO or custom thresholds) |
180:1920 |
Image height in pixels |
This assessment determines the width of the image in pixels. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
180:1080 |
Image aspect ratio |
This assessment determines the proportional ratio of the image width to height. It also compares the estimated value with the specified threshold |
0.74:0,8 |
Degree of illumination uniformity |
It is possible to evaluate the uniformity of illumination according to the requirements specified in the ICAO standard. It also compares the estimated value with the specified threshold (correct verification) |
0.3:1 |
Degree of image specularity |
Bright light artifacts and flash reflection from glasses are not allowed (indirect verification) |
0.3:1 |
Degree of image blureness |
The pixel colors of front-type photo images must be represented in the 24-bit RGB color space, in which each pixel has 8 bits for each color component: red, green, and blue (indirect verification) |
0.61:1 |
Degree of absence of underexposure in the photo |
An underexposure assessment is available. It also compares the estimated value with the specified threshold |
0.5:1 |
Degree of absence of overexposure in the photo |
Too much exposure assessment is available. It also compares the estimated value with the specified threshold |
0.57:1 |
Face illumination uniformity |
It is possible to evaluate the uniformity of illumination according to requirements specified in ICAO standard. The face should be evenly lit so that there are no shadows or glare on the face image. It also compares the estimated value with the specified threshold (correct verification) |
0.5:1 |
Skin tone dynamic range |
This assessment is a determination of the ratio of the brightness of the lightest and darkest areas of the face according to the requirements specified in the ICAO standard. It also compares the estimated value with the specified threshold (correct verification) |
0.5:1 |
Degree of uniformity of the background |
This assessment determines the degree of background uniformity from 0 to 1, where:
|
0.5:1 |
Degree of lightness of the background |
This rating determines the degree of background brightness from 0 to 1, where:
|
0.5:1 |
Presence of radial distortion (Fisheye effect) |
Possible values:
|
No |
Type of image color based on face |
Possible values: Color; Grayscale; Infraredโnear-infrared |
Color |
Shoulders position |
This assessment determines the position of the shoulders if they are in the frame: Parallel Non-parallel Hidden |
Parallel |
Face width in pixels |
This assessment determines the width of the face in pixels. It also compares the estimated value with the specified threshold |
180:1920 |
Face height in pixels |
This assessment determines the height of the face in pixels. It also compares the estimated value with the specified threshold |
180:1080 |
Face offset from the top edge of the image in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification) |
20:50 |
Face offset from the bottom edge of the image in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification) |
20:50 |
Face offset from the left edge of the image in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification) |
20:50 |
Face offset from the right edge of the image in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification) |
20:50 |
Head yaw angle |
Head rotation should be no more than 5ยฐ from the frontal position (correct verification) |
-5:5 |
Head pitch angle |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification). The tilt of the head should be no more than 5ยฐ from the frontal position (correct verification) |
-5:5 |
Head roll angle |
The image must contain a full front view of the person's head, including the left and right ear (if person has any), the top point of the forehead area and the chin (correct verification). The inclination of the head should be no more than 8ยฐ from the frontal position (correct verification) |
-8:8 |
Gaze yaw angle |
This assessment determines the direction of gaze (yaw) |
-5:5 |
Gaze pitch angle |
This assessment determines the direction of gaze (pitch) |
-5:5 |
Probability of smile presence |
The facial expression must be neutral (indirect verification). |
0:0.5 |
Probability of mouth occlusion |
It is not allowed to cover the face with hair or foreign objects along the entire width, from the eyebrows to the lower lip (indirect verification) |
0:0.5 |
Probability of open mouth presence |
This assessment determines the state of the mouth The mouth is closed (correct verification) |
0:0.5 |
Smile properties |
This assessment determines the state of the mouth The facial expression must be neutral (indirect verification). Possible values: Noneโsmile is not found; Smile with closed mouth; Smile with teeth |
None |
Glasses |
Sun glasses are not allowed (correct verification). Possible values: Sunglasses; Eyeglasses; No glasses |
No glasses |
Left eye status |
Both eyes are open normally for the respective subject (considering behavioral factors and/or medical conditions, correct verification). It is not allowed to cover the face with hair or foreign objects along the entire width, from the eyebrows to the lower lip (indirect verification) Possible values: Open; Closed; Occluded |
Open |
Right eye status |
Both eyes are open normally for the respective subject (considering behavioral factors and/or medical conditions, correct verification). It is not allowed to cover the face with hair or foreign objects along the entire width, from the eyebrows to the lower lip (indirect verification). Possible values: Open; Closed; Occluded |
Open |
Red eyes effect presence |
Possible values: Noโthere is no red-eye effect; Yesโthere is a red-eye effect |
No |
Distance between eye centers in pixels |
The image must contain a full front view of the person's head, including the left and right ear (if person has any),the top point of the forehead area and the chin (correct verification) The distance between the centers of the eyes must be at least 120 pixels or at least 45 pixels in accordance with paragraph 12 of the procedure for placing and updating biometric personal data in a unified biometric system (correct verification) |
90:100 |
Horizontal head size relative to image size |
This assessment determines the horizontal head size relative to the image size. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
0.5:75 |
Vertical head size relative to image size |
This assessment determines the vertical head size relative to the image size. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
0.6:0.9 |
The position of the center point of the face horizontally relative to the image |
This assessment determines the horizontal position of the center point relative to the image. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
0.45:0.55 |
The position of the center point of the face vertically relative to the image |
This assessment determines the vertical position of the center point relative to the image. It also compares the estimated values with thresholds (according to ISO or custom thresholds) |
0.3:0.5 |
Eyebrows state |
The facial expression must be neutral (indirect verification). Possible values: Neutral; Raised; Squnting; Frowning |
Neutral |
Headwear type |
Possible values: None; Baseball_cap; Beanie; Peaked_cap; Shawl; Hat with earflaps; Helmet; Hood; Hat; Other |
None |
Presence of natural lighting |
The face should be evenly lit so that there are no shadows or glare on the face image (correct verification) Possible values: Noโthe lighting is unnatural; Yesโthe lighting is natural |
Yes |
Table 23. Parameters of the verifier editing form: face occlusion check
Parameter |
Description |
Default value |
---|---|---|
Perform face occlusion check |
Detect presence of occlusion areas of the face |
Off |
Reject face images with occlusion of the following parts of the face |
Select which areas of the face should not be covered in frames from the video stream for the system to accept the image: face, forehead, eyes, nose, mouth, lower part of the face |
|
Acceptable hair occlusion |
Set the degree of acceptable face occlusion with hair from 0 to 1 |
0.15 |
Acceptable Face occlusion |
Set the degree of acceptable face occlusion from 0 to 1 |
0.07 |
Acceptable forehead occlusion |
Set the degree of acceptable forehead occlusion from 0 to 1 |
0.2 |
Acceptable eye occlusion |
Set the degree of acceptable eye occlusion from 0 to 1 |
0.15 |
Acceptable nose occlusion |
Set the degree of acceptable nose occlusion from 0 to 1 |
0.2 |
Acceptable mouth occlusion |
Set the degree of acceptable mouth occlusion from 0 to 1 |
0.15 |
Acceptable lower Face occlusion |
Set the degree of acceptable lower face occlusion from 0 to 1 |
0.2 |
Table 24. Parameters of the verifier editing form: save parameters
Parameter |
Description |
Default value |
---|---|---|
Save parameters |
||
Save image in database |
Saving the image in the LUNA PLATFORM 5 database. If enabled, the unconditional saving of images in the database is performed. |
Off |
Save biometric template in database |
Saving the created biometric template in the LUNA PLATFORM 5 database. If enabled, the unconditional saving of biometric templates in the database is performed. |
Off |
Table 25. Parameters of the verifier editing form: filters
Parameter |
Description |
Default value |
---|---|---|
Filters |
||
Discard images with multiple faces |
Determination of images containing multiple faces. Possible values: Select only one face of the best qualityโprocess an image containing several faces, but detect only a face of the best quality; Do not discardโdetect all faces in the image; Discardโignore an image containing multiple faces |
Do not discard |
Reject descriptors with quality below the specified threshold |
Ignoring low quality images. To use the filter, you must enable the determination of the descriptor in the determined attributes |
0,5 |
Process images only if detected |
Possible values: Missingโthe event is created when there is no overlap of the face by the medical mask (no mask); Occludedโthe event is created in case of detection of face overlapping; Medical maskโthe event is created when a medical mask is detected on the face. Several filter values can be specified. Available only when defining the โMedical maskโ attribute |
- |
Discard face images with head rotation angle (to the left or right, yaw) above |
Ignoring images in which the person's head is turned to the left or right at a too large angle โno information will be extracted when detecting a face and evaluating the angle of head rotation. Available only if the โHead positionโ attribute is enabled |
30 |
Discard face images with head tilt angle (to the left or right, roll) above |
Ignoring images in which a person's head is tilted to the left or right at a too large angleโno information will be extracted during face detection and head tilt evaluating. Available only if the โHead positionโ attribute is enabled |
40 |
Discard face images with head tilt angle (up or down, pitch) above |
Ignoring images in which the personโs head is tilted up or down at a too large angle โno information will be extracted during face detection and head tilt evaluating. Available only if the โHead positionโ attribute is enabled |
30 |
Process images of faces only with Liveness states |
Processing images with Liveness status: Spoofโthe absence of a โliveโ person in the frame; Realโthe presence of a โliveโ person in the frame; Unknown. Available only if the โPerform Liveness checkโ attribute is enabled |
- |
Process images of faces only with Deepfake states |
Processing images with Deepfake status: Fakeโthe absence of a โliveโ person in the frame; Realโthe presence of a โliveโ person in the frame. Available only if the Perform Deepfake checkโ attribute is enabled |
- |
Filter images based on face image quality assessment results |
Filter images according to the parameters set in the "Perform face image quality assessment" setting that comply with ISO/IEC 19794-5:2011 and ICAO Available only when the parameter โPerform face image quality assessment*โ is enabled |
Off |
After finishing editing the verifier, click on the โSaveโ button in the upper right corner.
Verifier deleting#
Deleting a verifier is performed by clicking on the button in the line (3 in the Figure 44).
In the pop-up window (Figure 50), you must confirm the action โ click on the โDeleteโ button or cancel the action by clicking on the โCancelโ button.
"Event sources" section#
The โEvent sourcesโ section is intended to display all event sources and video streams, event source status, preview, group, and parameters settings for each event source (Figure 51).
Sources can be webcams, USB, and IP cameras (via RTSP protocol), video files, and images.
โEvent sourcesโ section contains the following elements:
- buttons to control video stream processing (1):
โ โPlayโ button to start stream processing (sends a request for stream processing, the stream is distributed to a specific FaceStream 5 and it starts processing it);
โ โPauseโ button to pause stream processing, for example, to save resources (pauses the stream processing, but the stream remains assigned to the same FaceStream 5);
โ โStopโ button to stop stream processing (stops the stream processing, the stream is no longer assigned to the same FaceStream 5);
- the list of streams from LUNA Streams and their parameters (2):
- โLast frameโ โ event source video stream preview;
- โStatusโ โ current status (state) of video stream activity;
- โStream IDโ โ video stream identifier in LUNA Streams, generated when stream is created;
- โNameโ โ video stream name;
- โDescriptionโ โ additional user-defined information about the video stream;
- โGroupโ โ the name of the group to which the video stream is attached;
- list of event sources from LUNA Access (3):
- โPreviewโ โ event source video stream preview;
- โNameโ โ event source name;
- buttons for working with streams (4):
โ button to edit video stream parameters;
โ button to view the event source video stream;
โ button to delete video stream;
- โAddโ button โ button for adding a video stream;
- the number of video streams displayed on the page is set by the switch in the lower right corner of the page. There can be 10, 25, 50 or 100 video streams in total on one page (5).
For searching and filtering video streams use the following fields: โStream IDโ, โNameโ, โDescriptionโ and โGroupโ.
LUNA Streams UI is a service that allows user to interact with FaceStream 5 in terms of working with video streams (event sources).
The LUNA Streams interface allows the user to configure video stream parameters such as stream type, stream source address, filtering parameters, etc FaceStream 5 takes settings from LUNA Streams for further processing.
For a detailed description of how FaceStream 5 interacts with LUNA Streams, see the FaceStream 5 documentation Administrator manual.
The following video stream statuses are possible:
- โpendingโ โ the stream has been taken into operation, but no FaceStream 5 handler has been found yet;
- โin_progressโ โ the stream is being processed by the FaceStream 5 handler;
- โdoneโ โ the stream is fully processed (for video files and images);
- โpauseโ โ stream processing has been paused by the user (not applicable to video files and images);
- โrestartโ โ stream processing has been restarted by the server. The status is temporary and can only appear when using the โautorestartโ settings group (for more details, see the FaceStream 5 User manual);
- โcancelโ โ stream processing has been canceled by the user. The "cancel" status may occur after the โpendingโ, โrestartโ and โpauseโ statuses;
- โfailureโ โ Stream processing by FaceStream 5 handler failed (e.g., an error occurred);
- โhandler_lostโ โ FaceStream 5 handler is lost, needs to be transferred to another handler (not applicable for video files and images);
- โnot_foundโ โ the stream was deleted by the user while processing;
- โdeletedโ โ the stream was deleted by the user intentionally.
The โrestartโ and โhandler_lostโ statuses are temporary. With these statuses, it is not possible to get a stream, however, the transition through these statuses is logged as usual.
The โnot_foundโ status is internal and will be sent back for feedback if the stream has been removed during processing. With this status it is not possible to receive a stream.
The โdeletedโ status is virtual. A stream with such a status cannot exist, but this status can be seen in the stream's logs.
See the FaceStream 5 Administrator manual for details on transition from status to status.
Adding a video stream#
To add a video stream, click on the โAddโ button (Figure 51). โCreate streamโ form will open for specifying the settings (Figure 52).
Start specifying the values of the video stream parameters according to the parameters shown in the table (Table 26). If you need to go back to the page with source list while creating a video stream, press the Esc key on your keyboard.
Table 26. Event source settings
Parameter |
Description |
Default value |
---|---|---|
General stream parameters |
||
Account ID |
The parameter is used to bind the received data to a specific user. Required field |
Account ID in LP5 |
Stream name |
The stream name displayed in the Service. Used to identify the source of sent frames. |
|
Description |
User information about the video stream |
|
Group |
The name of the group to which the stream is attached |
|
Status |
The current state of the stream. Possible statuses: Pause Pending |
Pending; |
Stream data |
||
Type |
Video stream transmission type: UDP; TCP; Videofile; Images The TCP protocol implements an error control mechanism that minimizes information loss and missing key frames at the cost of increasing network latency. The UDP protocol does not implement an error control mechanism, so the stream is not protected from damage. The use of this protocol is recommended only if a high quality network infrastructure is available. For a large number of streams (10 or more), it is recommended to use the UDP protocol. When using the TCP protocol, there may be problems with reading streams |
UDP |
Full path to the source |
The path to the source of the video stream. Required field For example, for TCP/UDP type: rtsp://some_stream_address USB device number for TCP/UDP type: /dev/video0 To use a USB device, you must specify the --device flag with the address of the USB device when starting the FaceStream 5 Docker container (see โLaunching keysโ section in the FaceStream 5 Installation manual). The full path to video file for Videofile type: https://127.0.0.1:0000/super_server/ The full path to the directory with images for Images type: /example1/path/to/images/ To use video files and images, you must first transfer them to a Docker container |
|
ROI coordinates |
A limited frame area where face or body is detected and tracked (for
example, in a dense flow of people). Specify the ROI value in one of the
two formatsโ The first two values specify the coordinates of the top left point of the frame. The second two values indicate the width and height of the area of interest,
if the values are specified in Parameter setting can be done visually on the preview image when editing the event source. To do this, click on the gear button. In the opened window, grab the border of the detection area and move it. The width, height and coordinates of the detection area will take on new values. If you need to detect over the entire frame, click the "Full frame" button. Save your changes. |
|
DROI coordinates |
A limited area within the ROI zone. Face detection is performed in the
ROI region, but the best frame is selected only in the DROI region. The face
detection must be completely within the DROI so that the frame is considered as
the best one. Specify the DROI value in one of two formatsโ The first two values specify the coordinates of the top left point of the frame. The second two values indicate the width and height of the area of interest,
if the values are specified in
DROI is recommended for use when working with access control systems. This parameter is used only for working with faces. Parameter setting can be done visually on the preview image when editing the event source. To do this, click on the gear button. In the opened window, grab the border of the best frame selection area and move it. The width, height and coordinates of the area will take on new values. If you need to select the best frame over the entire frame, click the "Full frame" button. Save your changes. |
|
Rotation angle of the image from the source |
Used when the incoming video stream is rotated (for example, if the event source is installed on the ceiling) |
0 |
Frame width |
The parameter is used only for the TCP and UDP types and is designed to work with protocols that imply the existence of several channels with different bit rates and resolutions (e.g., HLS). If the stream has several such channels, then this parameter will allow you to select from all channels of the entire stream the channel which frame width is closer to the value specified in this parameter |
800 |
Endless |
This parameter allows you to control how the stream is restarted when a network error is received. The parameter is available only for the TCP and UDP types. If the parameter takes the Enabled value, then in case of receiving an error and successful reconnection, the stream processing will continue. If all attempts to reconnect failed, then the stream will take the status โfailureโ. If the parameter takes the Disabled value, then the stream processing will not continue, and the stream will take the โdoneโ status. When broadcasting a video file, the Disabled value is assumed. This will avoid re-processing an already processed video file fragment when an error is received. If the parameter value is Enabled when broadcasting a video file, then after the processing is completed, the video file will be processed from the beginning |
Enabled |
Stream handler parameters |
This group of parameters defines the parameters of the policy (handler) created in LP5, which will be used to process streams. Different handlers should be used for faces and bodies. The handler must be created in LP5 beforehand |
|
Handler URL |
The full network path to the deployed LP5 API service, including the LUNA Handlers and LUNA Events services required to generate an event by handler: http:// Where Required field |
|
API version |
API version for event generation in LP5. Required field API version 6 is currently supported |
|
Handler ID for best shots (static) |
The parameter allows using an external static handler_id of the LP5 policy for processing biometric samples of faces or bodies according to the specified rules. When using this policy, LP5 generates an event that contains all the information received from FaceStream 5 and processes it according to the processing rules. For example: aaba1111-2111-4111-a7a7-5caf86621b5a Required field |
|
URL to save original frames |
This parameter specifies the URL for saving original frames of faces or bodies in LP5. The URL can be either the address to the LUNA Image Store service container, or the address to the /images resource of the LUNA API service. When specifying an address to /images resource, original frame will be saved under the image_id identifier. To send a frame, the send_source_frame parameter must be enabled. An example of the address to the LUNA Image Store service container:
Where 127.0.0.1 is IP address where the LUNA Image Store service is deployed; 5020 is
the default port of the LUNA Image Store service; 1 is API version of the LUNA Image
Store service; An example of the address to the /images resource of the LUNA API service: http://127.0.0.1:5000/6/images Where 127.0.0.1 is the IP address where the LUNA API service is deployed; 6 is the API version of the LUNA API service; 5000 is the default API service port |
|
Authorization (Token) |
This parameter specifies either a token or an LP5 account ID for making requests to the LUNA API service. If the authorization field is not filled, then the LP5 account ID will be used, which is set when creating a stream |
|
Geoposition |
This group of parameters includes information about location of video stream source |
|
City Area District Street House number Longitude Latitude |
Event source geographical location |
|
Autorestart |
This group of parameters allows to configure the automatic restart of the stream |
|
Autorestart |
Whether or not to use automatic stream restart |
Enabled |
Attempt count |
Number of attempts to automatically restart the stream |
10 |
Autorestart delay (in seconds) |
Stream auto restart delay |
60 |
Sending parameters |
This group of parameters defines the period during which frames will be analyzed to select the best shot, as well as all parameters associated with compiling a collection of the best shots |
|
Frame analysis period after which the best shot will be sent |
The period starts from the moment a person appears in the frame โ the first detection. Decreasing this parameter allows to quickly determine person, but with a greater error. Possible values: number of frames; number of seconds; 1 โ frames are analyzed for all frames until the end of the track. At the end of the track (when the object leaves the frame) the best shot will be sent to LP5. |
-1 0 |
Wait duration between track analysis periods |
Specifies the timeout between two consecutive tracks. Possible values: number of frames; number of seconds; 0 โ there is no timeout; 1 โ timeout will last indefinitely |
0 |
Track analysis and waiting period duration measure |
Specifies the measurement type of the frame analysis period and the timeout period: Seconds; Frames The choice depends on the business task |
Seconds |
Number of frames that the user sets to receive from the track or certain periods of this track |
Assumes the creation of a collection of the best shots of the track or the time interval of the track, specified in the "Frame analysis period after which the best shot will be sent" parameter. This collection will be sent to LP5. Increasing the value increases the probability of correct recognition of the object, but affects the network load. Possible values are from 1 and more. |
1 |
Send only full set |
Allows to send data (best shots and detections) only if user have the required number of best shots (โNumber of frames that the user sets to receive from the track or certain periods of this trackโ) and the track length (โMinimum detection size for Primary Track modeโ) |
Enabled |
Delete bestshot and detection data |
Allows to delete the best shots and detections after sending data. If disabled, data remains in memory |
Disabled |
Use Primary Track |
This group of parameters is designed to work with access control systems (ACS, turnstiles at the entrances) to simplify control and of face recognition technology at the entrance to a protected area. This group of parameters is only used for working with faces. This group of parameters is not used for the Image type |
|
Use Primary Track |
If the value of this parameter is Enabled, then the implementation mode of the Primary Track is turned on. Of all the detections on the frame, the detection with the maximum size is selected and its track becomes the main one. Further analysis s performed based on this track. The best shot from that track is sent to LP5. When using the parameter at the checkpoint, the best shots of only the person closest to the turnstile will be sent (the condition for the largest detection is met) |
Disabled |
Minimum detection size for Primary Track mode |
Sets the minimum detection size (vertically in pixels) at which the analysis of stream frames begins and the determination of the best frame |
70 |
Size of detection for the main track |
Sets the detection size in pixels for the Primary Track. When the detection size in pixels reaches the specified value, the track immediately sends the best shot to the server |
140 |
Healthcheck parameters |
This parameter group is used only when working with streams (TCP, UDP) and video files. In this group, user can set the parameters for reconnecting to the stream in case of stream playback errors |
|
Maximum number of stream errors to reconnect to the stream |
The maximum number of errors during stream playback. The parameter works in conjunction with the parameters "Error count period duration (in seconds)" and " Time between reconnection attempts (in seconds)". After the first error is received, the timeout specified in the "Time between reconnection attempts (in seconds)" parameter is performed, and then the connection to the stream is retried. If during the time specified in the "Error count period duration (in seconds)" parameter, the number of errors is greater than or equal to the number specified in the parameter, then the processing of the stream will be terminated, and its status will change to "failure". Errors can be caused by a problem with the network or video availability |
10 |
Error count period duration (in seconds) |
Parameter-criterion of the time to reconnect to the video stream. If the maximum number of errors occurs within the specified time, an attempt is made to reconnect to the video stream |
3600 |
Time between reconnection attempts (in seconds) |
After receiving the first error, the timeout specified in the parameter is performed, then the connection to the stream is retried |
5 |
Filtering parameters |
The parameter group describes objects for image filtering and sending the resulting best shots |
|
Threshold value to filter detections |
Also called Approximate Garbage Score (AGS) for faces and Detector score for bodies โ threshold for filtering face or body detections sent to the server. All detections with a score above the value of the parameter can be sent to the server as an HTTP request, otherwise the detections are not considered acceptable for further work with them. The recommended threshold value was identified through research and analysis of detections on various images of faces and bodies |
0,5187 |
Head rotation angle threshold (to the left or right, yaw) |
The maximum value of the angle of rotation of the head to the left and right to the source of the stream (in degrees). If the head rotation angle on the frame is greater than the specified value, the frame is considered unacceptable for further processing. This parameter is used only for working with faces |
40 |
Head tilt angle threshold (up or down, pitch) |
The maximum value of the head tilt angle up and down relative to the source of the stream. If the head tilt angle on the frame is greater than the specified value, then the frame is considered unacceptable for further processing. This parameter is used only for working with faces |
40 |
Head tilt angle threshold (to the left or right, roll) |
The maximum head tilt angle to the left and right relative to the source of the stream. If the head tilt angle on the frame is greater than the specified value, then the frame is considered unacceptable for further processing. This parameter is used only for working with faces |
30 |
Number of frames used to filter photo images by the angle of rotation of the head |
Filtering cuts off images with faces strongly turned away from the stream source. Specifies the number of frames for analyzing head rotation angles on each of these frames. If the angle is drastically different from the group average, the frame will not be considered the best shot. This parameter is used only for working with faces With a value of 1, the parameter is disabled. Recommended value: 7 |
1 |
Number of frames the system must collect to analyze head yaw angle |
The parameter indicates to the system that it is necessary to collect the number of frames specified in the โNumber of frames used to filter photo images by the angle of rotation of the headโ parameter to analyze the head rotation angle. If the parameter is disabled, the Service will sequentially analyze incoming frames, i.e., first, two frames are analyzed, then three, and so on. The maximum number of frames in this sequence is set in "Number of frames used to filter photo images by the angle of rotation of the head". This parameter is used only for working with faces |
Disabled |
Mouth overlap threshold (minimum mouth visibility) |
If the received value exceeds the specified threshold, the image is considered unacceptable for further processing. For example, with the parameter value equals to 0.5, 50% of the mouth area is allowed to be covered. This parameter is used only for working with faces |
0 |
Minimum body detection size |
The parameter specifies the body detection size, less than which it will not be sent for processing. If the value is 0, then body detection will not be filtered by size |
0 |
Liveness parameters |
Liveness is used to check if there is a live person in the frame and prevents a printed photo or a photo from a phone from being used to pass the check. This group of parameters is only used for working with faces. This group of parameters is not used for the Image type |
|
Check RGB ACS Liveness |
Enables the mode of checking the presence of a person in the frame, based on working with the background. The check execution speed depends on the frame size of the video stream. If the processing speed drops with the enabled parameter, you need to reduce the video resolution in the event source settings |
Disabled |
Check FlyingFaces Liveness |
Enables the mode of checking the presence of a person in the frame, based on working with the environment of the face |
Disabled |
Track frames to run liveness check on |
The parameter specifies for which frames of track the Liveness check will be performed. Frame selection options: First N shots; Last N shots before best shot sending; All shots of track. The value โNโ is specified in the parameter โNumber of frames in the track for Liveness check when liveness-mode is enabledโ |
First N frames |
Number of frames in the track for Liveness check when liveness-mode โ is enabled |
The number of frames in a track for checking Liveness when using the parameter โTrack frames to run liveness check on |
0 |
Threshold value at which the system will consider that there is a real person in the frame |
The threshold value at which the Service considers that there is a living person in the frame. The Service verdict on the presence of a real person in the frame will follow only if Liveness returns a value higher than the specified threshold value |
0 |
Livenesses weights (RGB ACS, FlyingFaces) |
The coefficient of influence of each type of Liveness checking on the final estimate of the presence of a living person in the frame. Three values are indicated, referring to different types of Liveness. Values are indicated in fractions of a unit. The ratio is scaled based on the given numbers, regardless of whether they constitute a unit and which Liveness methods are enabled |
0 0 0 |
Number of background frames that are used for the corresponding checks |
Allows to set the number of background frames in the track for the Liveness check. Recommended value: 300. It is not recommended to change this parameter |
0 |
Additional parameters |
||
Frame processing |
The parameter is used only for TCP, UDP and Videofile types. Possible values: Auto; Full frame; Scale frame. The parameter is set for a specific instance of FaceStream 5. With Full frame value, frame is immediately converted to an RGB image of required size after decoding. This results in a better image quality, reduces the frame rate. When set to Scale frame, the image is scaled based on the TrackeEgine settings. The default value is Auto. In this case, one of the two modes is selected automatically |
Auto |
Number of threads for video decoding |
Sets the number of streams for video decoding with FFMPEG. With an increase in the number of streams, the number of processor cores involved in decoding increases. Increasing the number of streams is recommended when processing high- definition video (4K and above) |
0 |
Maximum FPS for video processing |
The parameter is used only for Videofile type. The video is processed at the specified. Video cannot be processed with FPS higherFPS. than specified in this parameter If the video has a high FPS value and FaceStream 5 cannot operate at the specified FPS, then frames will be skipped. Thus, the video file imitates a stream from a real video camera. This can be useful for performance tuning. The video will be played at the selected speed, which is convenient for load testing and further analysis. The parameter is not used if the value is 0 |
0 |
After saving the settings for the newly created video stream, the message โSource has been successfully createdโ will appear on the screen (Figure 53).
Video stream editing#
Clicking the (4 in the Figure 51) button to start editing a video stream, or click the โEditโ button on the page for viewing the video stream of the source (Figure 54).
General view of the form for editing a video stream is similar to the view of the form for creating a video stream (Figure 52).
Start editing stream parameters. If you need to go back to the page with the list of sources during editing, press the Esc key on your keyboard.
After saving the settings for the stream parameters, the message โSource was successfully updatedโ will appear (Figure 55).
Video stream deleting#
To delete a video stream, click on the button in the line with the video stream you want to delete (3 in the Figure 51). In the pop-up window (Figure 56), confirm the actionโclick on the โDeleteโ button or cancel the action by clicking on the โCancelโ button.
After clicking the โDeleteโ button, a message about successful stream deletion appears (Figure 57).
Tasks section#
The โTasksโ is intended for creating, deleting, and displaying tasks, downloading search results by events and persons. Export to a file is implemented in the Interface in the form of a task;
General view of the โTasksโ section is shown below (Figure 58).
The โLUNA PLATFORM. Deferred tasksโ tab contains the following elements:
- task counter (1);
- โCross-matchingโ button โ button for creating a task for cross-matching lists of faces;
- โExport facesโ button โ button for creating a task to export faces and information on them;
- โExport eventsโ button โ button for creating a task to export events and information on them;
- โBatch processingโ button โ button for creating a task for batch processing of photo image archives according to a specific policy;
- โBatch importโ button โ button for creating a task for batch import of photo image archive into the list;
- โBatch identificationโ โ button for creating a task for identifying an archive of photo images of references with candidates (faces or events with faces);
- โDeleting faces from the listโ โ button for creating a task for removing persons from the selected list;
- table of tasks:
- โIDโ โ task identifier;
- โDescriptionโ โ user who created the task;
- โTypeโ โ task type (cross-matching, export, batch processing, batch import, batch identification);
- โDate createdโ โ date and time of task creation;
- โExpiration dateโ โ date and time of completion of the task;
- โStatusโ โ task progress state;
โ button to stop the task (appears if the task status is โIn progressโ);
โ button for downloading the result of the task (2);
โ button for deleting a task (3);
- the number of tasks displayed on the page is set by the switch in the lower right corner of the page. There can be 10, 25, 50 or 100 tasks in total on one page (4).
The status changes during the task execution. In total, 4 statuses are applied to tasks in the Interface:
โ task is being performed;
- โCollecting resultsโ โ collecting the results of the task;
โ task completed;
โ an error occurred while executing a task.
The process of creating tasks and the values of the specified parameters are described below. If you need to go back to the task list page during creating a task, press the Esc key on your keyboard.
Configure notifications about task status using the "callbacks" functionality. Notifications will be sent to the external system at the specified URL. The notification settings block opens after filling in the required fields to create a task (Table 25).
Table 25. Notification settings in the task creation form
Parameter |
Description |
Default value |
---|---|---|
Add callback |
||
Type |
Protocol type when creating a notification. Users can receive notifications in the interface via http and Telegram |
HTTP |
URL |
Address of the external system where the notification will be sent |
- |
Authorization type |
Selecting the type of authorization into an external system and setting up authorization data. The basic type of authorization requires specifying login and password to enter an external system |
Basic |
Timeout (seconds) |
Maximum time to wait for a request to complete |
60 |
Request body format |
Data interchange format: JSON or MessagePack |
application/json |
HTTP Headers |
HTTP Request Headers |
- |
Creating a cross-matching task#
To create a task for cross-matching lists of faces, click on the โCross-matchingโ button (Figure 58). A general view of the form for creating a cross-matching task is shown below (Figure 59).
The โCross-matchingโ form contains the following elements:
- โListโ โ selection of a list for comparison. Required field;
- โFind matches inโ โ selection of a list for comparison. Required field;
- โThe maximum number of similar onesโ โ maximum number of similar candidates (the default is 3);
- โMinimum similarity threshold, %โ โ the lowest score of similarity in percentage between candidates that the Interface accepts as a possible match (the default is 50).
Fill in all the required parameters and click on the โCreate taskโ button (or Enter key on your keyboard)..
Resource-intensive tasks can take a while. In the pop-up window (Figure 60), you must confirm the action โ click on the โOkโ button or cancel the action by clicking on the โCancelโ button (Esc key on keyboard)..
After successfully creating a cross-matching task, the message โTask for cross-matching has been createdโ will appear in the upper right corner of the screen (Figure 61).
Creating an export faces task#
To create a task to export faces and information on them, click on the โExport facesโ button (Figure 58). A general view of the form for creating an export task is shown below (Figure 62).
Description of the parameters of the โExport facesโ form is given in Table 26.
Table 26. Export faces task parameters
Parameter |
Description |
Default value |
---|---|---|
Data upload parameters |
||
List |
Specifies the list for export |
- |
User data |
Indicates face data (up to 128 characters) |
- |
Face IDs |
Specifies the values of identifiers of faces in LUNA PLATFORM 5 in UUID format |
- |
Face external IDs |
Specifies the values of third-party external identifiers |
- |
Date from |
Specifies the lower limit of the period of creation of faces or events in LUNA PLATFORM 5 |
- |
Date to |
Specifies the upper limit of the period of creation of faces or events in LUNA PLATFORM 5 |
- |
ID of the first face |
Specify the values of the identifier of the first face from the exported faces |
- |
ID of the last face |
Specify the values of the identifier of the last face from the exported faces |
- |
Additional parameters |
||
Columns in the reportโselecting table columns to be included in the file upon export, and indication the order in which they are located |
Face ID User data External ID Time Avatar Event ID Lists |
On On On On On On Off |
Save face images |
Enabling this option allows you to upload face images into the archive with the .csv report |
Off |
Type of biometric template |
Specifies the biometric template of which objects will be exportedโfaces or bodies |
Faces |
Delimiter for .csv |
A special character that will be used in the file with export results to divide text into columns |
, |
Fill in all the required parameters and click on the โCreate taskโ button.
Resource-intensive tasks can take a while. In the pop-up window (Figure 63), you must confirm the action โ click on the โOkโ button or cancel the action by clicking on the โCancelโ button (Esc key on keyboard)..
After successfully creating an export task, the message โExport task has been successfully createdโ will appear in the upper right corner of the screen (Figure 64).
Creating an export events task#
To create a task to export objects events and information on them, click on the โExport eventsโ button (Figure 58). A general view of the form for creating an export events task is shown below (Figure 65).
Description of the parameters of the โExport eventsโ form is given in Table 27.
Table 27. Export events task parameters
Parameter |
Description |
Default value |
---|---|---|
Data upload parameters |
||
General data about the event |
||
Source |
Specifies the name of the event source |
- |
Event IDs |
Specifies the values of the event identifiers in UUID format for performing an accurate search separated by commas |
- |
User data |
Indicates face data (up to 128 characters) |
- |
Event external IDs |
Specifies the values of third-party external identifiers separated by commas |
- |
Face IDs |
Specifies the values of identifiers of faces in UUID format separated by commas |
- |
Similarity |
A value from 0 to 1 is specified |
- |
Tags |
Specifies a tag or tags separated by commas |
- |
Handling policies |
Specifies policy name, it is possible to specify several values |
- |
Date from |
Specifies the lower limit of the period of creation of events |
- |
Date to |
Specifies the upper limit of the period of creation of events |
- |
Advanced event filters |
||
End date from |
Indicates the lower limit of the event end period |
- |
End date to |
Indicates the upper limit of the event end period |
- |
ID of the first event |
Indicate the value of the identifier of the first event from the exported events |
- |
ID of the last event |
Indicates the value of the identifier of the last event from the exported events |
- |
Track IDs |
Specifies the values of the track identifiers in the UUID format separated by commas |
- |
Face attributes |
||
Gender |
Specifies male/female gender |
- |
Age category |
Specifies the age range |
- |
Emotions |
Specifies emotions |
- |
Medical mask |
Detection of the presence/absence of a medical mask, mouth occlusion |
- |
Liveness |
Specifies the result of checking for the presence of a living person in the frame |
|
Body attributes |
||
Gender by body |
Specifies the female, male, undefined gender |
- |
Age category by body |
Specifies the age range by body image |
- |
Headwear |
Specifies headdress |
- |
Upper body colors |
Specifies top clothing color |
- |
Sleeve |
Specifies sleeve length |
- |
Headwear color |
Specifies headdress color |
- |
Lower body colors |
Specifies bottom clothing color |
- |
Lower body type |
Specifies bottom clothing type |
- |
Shoes color |
Specifies shoe color (only for โIdentify among events); |
|
Backpack |
Specifies backpack presence |
- |
Best match data |
||
Face IDs |
Specifies the values of identifiers of faces 5 in UUID format |
- |
Face external IDs |
Specifies the values of third-party external identifiers |
- |
Label |
Name of the labelโrule by which the comparison occurred |
- |
Geoposition |
โDistrictโ; โAreaโ; โCityโ; โStreetโ; โHouse numberโ; |
|
Advanced geoposition filters |
โLongitude (-180โฆ180)โ; โAccuracy (0โฆ90)โ; โLatitude (-90โฆ90)โ; โAccuracy (0โฆ90)โ |
|
Other |
||
Add filter by meta |
Allows you to fill in a set of blocks to create a filter by the "meta" field. The number of meta filters is unlimited. The following blocks are required to be filled in when creating a filter by meta:
|
|
Additional parameters |
||
Columns in the reportโselecting table columns to be included in the file upon export, and indication the order in which parameters are located in the report |
Event data DFace data from the event Body data from the event Match data Geo position |
Off Off Off Off On |
Save face images |
Enabling this option allows you to upload face images into the archive with the .csv report |
Off |
Type of biometric template |
Specifies the biometric template of which objects will be exportedโfaces or bodies |
Faces |
Delimiter for .csv |
A special character that will be used in the file with export results to divide text into columns |
, |
Fill in all the required parameters and click on the โCreate taskโ button or the Enter key on your keyboard.
Resource-intensive tasks can take a while. In the pop-up window, confirm the action โ click on the โOkโ button or cancel the action by clicking on the โCancelโ button (Esc key on keyboard).
After successfully creating an export task, the message โExport task has been successfully createdโ will appear in the upper right corner of the screen.
Creating a batch processing task#
The batch processing task allows user to process several photos using a specified policy.
To create a task for batch processing of photo image archives according to a specific policy, click on the โBatch processingโ button (Figure 58). The general view of the form for creating a batch processing task is shown below (Figure 66).
By default the โBatch Processingโ form contains the following elements:
- โData source typeโ โ selection of the source type of the loaded data;
- โDescriptionโ โ description of the task;
- โHandling policyโ โ selection of a policy. Required field;
The resource can accept five types of sources with images for processing:
- File;
- ZIP;
- S3;
- Network disk;
- FTP;
- Samba.
Additional options appear depending on the selected data source type.
To quickly download a ZIP archive from your local machine without additional options, select โFileโ as the data source type. Then upload or drag-and-drop the archive with photo images in the field for uploading data.
Download file requirements:
- *.zip file format;
- there can be one or more people on the image (depends on policy settings);
- the image must contain a person's face or body;
- images must be located immediately inside the archive, and not in a folder inside the archive;
- the archive size is set using the ARCHIVE_MAX_SIZE parameter in the config.py configuration file of the Tasks component, the default size is 100 GB (for details, see โVisionLabs LUNA PLATFORM 5. Administrator manualโ).
When choosing a ZIP archive as image source for the batch processing task, the following parameters can be set:
- โFile URLโ โ URL address of the archive with images, the default archive size is 100 GB. Required field;
- โArchive passwordโ โ a password for the transferred archive protection;
- โFile key prefixโ โ a file key prefix that can be used to load images from a specific directory, for example, "2022/January";;
- โFile key postfixโ โ file key postfix that can be used to upload images with a specific extension;
- โWhether to estimate images from ZIP archive subdirectories recursively?โ switch โ allows you to recursively receive images from subdirectories.
- โInput image typeโ โ selection of the type of image that is input in the batch processing task โ "Raw image", "Face warped image", "Body warped image".
When choosing an S3-like storage as an image source for the batch processing task, the following parameters can be set:
- โStorage endpointโ โ only when specifying the bucket name;
- โBucket nameโ โ Access Point ARN / Outpost ARN. Required field;
- โFile key prefixโ โ file key prefix. It can be used to load images from a specific folder, such as "2022/January";
- โBucket regionโ โ only when specifying the bucket name;
- โPublic access keyโ โ public key for setting up authorization. Required field;
- โSecret access keyโ โ secret key for setting up authorization. Required field;
- โSignature versionโ โ signature "s3v2" / "s3v4";
- โWhether to estimate images from bucket subdirectories recursively?โ โ possibility to recursively download images from nested bucket folders;
- โWhether to save image origin?โ โ saving original images in the LUNA PLATFORM 5 database.
It is also possible to select the type of transferred images. For more information about working with S3-like repositories, see AWS User Guide.
When choosing a network disk as an image source for the batch processing task, the following parameters can be set:
- โPath to directory with imagesโ โ absolute path to the directory with images in the container. Required field;
- โFile key prefixโ โ a file key prefix that can be used to load images from a specific directory;
- โFile key postfixโ โ file key postfix that can be used to upload images with a specific extension;
- โWhether follow file system links?โ โ enable/disable of symbolic links processing.
As in the batch processing task using S3-like storage as image source, it is possible to recursively receive images from nested directories, and to select the type of transferred images.
When choosing a FTP server as an image source for the batch processing task, the following parameters can be set:
- โServer hostโ โ FTP server IP address or hostname. Required field;
- โPortโ โ FTP server port;
- โFTP sessionsโ โ maximum number of allowed sessions on the FTP server;
- โServer userโ and โServer passwordโ โ authorization parameters.
As in the batch processing task using network disk as image source, it is possible to set the path to the directory with images, recursively receive images from nested directories, select the type of transferred images, and specify the prefix and postfix.
When choosing a Samba as an image source for the batch processing task, the parameters are similar to those of an FTP server, except for the "max_sessions" parameter. Also, if authorization data is not specified, the connection to Samba will be performed as a guest.
Fill in all the required parameters and click on the โCreate taskโ button or the Enter key on your keyboard. Resource-intensive tasks can take a while. In the pop-up window (Figure 67), you must confirm the action โ click on the โOkโ button or cancel the action by clicking on the โCancelโ button.
After successfully creating a batch processing task, the message โBatch processing task has been successfully createdโ will appear in the upper right corner of the screen (Figure 68).
Creating a batch import task#
The batch import task allows you to batch import faces from photos into a specified list.
To create a task for batch import of photo image archive into the list, click on the โBatch importโ button (Figure 58). The general view of the form for creating a batch import task is shown below (Figure 69).
The โBatch importโ form contains the following elements:
- field for uploading an archive with photographs* โ it is possible to upload archives in *.ZIP format (required);
- โListโ โ selection of a list. Required field;
- โAdd a photo to the list only if it complies with the ISO/IEC standardโ โ the photo will be added to the list only after passing the ISO/IEC 19794-5:2011 verification.
โ button for deleting the loaded archive โ button for deleting the loaded archive.
Download file requirements:
- *.ZIP file format;
- there can be one or more people on the image (depends on policy settings);
- the image must contain a person's face;
- images must be located immediately inside the archive, and not in a folder inside the archive;
- the archive size is set using the ARCHIVE_MAX_SIZE parameter in the config.py configuration file of the Tasks component, the default size is 100 GB (for details, see โVisionLabs LUNA PLATFORM 5. Administrator manualโ).
Fill in all the required parameters and click on the โCreate taskโ button or the Enter key on your keyboard.
Resource-intensive tasks can take a while. In the pop-up window (Figure 70), you must confirm the action โ click on the โOkโ button or cancel the action by clicking on the โCancelโ button.
After successfully creating a batch import task, the message โBatch import task has been successfully createdโ will appear in the upper right corner of the screen (Figure 71).
Creating a batch identification task#
To create a task for identifying an archive of photo images (faces or events with faces), click on the โBatch identificationโ button (Figure 58). The general view of the form for creating a batch identification task is shown below (Figure 72).
The โBatch Identificationโ form contains the following elements:
- field for uploading an archive with photographs in *.ZIP format (required);
โbutton for deleting the loaded archive;
- โIdentify amongโโlook for matches among โFacesโ or โEventsโ;
- โFiltersโ blockโsettings for user identification. A description of the parameters of the โFiltersโ block, depending on the selected object for identification, is presented below (Table 28 and Table 29);
- โAdditional filter parametersโ blockโgeneral parameters for identification among faces and events:
- โSimilarity threshold, %โโthe lowest percentage similarity score between candidates that the Interface accepts as a possible match (default: 80).
- โNumber of records (from 1 to 100)โโthe number of lines with matches with a limit of 100 lines (default: 3).
Table 28. โFiltersโ block parameters of the batch identification task when searching for a match among faces
Name |
Description |
---|---|
List |
List name |
Comma-separated Face IDs |
Face ID from the list |
User data |
Information linked to the person from the database |
Comma-separated external face IDs |
External ID of persons face |
Date from |
Specifies the lower limit of the period of creation of faces or events in LUNA PLATFORM 5 |
Date to |
Specifies the upper limit of the period of creation of faces or events in LUNA PLATFORM 5 |
Table 29. โFiltersโ block parameters of the batch identification task when searching for a match among events
Name |
Description |
---|---|
Source |
List of available event sources |
User data |
Information linked to the person from the database |
Age category |
Age group: below 18; from 18 to 44; from 45 to 60; above 60 |
Gender |
Female; Male Not estimated |
Emotion |
Anger; Sadness; Neutral; Disgust; Fear; Happiness; Surprise. Its possibly to select multiple emotions. |
Medical mask |
Detection of the presence/absence of a medical mask or mouth occlusion. Missing; Medical mask; Occluded. Its possibly to select multiple variants. |
Creation date from |
Specifies the lower limit of the period of creation of faces or events in LUNA PLATFORM 5 |
Creation date to |
Specifies the upper limit of the period of creation of faces or events in LUNA PLATFORM 5 |
Comma-separated event IDs |
Event ID of detection and attribute retrieval |
Comma-separated external event IDs |
External event ID |
Comma-separated Face IDs |
Face ID from events that are created in LUNA PLATFORM 5 as a result of a detection event and attribute extraction |
Similarity |
The lower threshold on the similarity if the person was identified |
Handling policies |
Handling Policy ID |
Comma-separated track IDs |
Specifies the values of the track identifiers in LUNA PLATFORM 5 in the UUID format |
Comma-separated tags |
Specifies a tag or tags |
Gender by body |
female male undefined |
Headwear |
Specifies headdress |
Sleeve |
Specifies sleeve length |
Lower body colors |
Specifies bottom clothing color |
Shoes color |
Specifies shoe color (only for โIdentify among events); |
Age category by body |
Specifies the age range below 18; from 18 to 44; from 45 to 60; above 60 |
Upper body colors |
Specifies top clothing color |
Headwear color |
Specifies headdress color |
Lower body type |
Specifies bottom clothing type |
Backpack |
Specifies backpack presence |
Location |
โDistrictโ; โAreaโ; โCityโ; โStreetโ; โHouse numberโ; โLongitude (-180โฆ180)โ; โAccuracy (0โฆ90)โ; โLatitude (-90โฆ90)โ; โAccuracy (0โฆ90)โ |
To upload an archive with photo images of faces to be identified, click on in the โReferencesโ section and specify the path to the archive on the local computer.
Download file requirements:
- *.ZIP file format;
- there can be one or more people on the image (depends on policy settings);
- the image must contain a person's face;
- images must be located immediately inside the archive, and not in a folder inside the archive;
- the archive size is set using the ARCHIVE_MAX_SIZE parameter in the config.py configuration file of the Tasks component, the default size is 100 GB (for details, see โVisionLabs LUNA PLATFORM 5. Administrator manualโ).
Fill in all the necessary parameters and click the โCreate taskโ button or the Enter key on your keyboard..
Resource-intensive tasks can take a while. In the pop-up window (Figure 73), you must confirm the action โ click on the โOkโ button or cancel the action by clicking on the โCancelโ button.
After successfully creating a batch identification task, the message โBatch identification task has been successfully createdโ will appear in the upper right corner of the screen (Figure 74).
Creating a task for deleting faces from the list#
The task of removing persons from the list (Cleanup task) allows you to select faces based on specific parameters and then remove them from the selected list.
To create a Cleanup task, click on the โDeleting faces from the listโ button (Figure 58). The general view of the window for creating a task for batch import is shown below (Figure 75).
The โDeleting faces from the listโ window contains the following elements:
- โDescriptionโโa field for adding an explanatory note to the task;
- โStore resultsโ checkboxโif enabled, the results of the task will be saved in the Image Store service storage.
- โDelete samplesโ checkboxโif enabled, wrapped images obtained after detecting faces from the list will be deleted;
- โListโโselect a list from which faces will be removed. Required field;
- โInformationโโa field for specifying information about persons from the list. Allows you to remove only a few people from the list, for example, those for whom the same information is provided.
- โDelete data afterโโthe lower included threshold value of the face creation time;
- โDelete data beforeโโthe upper excluded threshold value of the face creation time.
Fill in all the required parameters and click on the โCreate taskโ button or the Enter key on your keyboard.
Resource-intensive tasks may take some time to complete. Confirm the action in the pop-up windowโclick the โOkโ button or cancel the action using the โCancelโ button (Figure 76).
After successfully creating a task for removing persons from the list, the message โCleanup task has been successfully createdโ appears on the screen (Figure 77).
Viewing the results of a task#
Viewing the results of a task is performed by pressing the button in the line (3 in the Figure 58). For viewing the *.ZIP archive for export tasks, the *.csv file for cross-matching tasks, the *.json file for batch processing, batch import and batch identification tasks (where * is the task ID) will be loaded.
The downloaded *.csv file contains a table with the export parameters selected in "Creating an export task" section (Figure 78) or with the results of cross-matching (Figure 79).
Task deleting#
Deleting a task is performed by clicking the button in the line (4 in the Figure 58).
In the pop-up window (Figure 80), you must confirm the action โ click on the โDeleteโ button or cancel the action by clicking on the โCancelโ button. After the successful deletion, a corresponding notification appears.
"Statistics" section#
The "Statistics" section is designed for getting statistical data on events and building graphs based on them. When you open the section, a page opens with a list of pre-configured charts (Figure 81):
- Age distribution;
- Attendance by gender;
- Distribution gender,
- Gender ratio;
- Emotion distribution;
- Average age by gender.
Go to the chart editing page by clicking next to the chart. Change the field values โโand build a new charts.
Chart settings#
The chart creation and configuration page consists of table (Table 28). Specify the data for building the chart. After selecting the parameters, click the "Refresh" button to show the specified data on the chart. Click "Reset to default" to reset values โโin the fields to the original ones.
Table 28. Description of the elements of the chart setup page
Name |
Description |
---|---|
"Basic parameters" |
Fields required for plotting a chart |
"Chart Title" |
Field for entering the chart title |
"Chart type" |
Drop-down list for selecting the chart type:
|
"Data" |
Drop-down list for selecting event characteristic. The chart is based on those events that have the selected characteristic filled in. You can select events that have, for example, "top ,atching candidates label", "Liveness" rating or event ID โ for grouping by all events. The event characteristics to select in the "Data" field are related to the data shown in the last events. |
"Aggregator" |
Drop-down list for selecting the type of data aggregation. The set of acceptable values โโin the field depends on the value selected in the "Data" field. Possible values โโof the "Aggregator" field:
|
"Grouping" tab |
|
Group by time |
Fields required for grouping data by time |
"Group by" |
Drop-down list for selecting the type of grouping by time. Available values:
The โCountโ field is used to specify the number of time intervals specified in the โDimensionโ field. The โDimensionโ field is a drop-down list of dimension values. Valid field values: "Minute", "Hour", "Day", "Week", "Month", "Year". The โFrequencyโ field is a drop-down list of frequency values. Valid field values: "By months of the year", "By weeks of the year", "By days of the year", "By days of the month", "By days of the week", "By hours of the day", "By minutes of the hour" |
Group by field |
Fields required to group data by field value |
"Field" |
Drop-down list for selecting the field by which events will be grouped. For example, specify "Age" to get the distribution of events by age. Leave the value empty, to get the distribution by all events containing the characteristic specified in the "Data" field, "Basic Parameters" tab. Add an additional field for distribution, for example "Gender", to get a combined chart (Figure 82) The field is available for input except when the "Donut" or "Pie" chart type or the group by by time intervals is selected. |
Filters tab |
|
"Period" |
Fields required to filter data by period |
"From" and "To" |
Fields for selecting a date. The chart includes events created between the specified dates and times. |
โDataโ |
Fields required to group data by field value |
โAdd fieldโ button |
Button for adding a new group of fields to fill in the filter parameters |
"Field" |
Drop-down list for selecting the field by which filtering will be performed |
"Function" |
Drop-down list of functions for filtering. Activated by selecting a value in the "Field". Valid field values: Contains/Does not contain symbol, Includes/Excludes value, Equals/Not equals, Greater than, Greater than or equal to, Less than, Less than or equal to, Empty, Included in the area |
โValueโ |
Field for entering a value corresponding to the function for filtering. It can be either text format or drop-down options, depending on the selected value in the "Field". Activated by selecting a value in the "Functions" field. For example, select filtering by "Emotions" in the "Field". Specify the value "Equals" in the "Function" field and "Happiness" in the "Value" field to find out the number of events in which the emotion of the detected of the person was identified as "Happiness". Add another field to expand the filtering. Specify "Age" in "Field", "Greater than or equal to" in the "Function" and "30" as a "Value". The chart will show the number of events where people with the emotion "Happiness" and age from 30 years were recognized (Figure 83) |
Download the event report that the graph is based on. Clicking the or
button to download the report in json or csv format.
"Checks" section#
The "Checks" section allows you to upload and check photos and videos.
Checks are available only if user has the corresponding licenses.
The โChecksโ section contains the following items (Figure 84):
- "Image" and "Video" tabs to upload photos or videos by clicking the file or using drag and drop;
- A list of checks, where the user selects which ones to use to check the image or video.
Image check#
To start a photo check:
- open the "Image" tab;
- upload or drag and drop your file;
- select checks from the list:
- Liveness;
- DeepFake;
- compliance with the requirements of the ISO/IEC 19794-5:2011 standard;
- compliance with ICAO standard;
- compliance with the biometric standards.
- click the "Check" button.
The page with the check results contains (Figure 85):
- Final assessment of whether the photo passed the selected checks;
- Results for each check criterion: if the photo meets the criterion, then the font color is green; if it doesn't, then the font color is red. A photo must meet all the criterion in order to comply with the check;
- The number of criteria that the photo does not meet if the result of the check is negative;
- Button for downloading results of checking in json;
- Button to reset old photo. Allows you to start a new scan for a different photo.
Image file requirements:
.jpeg
,.png
,.tif
or.bmp
format;- image size no more than 15 MB and no more than 3840x2160 pixels;
- image may contain one or more people;
- image must have a person's face.
Video check#
To start a video check:
- open the "Video" tab;
- upload or drag the file;
- select checks from the list:
- Liveness;
- DeepFake;
- Face attributes;
- Body attributes;
- click the "Check" button.
Video file requirements:
- file format
.avi
,.mp4
,.mov
,.mkv
,.flv
; - file size depends on specified timeout during which a connection with the service is established when uploading a video. For more information, see the
LUNA_IMAGE_STORE_OBJECTS_TIMEOUTS
parameters in the LUNA PLATFORM 5 documentation.
The page with the check results contains (Figure 86):
- A window with a processed video fileโthe recognition objects, bodies and faces, are bounded into bbox on the video frames (1);
- A carousel of frames with events (2);
- A time track with per-second marking indicating the current time of the video. Move the point to move along the video timecodes (3);
- Time lines with color-coded frame check results that are shown depending on checks selected (4);
- Detailed video segment evaluation results for the selected checks (the average result is calculated for the segment frames) (5);
- Button to reset old video. Allows you to start a new scan for a different video.
Hover the cursor over a timecode segment to see what time period of the video it belongs to. Click on the video segment timecode to show frames of particular events that occurred during the selected period. Switch by timecode and view the video segment check results (Figure 87).
- For Liveness and Deepfake:
- Live/real faceโgreen timecode;
- Fake faceโred timecode;
- Unknownโgray timecode.
- For face and body attributes:
- Availability of ratingโgreen timecode.
The absence of events has no color designation and is not clickable in the timecode. When events overlap, their color timecodes also overlap each other.
Users section#
The โUsersโ section is intended for showing user accounts, created in LUNA PLATFORM 5. General view of the โUsersโ section is shown below (Figure 86).
โUsersโ section contains the following elements:
- Table of existing user accounts containing columns:
- "Loginโaccount login;
- "Description"โaccount description
- "Account type"
- user โ the type of account with which you can create objects and use only your account data.
- advanced_user โ allows to interact with its own data and view other accounts data
- admin โ the type of account for which rights similar to "advanced_user" are available, and there is also access to the Admin service.
- "Create time" โ date and time of account creation;
- "Last update time" โ date and time of the last account update.
To sort a column in the table, click on the column name. The sorting arrow icon
indicates the current sorting by one of the parameters: alphabetically, ascending, or descending.
You can work with all types of accounts in the API service, but only "advanced_user" and "user" types of accounts can be created, while in the Admin service you can create all three types.
Creating a user account#
Create a user account using a POST request "create account" to the API service, or using the Admin service. When creating the account, you must specify the following data: login (email), password and account type.
Monitoring section#
The โMonitoringโ section is intended for viewing information and status of connected services, modules, components, and systems (Figure 87).
โMonitoringโ section contains the following elements:
- List of connected services, modules, components, and systems:
- โNameโ โ component/service/system name;
- โVersionโ โ component/service/system version;
- โStateโ โ current state (status) of a component/service/system;
- "Health check" โ checks access to the service. Useful in situations when the service is running, i.e. in the "Loaded" state, but is not available. The system response time in seconds is shown next to the check status;
- โDocumentationโ โ links to the documentation if it is present in the component/service/system.
This colors are used to indicate the current status of a service, module, component, or system:
โ green color โ component/service/system is up and running;
โ blue color โ component/service/system is loading;
โ red color โ component/service/system is temporarily unavailable.
Licenses section#
The โLicensesโ section provides the information about available licenses.
The section contains the following tabs:
โGeneral informationโโshows whether the license has expired (Figure 88);
โEstimationsโโshows the status (enabled/disabled) of licenses for (Figure 89):
- estimating the number of people in an image (PlatformPeopleCounter);
- checking quality of facial image (PlatformISO);
- estimation of body attributes (PlatformBodyAttributes);
- Liveness check (PlatformLiveness);
- Deepfake check (PlatformDeepfake);
โModulesโโshows the status (enabled/disabled) of service licenses for (Figure 90):
- storing data about events in the database (Events Service);
- completing tasks (Tasks Service);
- sending event notifications via a web socket (Sender Service);
- storing samples, reports on task execution, created clusters and additional metadata (Image Store Service);
- creating and storing handlers (Handlers Service).
Each tab allows you to go to the page for buying license or renewal.
Plugins section#
The โPluginsโ section provides the information about plugins imported into LUNA PLATFORM 5. Plugins are used to perform secondary actions for the userโs needs. For example, you can expand the standard functionality of the product using them. The general view of the โPluginsโ section is presented below (Figure 91).
โPluginsโ section contains the following elements:
- table of plugins:
- โNameโ โ name of the plugin;
- โStatusโโshows the current status (running/not running) of the plugin.
For more information on getting a list of imported plugins and their status, see LUNA PLATFORM 5 documentation .