[ aws . rekognition ]
For a given input face ID, searches for matching faces in the collection the face belongs to. You get a face ID when you add a face to the collection using the IndexFaces operation. The operation compares the features of the input face with faces in the specified collection.
Note
You can also search faces without indexing faces by using the SearchFacesByImage
operation.
The operation response returns an array of faces that match, ordered by similarity score with the highest similarity first. More specifically, it is an array of metadata for each face match that is found. Along with the metadata, the response also includes a confidence
value for each face match, indicating the confidence that the specific face matches the input face.
For an example, see Searching for a Face Using Its Face ID in the Amazon Rekognition Developer Guide.
This operation requires permissions to perform the rekognition:SearchFaces
action.
See also: AWS API Documentation
See ‘aws help’ for descriptions of global parameters.
search-faces
--collection-id <value>
--face-id <value>
[--max-faces <value>]
[--face-match-threshold <value>]
[--cli-input-json | --cli-input-yaml]
[--generate-cli-skeleton <value>]
--collection-id
(string)
ID of the collection the face belongs to.
--face-id
(string)
ID of a face to find matches for in the collection.
--max-faces
(integer)
Maximum number of faces to return. The operation returns the maximum number of faces with the highest confidence in the match.
--face-match-threshold
(float)
Optional value specifying the minimum confidence in the face match to return. For example, don’t return any matches where confidence in matches is less than 70%. The default value is 80%.
--cli-input-json
| --cli-input-yaml
(string)
Reads arguments from the JSON string provided. The JSON string follows the format provided by --generate-cli-skeleton
. If other arguments are provided on the command line, those values will override the JSON-provided values. It is not possible to pass arbitrary binary values using a JSON-provided value as the string will be taken literally. This may not be specified along with --cli-input-yaml
.
--generate-cli-skeleton
(string)
Prints a JSON skeleton to standard output without sending an API request. If provided with no value or the value input
, prints a sample input JSON that can be used as an argument for --cli-input-json
. Similarly, if provided yaml-input
it will print a sample input YAML that can be used with --cli-input-yaml
. If provided with the value output
, it validates the command inputs and returns a sample output JSON for that command.
See ‘aws help’ for descriptions of global parameters.
To search for faces in a collection that match a face ID.
The following search-faces
command searches for faces in a collection that match the specified face ID.
aws rekognition search-faces \
--face-id 8d3cfc70-4ba8-4b36-9644-90fba29c2dac \
--collection-id MyCollection
Output:
{
"SearchedFaceId": "8d3cfc70-4ba8-4b36-9644-90fba29c2dac",
"FaceModelVersion": "3.0",
"FaceMatches": [
{
"Face": {
"BoundingBox": {
"Width": 0.48166701197624207,
"Top": 0.20999999344348907,
"Left": 0.21250000596046448,
"Height": 0.36125001311302185
},
"FaceId": "bd4ceb4d-9acc-4ab7-8ef8-1c2d2ba0a66a",
"ExternalImageId": "image1.jpg",
"Confidence": 99.99949645996094,
"ImageId": "5e1a7588-e5a0-5ee3-bd00-c642518dfe3a"
},
"Similarity": 99.30997467041016
},
{
"Face": {
"BoundingBox": {
"Width": 0.18562500178813934,
"Top": 0.1618019938468933,
"Left": 0.5575000047683716,
"Height": 0.24770599603652954
},
"FaceId": "ce7ed422-2132-4a11-ab14-06c5c410f29f",
"ExternalImageId": "example-image.jpg",
"Confidence": 99.99340057373047,
"ImageId": "8d67061e-90d2-598f-9fbd-29c8497039c0"
},
"Similarity": 99.24862670898438
},
{
"Face": {
"BoundingBox": {
"Width": 0.18562500178813934,
"Top": 0.1618019938468933,
"Left": 0.5575000047683716,
"Height": 0.24770599603652954
},
"FaceId": "13692fe4-990a-4679-b14a-5ac23d135eab",
"ExternalImageId": "image3.jpg",
"Confidence": 99.99340057373047,
"ImageId": "8df18239-9ad1-5acd-a46a-6581ff98f51b"
},
"Similarity": 99.24862670898438
},
{
"Face": {
"BoundingBox": {
"Width": 0.5349419713020325,
"Top": 0.29124999046325684,
"Left": 0.16389399766921997,
"Height": 0.40187498927116394
},
"FaceId": "745f7509-b1fa-44e0-8b95-367b1359638a",
"ExternalImageId": "image9.jpg",
"Confidence": 99.99979400634766,
"ImageId": "67a34327-48d1-5179-b042-01e52ccfeada"
},
"Similarity": 96.73158264160156
},
{
"Face": {
"BoundingBox": {
"Width": 0.5307819843292236,
"Top": 0.2862499952316284,
"Left": 0.1564060002565384,
"Height": 0.3987500071525574
},
"FaceId": "2eb5f3fd-e2a9-4b1c-a89f-afa0a518fe06",
"ExternalImageId": "image10.jpg",
"Confidence": 99.99970245361328,
"ImageId": "3c314792-197d-528d-bbb6-798ed012c150"
},
"Similarity": 96.48291015625
},
{
"Face": {
"BoundingBox": {
"Width": 0.5074880123138428,
"Top": 0.3774999976158142,
"Left": 0.18302799761295319,
"Height": 0.3812499940395355
},
"FaceId": "086261e8-6deb-4bc0-ac73-ab22323cc38d",
"ExternalImageId": "image6.jpg",
"Confidence": 99.99930572509766,
"ImageId": "ae1593b0-a8f6-5e24-a306-abf529e276fa"
},
"Similarity": 96.43287658691406
},
{
"Face": {
"BoundingBox": {
"Width": 0.5574039816856384,
"Top": 0.37187498807907104,
"Left": 0.14559100568294525,
"Height": 0.4181250035762787
},
"FaceId": "11c4bd3c-19c5-4eb8-aecc-24feb93a26e1",
"ExternalImageId": "image5.jpg",
"Confidence": 99.99960327148438,
"ImageId": "80739b4d-883f-5b78-97cf-5124038e26b9"
},
"Similarity": 95.25305938720703
},
{
"Face": {
"BoundingBox": {
"Width": 0.5773710012435913,
"Top": 0.34437501430511475,
"Left": 0.12396000325679779,
"Height": 0.4337500035762787
},
"FaceId": "57189455-42b0-4839-a86c-abda48b13174",
"ExternalImageId": "image8.jpg",
"Confidence": 100.0,
"ImageId": "0aff2f37-e7a2-5dbc-a3a3-4ef6ec18eaa0"
},
"Similarity": 95.22837829589844
}
]
}
For more information, see Searching for a Face Using Its Face ID in the Amazon Rekognition Developer Guide.
SearchedFaceId -> (string)
ID of the face that was searched for matches in a collection.
FaceMatches -> (list)
An array of faces that matched the input face, along with the confidence in the match.
(structure)
Provides face metadata. In addition, it also provides the confidence in the match of this face with the input face.
Similarity -> (float)
Confidence in the match of this face with the input face.
Face -> (structure)
Describes the face properties such as the bounding box, face ID, image ID of the source image, and external image ID that you assigned.
FaceId -> (string)
Unique identifier that Amazon Rekognition assigns to the face.
BoundingBox -> (structure)
Bounding box of the face.
Width -> (float)
Width of the bounding box as a ratio of the overall image width.
Height -> (float)
Height of the bounding box as a ratio of the overall image height.
Left -> (float)
Left coordinate of the bounding box as a ratio of overall image width.
Top -> (float)
Top coordinate of the bounding box as a ratio of overall image height.
ImageId -> (string)
Unique identifier that Amazon Rekognition assigns to the input image.
ExternalImageId -> (string)
Identifier that you assign to all the faces in the input image.
Confidence -> (float)
Confidence level that the bounding box contains a face (and not a different object such as a tree).
FaceModelVersion -> (string)
Version number of the face detection model associated with the input collection (
CollectionId
).