iOSLivenessSDK
📜 Introduction
Brought to you by FaceX.io, this iOSLiveness SDK can now be used to integrate gesture-based liveness Detection into your iOS applications.
🔧 Functioning
This LivenessSDK is Liveness based on Motion detection. Users will be directed by the screen to perform facial gestures and actions which will be analysised to verify and identify live visitors.
📑 Index
- Documentation
- Features
- Supported Devices
- Software Prerequisite
- Installation Using CocoaPods
- How to use
- Delegates
- Customization
- Supported OS & SDK Versions
- License
📚 Documentation
- Step 1: Sign up for a developer account through FaceX User Portal
- Step 2: Navigate to Plans & Payments Tab & Select Plans
- Step 3: Scroll Down to Mobile SDK Section and select Active Liveness
- Step 4: Select the number of licenses you require, add the required details and Purchase your plan
- Step 5: Navigate to Mobile SDK Tab & Select License History
- Step 5: Download the plist file for iOS SDK from the Offline License Section
- Step 6: Download the Online Active Livenesss iOS SDK from this Github LivenessSDK
- Step 7: Add the plist file from the portal to your project.
🌟 Features
Utility
Ensure Secure use of facial recognition based applications by determining liveness of your users
Liveness Detection by Motion
With this iOSLiveness SDK, you can determine liveness of your users by detection of motion.
Users will be prompted by the screen to perform facial gestures and the response by the user is analyzed in a live streaming video.
✔️ Supported Devices
- Supports iOS 12.0+
- Supports binary
- Supports iPhone and iPad
🔍 Software Prerequisite
- Requires Swift 4/5 and Xcode 10.x
📲 Installation Using CocoaPods
CocoaPods is a dependency manager for Cocoa projects. For usage and installation instructions, visit their website. To integrate Alamofire into your Xcode project using CocoaPods, specify it in your Podfile
:
1. Create Podfile
and add pod 'LivenessSDK'
:
use_frameworks!
target 'YourApp' do
pod 'LivenessSDK'
end
2. Install pods:
$ pod install
3. Import the module:
Swift:
import LivenessSDK
Objective - C:
@import LivenessSDK;
🐒 How to use
Swift
import LivenessSDK
var liveness = Liveness()
liveness.delegate = self
liveness.enableEyes = true;
liveness.startLiveness()
// To stop and exit Liveness View
liveness.stopLiveness()
Objective - C
@import LivenessSDK;
Liveness * liveness = [Liveness new];
liveness.delegate = self;
liveness.enableEyes = true;
[liveness startLiveness];
// To stop and exit Liveness View
[liveness stopLiveness];
🏄 Delegates
LivenessDelegate
Swift
func livenessSuccess(live: Bool) {
}
func livenessError(live: Bool, error: NSError) {
}
Objective - C
- (void)livenessErrorWithLive:(BOOL)live error:(NSError * _Nonnull)error{
}
- (void)livenessSuccessWithLive:(BOOL)live face:(UIImage * _Nonnull)face {
}
🎛 Customization
You can set some properties for liveness.
Steps
| Steps | Value | Default |
| ——- | ——- |——- |
| Eyes(enableEyes) | Bool
| false
|
| Mouth(enableMouth) | Bool
| false
|
| Yaw(enableYaw) | Bool
| false
|
| Random Steps(randomSteps) | Bool
| true
|
Thresholds
| Property | Min. Value | Max. Value | Default |
| ——- | ——- | ——- |——- |
| Eyes closed Threshold(eyeThreshold) | 0.05
| 0.2
| 0.1
|
| Mouth opened Threshold(mouthThreshold) | 0.1
| 0.6
| 0.35
|
| Each step timer(timerSeconds) | Seconds
| Seconds
| 5 seconds
|
📋 Supported OS & SDK Versions
- iOS 12.0+
- iPadOS 13.0+
- Swift 5