Since the LiteRT builtin operator library only supports a limited number of TensorFlow operators, not every model is convertible. For details, refer to operator compatibility.
To allow conversion, users can enable the usage of certain TensorFlow ops in their LiteRT model. However, running LiteRT models with TensorFlow ops requires pulling in the core TensorFlow runtime, which increases the LiteRT interpreter binary size. For Android, you can avoid this by selectively building only required Tensorflow ops. For the details, refer to reduce binary size.
This document outlines how to convert and run a LiteRT model containing TensorFlow ops on a platform of your choice. It also discusses performance and size metrics and known limitations.
Convert a model
The following example shows how to generate a LiteRT model with select TensorFlow ops.
import tensorflow as tf
converter = tf.lite.TFLiteConverter.from_saved_model(saved_model_dir)
converter.target_spec.supported_ops = [
tf.lite.OpsSet.TFLITE_BUILTINS, # enable LiteRT ops.
tf.lite.OpsSet.SELECT_TF_OPS # enable TensorFlow ops.
]
tflite_model = converter.convert()
open("converted_model.tflite", "wb").write(tflite_model)
Run Inference
When using a LiteRT model that has been converted with support for select TensorFlow ops, the client must also use a LiteRT runtime that includes the necessary library of TensorFlow ops.
Android AAR
To reduce the binary size, please build your own custom AAR files as guided in the next section. If the binary size is not a considerable concern, we recommend using the prebuilt AAR with TensorFlow ops hosted at MavenCentral.
You can specify this in your build.gradle
dependencies by adding it alongside
the standard LiteRT AAR as follows:
dependencies {
implementation 'org.tensorflow:tensorflow-lite:0.0.0-nightly-SNAPSHOT'
// This dependency adds the necessary TF op support.
implementation 'org.tensorflow:tensorflow-lite-select-tf-ops:0.0.0-nightly-SNAPSHOT'
}
To use nightly snapshots, make sure that you have added Sonatype snapshot repository.
Once you've added the dependency, the necessary delegate for handling the graph's TensorFlow ops should be automatically installed for graphs that require them.
Note: The TensorFlow ops dependency is relatively large, so you'll probably
want to filter out unnecessary x86 ABIs in your .gradle
file by setting up
your abiFilters
.
android {
defaultConfig {
ndk {
abiFilters 'armeabi-v7a', 'arm64-v8a'
}
}
}
Building the Android AAR
For reducing the binary size or other advanced cases, you can also build the library manually. Assuming a working LiteRT build environment, build the Android AAR with select TensorFlow ops as follows:
sh tensorflow/lite/tools/build_aar.sh \
--input_models=/a/b/model_one.tflite,/c/d/model_two.tflite \
--target_archs=x86,x86_64,arm64-v8a,armeabi-v7a
This will generate the AAR file bazel-bin/tmp/tensorflow-lite.aar
for
LiteRT built-in and custom ops; and generate the AAR file
bazel-bin/tmp/tensorflow-lite-select-tf-ops.aar
for TensorFlow ops. If you
don't have a working build environment, You can also build above files with
docker.
From there, you can either import the AAR files directly into your project, or publish the custom AAR files to your local Maven repository:
mvn install:install-file \
-Dfile=bazel-bin/tmp/tensorflow-lite.aar \
-DgroupId=org.tensorflow \
-DartifactId=tensorflow-lite -Dversion=0.1.100 -Dpackaging=aar
mvn install:install-file \
-Dfile=bazel-bin/tmp/tensorflow-lite-select-tf-ops.aar \
-DgroupId=org.tensorflow \
-DartifactId=tensorflow-lite-select-tf-ops -Dversion=0.1.100 -Dpackaging=aar
Finally, in your app's build.gradle
, ensure you have the mavenLocal()
dependency and replace the standard LiteRT dependency with the one that
has support for select TensorFlow ops:
allprojects {
repositories {
mavenCentral()
maven { // Only for snapshot artifacts
name 'ossrh-snapshot'
url 'https://oss.sonatype.org/content/repositories/snapshots'
}
mavenLocal()
}
}
dependencies {
implementation 'org.tensorflow:tensorflow-lite:0.1.100'
implementation 'org.tensorflow:tensorflow-lite-select-tf-ops:0.1.100'
}
iOS
Using CocoaPods
LiteRT provides nightly prebuilt select TF ops CocoaPods for arm64
,
which you can depend on alongside the TensorFlowLiteSwift
or
TensorFlowLiteObjC
CocoaPods.
Note: If you need to use select TF ops in an x86_64
simulator, you can build
the select ops framework yourself. See Using Bazel + Xcode
section for more details.
# In your Podfile target:
pod 'TensorFlowLiteSwift' # or 'TensorFlowLiteObjC'
pod 'TensorFlowLiteSelectTfOps', '~> 0.0.1-nightly'
After running pod install
, you need to provide an additional linker flag to
force load the select TF ops framework into your project. In your Xcode project,
go to Build Settings
-> Other Linker Flags
, and add:
For versions >= 2.9.0:
-force_load $(SRCROOT)/Pods/TensorFlowLiteSelectTfOps/Frameworks/TensorFlowLiteSelectTfOps.xcframework/ios-arm64/TensorFlowLiteSelectTfOps.framework/TensorFlowLiteSelectTfOps
For versions < 2.9.0:
-force_load $(SRCROOT)/Pods/TensorFlowLiteSelectTfOps/Frameworks/TensorFlowLiteSelectTfOps.framework/TensorFlowLiteSelectTfOps
You should then be able to run any models converted with the SELECT_TF_OPS
in
your iOS app. For example, you can modify the Image Classification iOS
app
to test the select TF ops feature.
- Replace the model file with the one converted with
SELECT_TF_OPS
enabled. - Add
TensorFlowLiteSelectTfOps
dependency to thePodfile
as instructed. - Add the additional linker flag as above.
- Run the example app and see if the model works correctly.
Using Bazel + Xcode
LiteRT with select TensorFlow ops for iOS can be built using Bazel.
First, follow the iOS build instructions to configure your
Bazel workspace and .bazelrc
file correctly.
Once you have configured the workspace with iOS support enabled, you can use the
following command to build the select TF ops addon framework, which can be added
on top of the regular TensorFlowLiteC.framework
. Note that the select TF ops
framework cannot be built for i386
architecture, so you need to explicitly
provide the list of target architectures excluding i386
.
bazel build -c opt --config=ios --ios_multi_cpus=arm64,x86_64 \
//tensorflow/lite/ios:TensorFlowLiteSelectTfOps_framework
This will generate the framework under bazel-bin/tensorflow/lite/ios/
directory. You can add this new framework to your Xcode project by following
similar steps described under the Xcode project
settings section in the
iOS build guide.
After adding the framework into your app project, an additional linker flag
should be specified in your app project to force load the select TF ops
framework. In your Xcode project, go to Build Settings
-> Other Linker
Flags
, and add:
-force_load <path/to/your/TensorFlowLiteSelectTfOps.framework/TensorFlowLiteSelectTfOps>
C/C++
If you're using Bazel or CMake to build LiteRT interpreter, you can enable Flex delegate by linking a LiteRT Flex delegate shared library. You can build it with Bazel as the following command.
bazel build -c opt --config=monolithic tensorflow/lite/delegates/flex:tensorflowlite_flex
This command generates the following shared library in
bazel-bin/tensorflow/lite/delegates/flex
.
Platform | Library name |
---|---|
Linux | libtensorflowlite_flex.so |
macOS | libtensorflowlite_flex.dylib |
Windows | tensorflowlite_flex.dll |
Note that the necessary TfLiteDelegate
will be installed automatically when
creating the interpreter at runtime as long as the shared library is linked. It
is not necessary to explicitly install the delegate instance as is typically
required with other delegate types.
Note: This feature is available since version 2.7.
Python
LiteRT with select TensorFlow ops will be installed automatically with the TensorFlow pip package. You can also choose to only install the LiteRT Interpreter pip package.
Metrics
Performance
When using a mixture of both builtin and select TensorFlow ops, all of the same LiteRT optimizations and optimized builtin ops will be available and usable with the converted model.
The following table describes the average time taken to run inference on
MobileNet on a Pixel 2. The listed times are an average of 100 runs. These
targets were built for Android using the flags: --config=android_arm64 -c opt
.
Build | Time (milliseconds) |
---|---|
Only built-in ops (TFLITE_BUILTIN ) |
260.7 |
Using only TF ops (SELECT_TF_OPS ) |
264.5 |
Binary size
The following table describes the binary size of LiteRT for each build.
These targets were built for Android using --config=android_arm -c opt
.
Build | C++ Binary Size | Android APK Size |
---|---|---|
Only built-in ops | 796 KB | 561 KB |
Built-in ops + TF ops | 23.0 MB | 8.0 MB |
Built-in ops + TF ops (1) | 4.1 MB | 1.8 MB |
(1) These libraries are selectively built for i3d-kinetics-400 model with 8 TFLite builtin ops and 3 Tensorflow ops. For more details, please see the Reduce LiteRT binary size section.
Known limitations
- Unsupported types: Certain TensorFlow ops may not support the full set of input/output types that are typically available in TensorFlow.
Updates
- Version 2.6
- Supports for GraphDef-attribute based operators and HashTable resource initializations have improved.
- Version 2.5
- You can apply an optimization known as post training quantization
- Version 2.4
- Compatibility with hardware accelerated delegates has improved