Editing with Imagen is only supported if you're using the Vertex AI Gemini API . It's also currently only supported for Android and Flutter apps. Support for other platforms is coming later in the year.
This page describes how to use Imagen to replace the background of an imageusing the Firebase AI Logic SDKs.
Background replacement is a type of mask-based editing (specifically inpainting ). A mask is a digital overlay defining the specific area you want to edit.
How it works: You provide an original image and a corresponding masked image that defines a mask over the background — either using automatic background detection or providing the mask of the background yourself. You also provide a text prompt describing what you want to change. The model then generates and applies a new background.
For example, you can change the setting around a subject or object without affecting the foreground (for example, in a product image).
Jump to code for auto-detected background Jump to code for providing the background mask
Before you begin
If you haven't already, complete the getting started guide
, which
describes how to set up your Firebase project, connect your app to Firebase,
add the SDK, initialize the backend service for your chosen API provider, and
create an ImagenModel
instance.
Models that support this capability
Imagen
offers image editing through its capability
model:
-
imagen-3.0-capability-001
Note that for Imagen
models, the global
location is not
supported.
Replace background using automatic background detection
The following sample shows how to replace the background of an image — using automatic background detection. You provide the original image and a text prompt, and Imagen automatically detects and creates a mask of the background to modify the original image.
Swift
Image editing with Imagen models isn't supported for Swift. Check back later this year!
Kotlin
To replace the background using automatic background detection, specify ImagenBackgroundMask
. Use editImage()
and set the editing config to use ImagenEditMode.INPAINT_INSERTION
.
// Using this SDK to access Imagen models is a Preview release and requires opt-in
@OptIn
(
PublicPreviewAPI
::
class
)
suspend
fun
customizeImage
()
{
// Initialize the Vertex AI Gemini API backend service
// Optionally specify the location to access the model (for example, `us-central1`)
val
ai
=
Firebase
.
ai
(
backend
=
GenerativeBackend
.
vertexAI
(
location
=
"us-central1"
))
// Create an `ImagenModel` instance with an Imagen "capability" model
val
model
=
ai
.
imagenModel
(
"imagen-3.0-capability-001"
)
// This example assumes 'originalImage' is a pre-loaded Bitmap.
// In a real app, this might come from the user's device or a URL.
val
originalImage
:
Bitmap
=
TODO
(
"Load your original image Bitmap here"
)
// Provide the prompt describing the new background.
val
prompt
=
"space background"
// Use the editImage API to replace the background.
// Pass the original image, the prompt, and an editing configuration.
val
editedImage
=
model
.
editImage
(
referenceImages
=
listOf
(
ImagenRawImage
(
originalImage
.
toImagenInlineImage
()),
ImagenBackgroundMask
(),
// Use ImagenBackgroundMask() to auto-generate the mask.
),
prompt
=
prompt
,
// Define the editing configuration for inpainting and background replacement.
config
=
ImagenEditingConfig
(
ImagenEditMode
.
INPAINT_INSERTION
)
)
// Process the resulting 'editedImage' Bitmap, for example, by displaying it in an ImageView.
}
Java
To replace the background using automatic background detection, specify ImagenBackgroundMask
. Use editImage()
and set the editing config to use ImagenEditMode.INPAINT_INSERTION
.
// Initialize the Vertex AI Gemini API backend service
// Optionally specify the location to access the model (for example, `us-central1`)
// Create an `ImagenModel` instance with an Imagen "capability" model
ImagenModel
imagenModel
=
FirebaseAI
.
getInstance
(
GenerativeBackend
.
vertexAI
(
"us-central1"
))
.
imagenModel
(
/* modelName */
"imagen-3.0-capability-001"
);
ImagenModelFutures
model
=
ImagenModelFutures
.
from
(
imagenModel
);
// This example assumes 'originalImage' is a pre-loaded Bitmap.
// In a real app, this might come from the user's device or a URL.
Bitmap
originalImage
=
null
;
// TODO("Load your image Bitmap here");
// Provide the prompt describing the new background.
String
prompt
=
"space background"
;
// Define the list of sources for the editImage call.
// This includes the original image and the auto-generated mask.
ImagenRawImage
rawOriginalImage
=
new
ImagenRawImage
(
ImagenInlineImageKt
.
toImagenInlineImage
(
originalImage
));
// Use ImagenBackgroundMask() to auto-generate the mask.
ImagenBackgroundMask
rawMaskedImage
=
new
ImagenBackgroundMask
();
ImagenEditingConfig
config
=
new
ImagenEditingConfig
();
// Use the editImage API to replace the background.
// Pass the original image, the auto-generated masked image, the prompt, and an editing configuration.
Futures
.
addCallback
(
model
.
editImage
(
Arrays
.
asList
(
rawOriginalImage
,
rawMaskedImage
),
prompt
,
config
),
new
FutureCallback
<
ImagenGenerationResponse
>
()
{
@Override
public
void
onSuccess
(
ImagenGenerationResponse
result
)
{
if
(
result
.
getImages
().
isEmpty
())
{
Log
.
d
(
"ImageEditor"
,
"No images generated"
);
}
Bitmap
editedImage
=
((
ImagenInlineImage
)
result
.
getImages
().
get
(
0
)).
asBitmap
();
// Process and use the bitmap to display the image in your UI
}
@Override
public
void
onFailure
(
Throwable
t
)
{
// ...
}
},
Executors
.
newSingleThreadExecutor
());
Web
Image editing with Imagen models isn't supported for Web apps. Check back later this year!
Dart
To replace the background using automatic background detection, specify ImagenBackgroundMask
. Use editImage()
and set the editing config to use ImagenEditMode.inpaintInsertion
.
import
'dart:typed_data'
;
import
'package:firebase_ai/firebase_ai.dart'
;
import
'package:firebase_core/firebase_core.dart'
;
import
'firebase_options.dart'
;
// Initialize FirebaseApp
await
Firebase
.
initializeApp
(
options:
DefaultFirebaseOptions
.
currentPlatform
,
);
// Initialize the Vertex AI Gemini API backend service
// Optionally specify a location to access the model (for example, `us-central1`)
final
ai
=
FirebaseAI
.
vertexAI
(
location:
'us-central1'
);
// Create an `ImagenModel` instance with an Imagen "capability" model
final
model
=
ai
.
imagenModel
(
model:
'imagen-3.0-capability-001'
);
// This example assumes 'originalImage' is a pre-loaded Uint8List.
// In a real app, this might come from the user's device or a URL.
final
Uint8List
originalImage
=
Uint8List
(
0
);
// TODO: Load your original image data here.
// Provide the prompt describing the new background.
final
prompt
=
'space background'
;
try
{
// Use the editImage API to replace the background.
// Pass the original image, the prompt, and an editing configuration.
final
response
=
await
model
.
editImage
(
sources:
[
ImagenRawImage
(
originalImage
),
ImagenBackgroundMask
(),
// Use ImagenBackgroundMask() to auto-generate the mask.
],
prompt:
prompt
,
// Define the editing configuration for inpainting and background replacement.
config:
const
ImagenEditingConfig
(
editMode:
ImagenEditMode
.
inpaintInsertion
,
),
);
// Process the result.
if
(
response
.
images
.
isNotEmpty
)
{
final
editedImage
=
response
.
images
.
first
.
bytes
;
// Use the editedImage (a Uint8List) to display the image, save it, etc.
print
(
'Image successfully generated!'
);
}
else
{
// Handle the case where no images were generated.
print
(
'Error: No images were generated.'
);
}
}
catch
(
e
)
{
// Handle any potential errors during the API call.
print
(
'An error occurred:
$
e
'
);
}
Unity
Image editing with Imagen models isn't supported for Unity. Check back later this year!
Replace background using a provided mask
The following sample shows how to replace the background of an image — using a background mask defined in an image that you provide. You provide the original image, a text prompt, and the masked image.
Swift
Image editing with Imagen models isn't supported for Swift. Check back later this year!
Kotlin
To replace the background using a mask that you provide, specify ImagenRawMask
with the masked image. Use editImage()
and set the editing config to use ImagenEditMode.INPAINT_INSERTION
.
// Using this SDK to access Imagen models is a Preview release and requires opt-in
@OptIn
(
PublicPreviewAPI
::
class
)
suspend
fun
customizeImage
()
{
// Initialize the Vertex AI Gemini API backend service
// Optionally specify the location to access the model (for example, `us-central1`)
val
ai
=
Firebase
.
ai
(
backend
=
GenerativeBackend
.
vertexAI
(
location
=
"us-central1"
))
// Create an `ImagenModel` instance with an Imagen "capability" model
val
model
=
ai
.
imagenModel
(
"imagen-3.0-capability-001"
)
// This example assumes 'originalImage' is a pre-loaded Bitmap.
// In a real app, this might come from the user's device or a URL.
val
originalImage
:
Bitmap
=
TODO
(
"Load your original image Bitmap here"
)
// This example assumes 'maskImage' is a pre-loaded Bitmap that contains the masked area.
// In a real app, this might come from the user's device or a URL.
val
maskImage
:
Bitmap
=
TODO
(
"Load your masked image Bitmap here"
)
// Provide the prompt describing the new background.
val
prompt
=
"space background"
// Use the editImage API to replace the background.
// Pass the original image, the masked image, the prompt, and an editing configuration.
val
editedImage
=
model
.
editImage
(
referenceImages
=
listOf
(
ImagenRawImage
(
originalImage
.
toImagenInlineImage
()),
ImagenRawMask
(
maskImage
.
toImagenInlineImage
()),
// Use ImagenRawMask() to provide your own masked image.
),
prompt
=
prompt
,
// Define the editing configuration for inpainting and background replacement.
config
=
ImagenEditingConfig
(
ImagenEditMode
.
INPAINT_INSERTION
)
)
// Process the resulting 'editedImage' Bitmap, for example, by displaying it in an ImageView.
}
Java
To replace the background using a mask that you provide, specify ImagenRawMask
with the masked image. Use editImage()
and set the editing config to use ImagenEditMode.INPAINT_INSERTION
.
// Initialize the Vertex AI Gemini API backend service
// Optionally specify the location to access the model (for example, `us-central1`)
// Create an `ImagenModel` instance with an Imagen "capability" model
ImagenModel
imagenModel
=
FirebaseAI
.
getInstance
(
GenerativeBackend
.
vertexAI
(
"us-central1"
))
.
imagenModel
(
/* modelName */
"imagen-3.0-capability-001"
);
ImagenModelFutures
model
=
ImagenModelFutures
.
from
(
imagenModel
);
// This example assumes 'originalImage' is a pre-loaded Bitmap.
// In a real app, this might come from the user's device or a URL.
Bitmap
originalImage
=
null
;
// TODO("Load your original image Bitmap here");
// This example assumes 'maskImage' is a pre-loaded Bitmap that contains the masked area.
// In a real app, this might come from the user's device or a URL.
Bitmap
maskImage
=
null
;
// TODO("Load your masked image Bitmap here");
// Provide the prompt describing the new background.
String
prompt
=
"space background"
;
// Define the list of source images for the editImage call.
ImagenRawImage
rawOriginalImage
=
new
ImagenRawImage
(
ImagenInlineImageKt
.
toImagenInlineImage
(
originalImage
));
// Use ImagenRawMask() to provide your own masked image.
ImagenBackgroundMask
rawMaskedImage
=
new
ImagenRawMask
(
ImagenInlineImageKt
.
toImagenInlineImage
(
maskImage
));
ImagenEditingConfig
config
=
new
ImagenEditingConfig
();
// Use the editImage API to replace the background.
// Pass the original image, the masked image, the prompt, and an editing configuration.
Futures
.
addCallback
(
model
.
editImage
(
Arrays
.
asList
(
rawOriginalImage
,
rawMaskedImage
),
prompt
,
config
),
new
FutureCallback
<
ImagenGenerationResponse
>
()
{
@Override
public
void
onSuccess
(
ImagenGenerationResponse
result
)
{
if
(
result
.
getImages
().
isEmpty
())
{
Log
.
d
(
"ImageEditor"
,
"No images generated"
);
}
Bitmap
editedImage
=
((
ImagenInlineImage
)
result
.
getImages
().
get
(
0
)).
asBitmap
();
// Process and use the bitmap to display the image in your UI
}
@Override
public
void
onFailure
(
Throwable
t
)
{
// ...
}
},
Executors
.
newSingleThreadExecutor
());
Web
Image editing with Imagen models isn't supported for Web apps. Check back later this year!
Dart
To replace the background using a mask that you provide, specify ImagenRawMask
with the masked image. Use editImage()
and set the editing config to use ImagenEditMode.INPAINT_INSERTION
.
import
'dart:typed_data'
;
import
'package:firebase_ai/firebase_ai.dart'
;
import
'package:firebase_core/firebase_core.dart'
;
import
'firebase_options.dart'
;
// Initialize FirebaseApp
await
Firebase
.
initializeApp
(
options:
DefaultFirebaseOptions
.
currentPlatform
,
);
// Initialize the Vertex AI Gemini API backend service
// Optionally specify a location to access the model (for example, `us-central1`)
final
ai
=
FirebaseAI
.
vertexAI
(
location:
'us-central1'
);
// Create an `ImagenModel` instance with an Imagen "capability" model
final
model
=
ai
.
imagenModel
(
model:
'imagen-3.0-capability-001'
);
// This example assumes 'originalImage' is a pre-loaded Uint8List.
// In a real app, this might come from the user's device or a URL.
final
Uint8List
originalImage
=
Uint8List
(
0
);
// TODO: Load your original image data here.
// This example assumes 'maskImage' is a pre-loaded Uint8List that contains the masked area.
// In a real app, this might come from the user's device or a URL.
final
Uint8List
maskImage
=
Uint8List
(
0
);
// TODO: Load your masked image data here.
// Provide the prompt describing the new background.
final
prompt
=
'space background'
;
try
{
// Use the editImage API to replace the background.
// Pass the original image, the prompt, and an editing configuration.
final
response
=
await
model
.
editImage
(
sources:
[
ImagenRawImage
(
originalImage
),
ImagenRawMask
(
maskImage
),
// Use ImagenRawMask() to provide your own masked image.
],
prompt:
prompt
,
// Define the editing configuration for inpainting and background replacement.
config:
const
ImagenEditingConfig
(
editMode:
ImagenEditMode
.
inpaintInsertion
,
),
);
// Process the result.
if
(
response
.
images
.
isNotEmpty
)
{
final
editedImage
=
response
.
images
.
first
.
bytes
;
// Use the editedImage (a Uint8List) to display the image, save it, etc.
print
(
'Image successfully generated!'
);
}
else
{
// Handle the case where no images were generated.
print
(
'Error: No images were generated.'
);
}
}
catch
(
e
)
{
// Handle any potential errors during the API call.
print
(
'An error occurred:
$
e
'
);
}
Unity
Image editing with Imagen models isn't supported for Unity. Check back later this year!
Best practices and limitations
We recommend dilating the mask when editing an image. This can help smooth
the borders of an edit and make it seem more convincing. Generally, a dilation
value of 1% or 2% is recommended ( 0.01
or 0.02
).
Give feedback about your experience with Firebase AI Logic

