Reusable containers
By default, each task execution in Flyte and Union runs in a fresh container instance that is created just for that execution and then discarded. With reusable containers, the same container can be reused across multiple executions and tasks. This approach reduces start up overhead and improves resource efficiency.
The reusable container feature is only available when running your Flyte code on a Union backend.
How It Works
With reusable containers, the system maintains a pool of persistent containers that can handle multiple task executions.
When you configure a TaskEnvironment with a ReusePolicy, the system does the following:
- Creates a pool of persistent containers.
- Routes task executions to available container instances.
- Manages container lifecycle with configurable timeouts.
- Supports concurrent task execution within containers (for async tasks).
- Preserves the Python execution environment across task executions, allowing you to maintain state through global variables.
Basic Usage
The reusable containers feature currently requires a dedicated runtime library
(
unionai-reuse) to be installed in the task image used by the reusable task.
You can add this library to your task image using the flyte.Image.with_pip_packages method, as shown below.
This library only needs to be added to the task image.
It does not need to be installed in your local development environment.
Enable container reuse by adding a ReusePolicy to your TaskEnvironment:
# /// script
# requires-python = "==3.13"
# dependencies = [
# "flyte>=2.0.0b25",
# ]
# main = "main"
# params = "n=500"
# ///
import flyte
from datetime import timedelta
# Currently required to enable resuable containers
reusable_image = flyte.Image.from_debian_base().with_pip_packages("unionai-reuse>=0.1.3")
env = flyte.TaskEnvironment(
name="reusable-env",
resources=flyte.Resources(memory="1Gi", cpu="500m"),
reusable=flyte.ReusePolicy(
replicas=2, # Create 2 container instances
concurrency=1, # Process 1 task per container at a time
scaledown_ttl=timedelta(minutes=10), # Individual containers shut down after 5 minutes of inactivity
idle_ttl=timedelta(hours=1) # Entire environment shuts down after 30 minutes of no tasks
),
image=reusable_image # Use the container image augmented with the unionai-reuse library.
)
@env.task
async def compute_task(x: int) -> int:
return x * x
@env.task
async def main() -> list[int]:
# These tasks will reuse containers from the pool
results = []
for i in range(10):
result = await compute_task(i)
results.append(result)
return results
if __name__ == "__main__":
flyte.init_from_config()
r = flyte.run(main)
print(r.name)
print(r.url)
r.wait()ReusePolicy parameters
The ReusePolicy class controls how containers are managed in a reusable environment:
flyte.ReusePolicy(
replicas: typing.Union[int, typing.Tuple[int, int]],
concurrency: int,
scaledown_ttl: typing.Union[int, datetime.timedelta],
idle_ttl: typing.Union[int, datetime.timedelta]
)replicas: Container pool size
Controls the number of container instances in the reusable pool:
- Fixed size:
replicas=3Creates exactly 3 container instances. These 3 replicas will be shutdown afteridle_ttlexpires. - Auto-scaling:
replicas=(2, 5)Starts with 2 containers and can scale up to 5 based on demand.- If the task is running on 2 replicas and demand drops to zero then these 2 containers will be shutdown after
idle_ttlexpires. - If the task is running on 2 replicas and demand increases, new containers will be created up to the maximum of 5.
- If the task is running on 5 replicas and demand drops, container 5 will be shutdown after
scaledown_ttlexpires. - If demand drops again, container 4 will be also shutdown after another period of
scaledown_ttlexpires.
- If the task is running on 2 replicas and demand drops to zero then these 2 containers will be shutdown after
- Resource impact: Each replica consumes the full resources defined in
TaskEnvironment.resources.
# Fixed pool size
fixed_pool_policy = flyte.ReusePolicy(
replicas=3,
concurrency=1,
scaledown_ttl=timedelta(minutes=10),
idle_ttl=timedelta(hours=1)
)
# Auto-scaling pool
auto_scaling_policy = flyte.ReusePolicy(
replicas=(1, 10),
concurrency=1,
scaledown_ttl=timedelta(minutes=10),
idle_ttl=timedelta(hours=1)
)concurrency: Tasks per container
Controls how many tasks can execute simultaneously within a single container:
- Default:
concurrency=1(one task per container at a time). - Higher concurrency:
concurrency=5allows 5 tasks to run simultaneously in each container. - Total capacity:
replicas × concurrency= maximum concurrent tasks across the entire pool.
# Sequential processing (default)
sequential_policy = flyte.ReusePolicy(
replicas=2,
concurrency=1, # One task per container
scaledown_ttl=timedelta(minutes=10),
idle_ttl=timedelta(hours=1)
)
# Concurrent processing
concurrent_policy = flyte.ReusePolicy(
replicas=2,
concurrency=5, # 5 tasks per container = 10 total concurrent tasks
scaledown_ttl=timedelta(minutes=10),
idle_ttl=timedelta(hours=1)
)idle_ttl vs scaledown_ttl: Container lifecycle
These parameters work together to manage container lifecycle at different levels:
idle_ttl: Environment timeout
- Scope: Controls the entire reusable environment infrastructure.
- Behavior: When there are no active or queued tasks, the entire environment scales down after
idle_ttlexpires. - Purpose: Manages the lifecycle of the entire container pool.
- Typical values: 1-2 hours, or
Nonefor always-on environments
scaledown_ttl: Individual container timeout
- Scope: Controls individual container instances.
- Behavior: When a container finishes a task and becomes inactive, it will be terminated after
scaledown_ttlexpires. - Purpose: Prevents resource waste from inactive containers.
- Typical values: 5-30 minutes for most workloads.
from datetime import timedelta
lifecycle_policy = flyte.ReusePolicy(
replicas=3,
concurrency=2,
scaledown_ttl=timedelta(minutes=10), # Individual containers shut down after 10 minutes of inactivity
idle_ttl=timedelta(hours=1) # Entire environment shuts down after 1 hour of no tasks
)Understanding parameter relationships
The four ReusePolicy parameters work together to control different aspects of container management:
reuse_policy = flyte.ReusePolicy(
replicas=4, # Infrastructure: How many containers?
concurrency=3, # Throughput: How many tasks per container?
scaledown_ttl=timedelta(minutes=10), # Individual: When do idle containers shut down?
idle_ttl=timedelta(hours=1) # Environment: When does the whole pool shut down?
)
# Total capacity: 4 × 3 = 12 concurrent tasks
# Individual containers shut down after 10 minutes of inactivity
# Entire environment shuts down after 1 hour of no tasksKey relationships
- Total throughput =
replicas × concurrency - Resource usage =
replicas × TaskEnvironment.resources - Cost efficiency: Higher
concurrencyreduces container overhead, morereplicasprovides better isolation - Lifecycle management:
scaledown_ttlmanages individual containers,idle_ttlmanages the environment
Simple example
Here is a simple, but complete, example of reuse with concurrency
First, import the needed modules, set upf logging:
import asyncio
import logging
import flyte
logger = logging.getLogger(__name__)
logger.setLevel(logging.DEBUG)env = flyte.TaskEnvironment(
name="reuse_concurrency",
resources=flyte.Resources(cpu=1, memory="1Gi"),
reusable=flyte.ReusePolicy(
replicas=2,
idle_ttl=60,
concurrency=100,
scaledown_ttl=60,
),
image=flyte.Image.from_debian_base().with_pip_packages("unionai-reuse==0.1.5b0", pre=True),
)reuse_concurrency task (the main driver task of the workflow) and the noop task that will be executed multiple times reusing the same containers:@env.task
async def noop(x: int) -> int:
logger.debug(f"Task noop: {x}")
return x
@env.task
async def main(n: int = 50) -> int:
coros = [noop(i) for i in range(n)]
results = await asyncio.gather(*coros)
return sum(results)python reuse_concurrency.py to see it in action:if __name__ == "__main__":
flyte.init_from_config()
r = flyte.run(main, n=500)
print(r.name)
print(r.url)
r.wait()