HTTP Endpoints for Agents and AI Workflows
Tensorlake automatically creates an HTTP endpoint when you deploy your agents and workflows. This makes it easy to integrate them with applications and other systems.Autoscaling for High Throughput Agents
Agents often run much longer than a typical web request. Tensorlake scales function containers based on concurrent requests, so you can run many agent sessions in parallel without operating queues or orchestrators.Built-in Durable Execution for Recovering from Failures
A single agent request can incur many calls to tools and LLMs. Failures are common at scale. Durable execution means retries can resume from the point of failure instead of starting over, which reduces cost and time to complete the task.Sandboxes for Running Untrusted Code
Code generation is often the most reliable way to get deterministic results from LLMs, like answering questions to mathematical problems, or when information is behind an HTTP API. Running untrusted code requires sandbox infrastructure to ensure the code is safe to run. Tensorlake functions run in sandboxes automatically, so you have sandbox infrastructure built into the platform.Tracing and Observability
Understanding the path of a request through the agentic application is crucial for debugging and observability. Every function is automatically traced in your application. You can observe logs, exceptions, timing information, and cold starts by looking at the execution timeline.Distributed Execution for Complex Applications
Applications can be composed of multiple functions with different compute requirements and dependencies. Tensorlake builds function images and runs functions in separate containers while serving requests transparently, without you wiring RPC calls or queues. Function-to-function calls are durable by default, so if a step fails, the request can resume from the last successful step.Declarative Concurrency Control
Tensorlake functions processes one request at a time by default. But, in some cases you may want to process multiple requests concurrently in the same container. Use themax_concurrency setting to allow up to N concurrent requests per container. This is useful when your agent loop is I O bound and you want to reuse a warm container across multiple sessions.
File Systems for Context Offloading
State-of-the-art agents use the filesystem to offload context and to download and inspect large files without loading them fully into memory. You can attach ephemeral disk to your functions by specifying the disk size.Declarative Dependency Management
Tensorlake uses real containers to run your functions, so you can configure the base image and install any dependencies you need.Map-Reduce for On-demand Large-scale Data Processing
Use Map-Reduce for on-demand processing of large volumes of data. Fan out work across a list in parallel, then aggregate the results.Next Steps
Jump straight into building a serverless agentic code interpreter in under 5 minutes.