Releases: roboflow/inference
v0.62.2
💪 Added
- Jetpack 6.2 Support by @alexnorell in #1730
- batch processing with cloud storage by @japrescott in #1726
- Parametrize modal function to enable conditional preloading of models by @grzegorz-roboflow in #1744
- Gemini workflow block v2 by @brunopicinin in #1753
🔧 Fixed
- lower bound confidence threshold was not applied due to incorrect condition by @grzegorz-roboflow in #1746
🏡 Maintenance
- improve rfdetr segmentation post processing by @hansent in #1747
- Proxy sam3 endpoint by @hansent in #1743
- Accept newer version of numpy by @PawelPeczek-Roboflow in #1745
Full Changelog: v0.62.0...v0.62.2
v0.62.0
💪 Added
♾️ Segment Anything 3 🤝 inference
sam-3-build-an-app.mp4
We are happy to share that newest Meta AI model - Segment Anything 3, just got added to inference 🔥
Staring from now, you can use the model directly in inference or explore various Roboflow Platform features powered by SAM3 - plenty of options are on the table :
- use the model to power your Workflow
- support your labelling efforts (or even offload the task to the model completely)
- build your model within few minutes with Roboflow Rapid
Our Blog 📖 provides further reading and is an excellent source of inspiration. You can also visit our docs 📝 providing instructions on how to use the model in the code.
PRs contributing to the release: #1729, #1731, #1732, #1704, #1734, #1736, #1737, #1738, #1739, #1740, #1741
Full Changelog: v0.61.0...v0.62.0
v0.61.0
💪 Added
- Api key passthrough/email - with this change @yeldarby enabled our users to send emails from workflows using Roboflow's email relay - change by @yeldarby in #1685
- WebRTC data decoupling - video channel and data channel are now created independently, making it possible to process video without streaming preview back to preserve bandwidth - change by @lrosemberg in #1700
- Dinov3 classifier by @mkaic in #1705
🔧 Fixed
- Fix exp integration tests / re-enable clip baseline tests by @hansent in #1716
- Add missing 'workflows' marker to pytest.ini by @grzegorz-roboflow in #1724
- Handle Supervision 0.27.0 by @grzegorz-roboflow in #1725
🏡 Maintenance
- Validate workflow accepts api key by @SolomonLake in #1697
- switch modal to app cls by @grzegorz-roboflow in #1707
- Parametrize gpu type by @grzegorz-roboflow in #1709
- Parametrize webrtc worker timeout by @grzegorz-roboflow in #1708
- Make GPU optional and allow user to run on CPU-only functions by @grzegorz-roboflow in #1713
- OPC-UA improvements for ignition by @sberan in #1706
- OPC-UI write block: support additional type conversions by @sberan in #1719
- Accept newest numpy in inference dependencies by @PawelPeczek-Roboflow in #1721
- Add shared thread-pool to HTTP API workflows runner by @PawelPeczek-Roboflow in #1717
- OPC-UA Writer Block: Convert incoming values to primitive type prior to setting value by @sberan in #1723
- Modal parametrization / webrtc turn config / perspective correction fix by @grzegorz-roboflow in #1722
New Contributors
Full Changelog: v0.60.0...v0.61.0
v0.60.0
What's Changed
- Serialize parent origin metadata for crops/slices by @brunopicinin in #1672
- Update perspective correction for dimensionality by @lou-roboflow in #1674
- "USE_INFERENCE_EXP_MODELS" by @hansent in #1657
- Add support for ingesting and processing rtmp stream remotely (on modal or process) and stream back processed frames over webrtc by @grzegorz-roboflow in #1675
- Add GitHub Actions workflow for Jetson 6.2.0 builds by @alexnorell in #1671
- uv lock inference-experimental by @hansent in #1677
- Add rtsp placeholder to webrtc_worker by @grzegorz-roboflow in #1683
- Set HF_HOME when running webrtc_worker on Modal by @grzegorz-roboflow in #1682
- Make webrtc_worker time out gracefully if WEBRTC_MODAL_FUNCTION_TIME_LIMIT is set by @grzegorz-roboflow in #1684
- add file lock for downloading core models by @hansent in #1680
- TRT Compilation of models in
inference-cliby @PawelPeczek-Roboflow in #1679 - Remove (currently broken)
trt-*-cu13dependencies frominference-expby @PawelPeczek-Roboflow in #1689 - Handle bursty rtsp by @grzegorz-roboflow in #1691
- Optimize bytes_to_unicode by @grzegorz-roboflow in #1676
- Drain rtsp queue by @grzegorz-roboflow in #1693
- InferenceExpObjectDetectionModelAdapter: dont optimize_for_inference by @hansent in #1694
Full Changelog: v0.59.1...v0.60.0
v0.59.1
💪 Added
📺 hosted webrtc
Thanks to @grzegorz-roboflow we have a great news 🔥
With this release of inference it is now possible to easily process video streams over webrtc with inference!
We prepared example script to showcase how to process streams from your python scripts. Start inference server, and then stream web-cam in order to start processing the stream and to see live preview produced by inference:
python examples/webrtc/webrtc_worker.py \
--workspace-id="<your workspace>" \
--workflow-id="<your workflow id>" \
--api-key="<your API key>" \
--inference-server-url="http://127.0.0.1:9001" \
--realtimeLocally stored files can also be streamed and processed:
python examples/webrtc/webrtc_worker.py \
--workspace-id="<your workspace>" \
--workflow-id="<your workflow id>" \
--api-key="<your API key>" \
--inference-server-url="http://127.0.0.1:9001" \
--source "/path/to/file.mp4" \
--realtimeThe processing can be performed on Modal simply export WEBRTC_MODAL_TOKEN_ID and WEBRTC_MODAL_TOKEN_SECRET with your credentials, and inference takes care of creating Modal app and processing stream on Modal function. All aspects of Modal execution are controlled through environmental variables. On-demand stream processing on CPU or GPU is extremely easy with this feature!
The feature is also available through our app https://app.roboflow.com:
Check out details #1612
Other new features
- Add RLE format support to SAM2 by @lrosemberg in #1651
- Seg preview workflow block by @hansent in #1647
🔧 Fixed
- Fix assertion in platform tests by @PawelPeczek-Roboflow in #1644
- Fix base detection docstrings default confidence and IOU by @dagleaves in #1660
- Fix ModelConfig incorrect confidence environment variable by @dagleaves in #1659
- Fix image dimensions in detections stitch block output by @brunopicinin in #1654
- Claude 3.5 is deprecated as of 28th of October 2025, replacing with Claude 4.5 by @grzegorz-roboflow in #1668
🏡 Maintenance
- Healthz and Readiness should not be authed by @bigbitbus in #1650
🥼 inference-experimental
- Add RFDetr Segmentation to inference-exp by @PawelPeczek-Roboflow in #1645
- Loose RFDetr Seg integration tests assertions for ONNX which seems to be quite numerically unstable by @PawelPeczek-Roboflow in #1646
- Fix assertions in RFDetr segmentation by @PawelPeczek-Roboflow in #1649
- Add fixes for TRT error handling by @PawelPeczek-Roboflow in #1665
- Bump version of
inference-expby @PawelPeczek-Roboflow in #1653 - TRT inplementation for more models by @PawelPeczek-Roboflow in #1666
- Fix issue with inference resolution not denoted correctly in RFDetr by @PawelPeczek-Roboflow in #1667
Full Changelog: v0.59.0...v0.59.1
v0.59.0
💪 Added
🍾 Workflows Execution Engine v1.7.0
This release refines how Workflow step errors are reported when interacting with the Roboflow platform. Client-side issues such as invalid API keys or model IDs now raise ClientCausedStepExecutionError with accurate HTTP status codes (400–404) instead of a generic StepExecutionError with HTTP 500.
Motivation
The change ensures that client misconfigurations are clearly distinguished from server-side errors, improving observability and making debugging easier.
Compatibility
Most users should not be affected. However, if an integration relies on the legacy error type or status code, it can be temporarily restored by setting:
DEFAULT_WORKFLOWS_STEP_ERROR_HANDLER=legacyThis update brings more accurate error reporting with minimal impact on existing workflows. See details of the change in the Workflows change log and in PR #1640
🥼 Changes inference-experimental
- Add changes to inference-exp deeplab sem-seg model by @PawelPeczek-Roboflow in #1629
- Adjust the name of yolov5 weights by @PawelPeczek-Roboflow in #1633
- Add tests for Yolov7 implementation by @PawelPeczek-Roboflow in #1634
- Add ability for inference-exp to parse modelVariant from RFAPI by @PawelPeczek-Roboflow in #1641
- Add PaliGemma 1 to inference-exp by @PawelPeczek-Roboflow in #1639
- YOLACT in inference experimental by @PawelPeczek-Roboflow in #1637
- Adjust YOLACT tests assertions for GPU tests to pass by @PawelPeczek-Roboflow in #1638
Full Changelog: v0.58.3...v0.59.0
v0.58.3
What's Changed
- Fix RFDetr issue with class remapping and ensure model cache paths sanitisation by @PawelPeczek-Roboflow in #1619
- Update Overlap block UI configuration by @brunopicinin in #1622
- Add changes to fix multi-class keypoints detection by @PawelPeczek-Roboflow in #1626
- Add implementation of yolov8-cls to inference-exp by @PawelPeczek-Roboflow in #1627
- bugfix: incorrect parsing of content-type breaks API_KEY authenticati… by @tonylampada in #1628
- Update size measurement video URL by @rfjames in #1625
- Fix JetPack 6.2 Dockerfile: Build ONNX Runtime from source with ARM64-compatible headers by @alexnorell in #1623
- Feat/include errors in usage by @grzegorz-roboflow in #1621
- Set workflow execution log level to debug by @bigbitbus in #1620
- bump version to 0.58.3 by @hansent in #1630
- Add Claude Haiku 4.5 to workflow block by @brunopicinin in #1632
New Contributors
Full Changelog: v0.58.2...v0.58.3
v0.58.2
🔧 Fixed
- Honor mask decode args for rf-detr seg model by @probicheaux in #1613
- Fix image preprocessing block to accept dynamic parameter references by @rf-elliot in #1616
🏡 Maintenance
- Report total remote execution time when running benchmarks by @grzegorz-roboflow in #1615
- add gcs/s3 example scripts by @japrescott in #1617
- Bump the version by @PawelPeczek-Roboflow in #1618
🏅 New Contributors
- @japrescott made their first contribution in #1617
- @rf-elliot made their first contribution in #1616
Full Changelog: v0.58.1...v0.58.2
v0.58.1
🔧 Fixed
Warning
Potential OOM errors in inference server as a result of bug introduced in v0.55.0
We've detected a bug that was in inference since 0.55.0 release - setting confidence to 0.0 was historically prevented globally in inference server. Since we wanted to remove that constraint, in 0.55.0 release, we make it possible to set such confidence threshold, but that turned out to lead to OOM issues with some models - namely instance segmentation ones, which in some cases may yield large amount of segmentation masks using wast amount of RAM. To provide quick-fix, we've added CONFIDENCE_LOWER_BOUND_OOM_PREVENTION env variable flag to control the minimal allowed confidence, set by default to 0.01.
Review the change here: #1611
🏡 Maintenance
- Add disable model monitoring on legacy requests by @rvirani1 in #1607
- Make triggering anchor optional in line_counter block by @grzegorz-roboflow in #1608
Full Changelog: v0.58.0...v0.58.1
v0.58.0
💪 Added
🤯 RFDetr Seg — Fast, Accurate Segmentation in Roboflow 💥
We’re excited to announce the addition of RF-DETR Seg (Preview), bringing real-time instance segmentation to the Roboflow platform. Thanks to the outstanding work of the RF research team, users can now:- Train and deploy RF-DETR Seg models directly in Roboflow
- Enjoy state-of-the-art accuracy with latency
- Run models seamlessly through Roboflow
inferenceor the rf-detr packages
This release pushes the boundary of fast, accurate segmentation — and it’s now available to explore in Roboflow. Check out our 📖 blog for more details and try out your RFDetr model in inference:
import os
import supervision as sv
from inference import get_model
from PIL import Image
from io import BytesIO
import requests
url = "https://media.roboflow.com/dog.jpeg"
image = Image.open(BytesIO(requests.get(url).content))
model = get_model("<your-rf-model-id>")
predictions = model.infer(image)[0]
detections = sv.Detections.from_inference(predictions)🔌 Improved
Full Changelog: v0.57.4...v0.58.0

