[BETA] Enable torch.compile ============================ .. warning:: The support for `torch.compile` is in beta. Not all models are compatible with this feature. Overview -------- OpenVINO™ Training Extensions now integrates the `torch.compile` feature from PyTorch, allowing users to optimize their models for better performance. This feature compiles the model's operations into optimized lower-level code, which can significantly improve execution speed and reduce memory usage. Benefits of torch.compile ------------------------- - **Performance Optimization**: Compiled models run faster by executing optimized low-level operations. - **Reduced Memory Footprint**: Optimized models can use less memory, which is beneficial for deploying models on resource-constrained devices. For more information on the benefits of `torch.compile`, refer to the official `PyTorch documentation `_. How to Use torch.compile in OpenVINO™ Training Extensions ---------------------------------------------------------- **Prepare OTXModel**: Ensure that model is compatible with `torch.compile`. When building the model, give the `torch_compile` option `True`. .. tab-set:: .. tab-item:: API .. code-block:: python from otx.algo.classification.multiclass_models.vit import VisionTransformerMulticlassCls model = VisionTransformerMulticlassCls(..., torch_compile=True) .. tab-item:: CLI .. code-block:: bash (otx) ...$ otx train ... --model.torch_compile True