diff --git a/examples/sam3_agent.ipynb b/examples/sam3_agent.ipynb index 3a1f9a3b..7b825244 100644 --- a/examples/sam3_agent.ipynb +++ b/examples/sam3_agent.ipynb @@ -53,7 +53,8 @@ "torch.autocast(\"cuda\", dtype=torch.bfloat16).__enter__()\n", "\n", "# inference mode for the whole notebook. Disable if you need gradients\n", - "torch.inference_mode().__enter__()" + "_inference_mode = torch.inference_mode()\n", + "_inference_mode.__enter__()" ] }, { diff --git a/examples/sam3_image_batched_inference.ipynb b/examples/sam3_image_batched_inference.ipynb index dddbfd83..1cd9163e 100644 --- a/examples/sam3_image_batched_inference.ipynb +++ b/examples/sam3_image_batched_inference.ipynb @@ -67,7 +67,7 @@ }, { "cell_type": "code", - "execution_count": 8, + "execution_count": null, "metadata": {}, "outputs": [], "source": [ @@ -81,7 +81,8 @@ "torch.autocast(\"cuda\", dtype=torch.bfloat16).__enter__()\n", "\n", "# inference mode for the whole notebook. Disable if you need gradients\n", - "torch.inference_mode().__enter__()\n" + "_inference_mode = torch.inference_mode()\n", + "_inference_mode.__enter__()" ] }, { @@ -650,7 +651,7 @@ "fileUid": "76928cb6-3532-4024-bafd-4d3a609dfe2a", "isAdHoc": false, "kernelspec": { - "display_name": "Python 3 (ipykernel)", + "display_name": "sam3", "language": "python", "name": "python3" }, @@ -664,7 +665,7 @@ "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", - "version": "3.12.11" + "version": "3.12.12" } }, "nbformat": 4, diff --git a/examples/sam3_image_interactive.ipynb b/examples/sam3_image_interactive.ipynb index 7e1e7666..5961613a 100644 --- a/examples/sam3_image_interactive.ipynb +++ b/examples/sam3_image_interactive.ipynb @@ -60,7 +60,7 @@ }, { "cell_type": "code", - "execution_count": 5, + "execution_count": null, "id": "13325376-658b-48d6-8528-2a006f223d44", "metadata": {}, "outputs": [], @@ -75,7 +75,8 @@ "torch.autocast(\"cuda\", dtype=torch.bfloat16).__enter__()\n", "\n", "# inference mode for the whole notebook. Disable if you need gradients\n", - "torch.inference_mode().__enter__()" + "_inference_mode = torch.inference_mode()\n", + "_inference_mode.__enter__()" ] }, {