Skip to content

Commit

Permalink
- reformat
Browse files Browse the repository at this point in the history
  • Loading branch information
MarcelRosier committed Jan 16, 2024
1 parent 0c8b34a commit 693fb82
Showing 1 changed file with 7 additions and 27 deletions.
34 changes: 7 additions & 27 deletions Tutorial.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -72,51 +72,31 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"__Minimal example__"
"__Minimal example__\n",
"\n",
"Logging will be messed up when used from a juypter notebook."
]
},
{
"cell_type": "code",
"execution_count": 1,
"execution_count": 6,
"metadata": {},
"outputs": [
{
"name": "stderr",
"output_type": "stream",
"text": [
"2024-01-16 20:13:13 INFO: Initialized AuroraGPUInferer with config: AuroraInfererConfig(output_mode=<DataMode.NIFTI_FILE: 'NIFTI_FILEPATH'>, log_level=20, tta=False, sliding_window_batch_size=1, workers=0, threshold=0.5, sliding_window_overlap=0.5, crop_size=(192, 192, 32), model_selection=<ModelSelection.BEST: 'best'>, include_whole_network_in_numpy_output_mode=False, include_metastasis_network_in_numpy_output_mode=False)\n",
"2024-01-16 20:13:13 INFO: Using device: cuda\n",
"2024-01-16 20:13:13 INFO: Logging to: test_output/segmentation.nii.log\n",
"2024-01-16 20:13:13 INFO: Running inference on cuda\n",
"2024-01-16 20:13:13 INFO: Successfully validated input images. Input mode: NIFTI_FILEPATH\n",
"2024-01-16 20:13:13 INFO: Received files: T1: True, T1C: False, T2: False, FLAIR: False\n",
"2024-01-16 20:13:13 INFO: Inference mode: t1-o\n",
"2024-01-16 20:13:13 INFO: No loaded compatible model found. Loading Model and weights\n",
"2024-01-16 20:13:13 INFO: Setting up Dataloader\n"
]
},
{
"name": "stdout",
"output_type": "stream",
"text": [
"BasicUNet features: (32, 32, 64, 128, 256, 32).\n"
]
},
{
"name": "stderr",
"output_type": "stream",
"text": [
"2024-01-16 20:13:20 INFO: Saved segmentation to test_output/segmentation.nii.gz\n",
"2024-01-16 20:13:20 INFO: Finished inference \n",
"\n"
]
}
],
"source": [
"from brainles_aurora.inferer.inferer import AuroraGPUInferer, AuroraInferer\n",
"from brainles_aurora.inferer.dataclasses import AuroraInfererConfig\n",
"\n",
"config = AuroraInfererConfig(tta=False) # disable tta for faster inference in this showcase\n",
"config = AuroraInfererConfig(\n",
" tta=False\n",
") # disable tta for faster inference in this showcase\n",
"\n",
"# If you don-t have a GPU that supports CUDA use the CPU version: AuroraInferer(config=config)\n",
"inferer = AuroraGPUInferer(config=config)\n",
Expand Down

0 comments on commit 693fb82

Please sign in to comment.