Code Monkey home page Code Monkey logo

christophreich1996 / cell-detr Goto Github PK

View Code? Open in Web Editor NEW
99.0 2.0 23.0 33.69 MB

Official and maintained implementation of the paper "Attention-Based Transformers for Instance Segmentation of Cells in Microstructures" [BIBM 2020].

Home Page: https://arxiv.org/pdf/2011.09763

License: MIT License

Python 74.92% C++ 1.41% Cuda 14.74% Jupyter Notebook 8.93%
bibm-2020 deep-learning attention instance-segmentation cell-segmentation synthetic-biology cell-detr transformer system-biology

cell-detr's Introduction

Cell-DETR: Attention-Based Transformers for Instance Segmentation of Cells in Microstructures

arXiv License: MIT

Tim Prangemeier, Christoph Reich & Heinz Koeppl

This repository includes the official and maintained implementation of the paper Attention-Based Transformers for Instance Segmentation of Cells in Microstructures (BIBM 2020).

drawingdrawingdrawingdrawingdrawing

Abstract

Detecting and segmenting object instances is a common task in biomedical applications. Examples range from detecting lesions on functional magnetic resonance images, to the detection of tumours in histopathological images and extracting quantitative single-cell information from microscopy imagery, where cell segmentation is a major bottleneck. Attention-based transformers are state-of-the-art in a range of deep learning fields. They have recently been proposed for segmentation tasks where they are beginning to outperforming other methods. We present a novel attention-based cell detection transformer (Cell-DETR) for direct end-to-end instance segmentation. While the segmentation performance is on par with a state-of-the-art instance segmentation method, Cell-DETR is simpler and faster. We showcase the method's contribution in a the typical use case of segmenting yeast in microstructured environments, commonly employed in systems or synthetic biology. For the specific use case, the proposed method surpasses the state-of-the-art tools for semantic segmentation and additionally predicts the individual object instances. The fast and accurate instance segmentation performance increases the experimental information yield for a posteriori data processing and makes online monitoring of experiments and closed-loop optimal experimental design feasible.

Architecture

Cell-DETR_Architecture
Architecture of the end-to-end instance segmentation network, with brightfield specimen image input and an instance segmentation prediction as output. The backbone CNN encoder extracts image features that then feed into both the transformer encoder-decoder for class and bounding box prediction, as well as to the CNN decoder for segmentation. The transformer encoded features, as well as the transformer decoded features, are feed into a multi-head-attention module and together with the image features from the CNN backbone feed into the CNN decoder for segmentation. Skip connections additionally bridge between the backbone CNN encoder and the CNN decoder. Input and output resolution is 128 × 128 pixels.

Dependencies

The Cell-DETR implementation uses multiple existing implementations. First, deformable convolutions v2 [2] are used based on the implementation of Dazhi Cheng. Second, pade activation units [4] are utilized base on the official implementation, by the authors. And third, the pixel adaptive convolutions [3] implementation by Nvidia is used. The pade activation unit implementation as well as the pixel adaptive convolution implementation are adopted slightly and are included in this repository. All dependencies can be installed by executing the following commands:

git clone https://github.com/ChristophReich1996/Cell-DETR.git
cd Cell-DETR
pip install --no-deps -r requirements.txt -f https://download.pytorch.org/whl/torch_stable.html
git clone https://github.com/chengdazhi/Deformable-Convolution-V2-PyTorch
cd Deformable-Convolution-V2-PyTorch
git checkout pytorch_1.0.0
python setup.py build install
cd ../pade_activation_unit/cuda
python setup.py build install

The transformer [5] implementation is based on the official implementation of DETR [1].

Usage

CELL-DETR can be trained, validated and testes by using the main.py script. The following command line arguments define what actions are performed.

python main.py {+ args}
Argument Default value Info
--train False Binary flag. If set training will be performed.
--val False Binary flag. If set validation will be performed.
--test False Binary flag. If set testing will be performed.
--cuda_devices "0" String of cuda device indexes to be used. Indexes must be separated by a comma
--data_parallel False Binary flag. If multi GPU training should be utilized set flag.
--cpu False Binary flag. If set all operations are performed on the CPU.
--epochs 200 Number of epochs to perform while training.
--lr_schedule False Binary flag. If set the learning rate will be reduced after epoch 50 and 100.
--ohem False Binary flag. If set online heard example mining is utilized.
--ohem_fraction 0.75 Ohem fraction to be applied when performing ohem.
--batch_size 4 Batch size to be utilized while training.
--path_to_data "trapped_yeast_cell_dataset" Path to dataset.
--augmentation_p 0.6 Probability that data augmentation is applied on training data sample.
--lr_main 1e-04 Learning rate of the detr model (excluding backbone).
--lr_backbone 1e-05 Learning rate of the backbone network.
--no_pac False Binary flag. If set no pixel adaptive convolutions will be utilized in the segmentation head.
--load_model "" Path to model to be loaded.
--dropout 0.0 Dropout factor to be used in model.
--three_classes False Binary flag, If set three classes (trap, cell of interest and add. cells) will be utilized.
--softmax False Binary flag, If set a softmax will be applied to the segmentation prediction instead sigmoid.
--only_train_segmentation_head_after_epoch 200 Number of epoch where only the segmentation head is trained.
--lr_segmentation_head 1e-05 Learning rate of the segmentation head, only applied when seg head is trained exclusively.
--no_deform_conv False Binary flag. If set no deformable convolutions will be utilized.
--no_pau False Binary flag. If set no pade activation unit is utilized, however, a leaky ReLU is utilized.

For training, validating and testing of the Cell-DETR B architecture run

python main.py --train --val --test --path_to_data "trapped_yeast_cell_dataset" --lr_schedule --batch_size 10 --data_parallel --cuda_devices "0, 1" --softmax

For training, validating and testing of the Cell-DETR A architecture run

python main.py --train --val --test --path_to_data "trapped_yeast_cell_dataset" --lr_schedule --batch_size 10 --data_parallel --cuda_devices "0, 1" --softmax --no_pac --no_deform_conv --no_pau

Trained Models

Our trained models (Cell-DETR A & B) are included in the folder trained_models.

To load and test the trained Cell-DETR A model run

python main.py --test --path_to_data "trapped_yeast_cell_dataset" --cuda_devices "0" --softmax --no_pac --no_deform_conv --no_pau --load_model "trained_models/Cell_DETR_A"

to load and test the trained Cell-DETR B model run

python main.py --test --path_to_data "trapped_yeast_cell_dataset" --cuda_devices "0" --softmax --load_model "trained_models/Cell_DETR_B"

Data

A few toy/test examples of the trapped yeast cell instance segmentation dataset are included in folder trapped_yeast_cell_dataset. The full dataset can be requested from the authors.

Results

Qualitative results

drawingdrawing
Example segmentations of our Cell-DETR B model.

Segmentation results

Model Dice Accuracy mIoU (mean over instances) Cell IoU
Cell-DETR A 0.92 0.96 0.84 0.83
Cell-DETR B 0.92 0.96 0.85 0.84

Bounding box results

Model MSE L1 IoU gIoU
Cell-DETR A 0.0006 0.016 0.81 0.80
Cell-DETR B 0.0005 0.016 0.81 0.81

Classification results

Model Accuracy
Cell-DETR A 1.0
Cell-DETR B 1.0

Citation

If you find this research useful in your work, please acknowledge it appropriately and cite the paper:

@article{prangemeier2020c,
        title={Attention-Based Transformers for Instance Segmentation of Cells in Microstructures},
        author={Prangemeier, Tim and Reich, Christoph and Koeppl, Heinz},
        booktitle={2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)},
        year={2020}
}

References

[1] @article{carion2020end,
        title={End-to-End Object Detection with Transformers},
        author={Carion, Nicolas and Massa, Francisco and Synnaeve, Gabriel and Usunier, Nicolas and Kirillov, Alexander and Zagoruyko, Sergey},
        journal={arXiv preprint arXiv:2005.12872},
        year={2020}
}
[2] @inproceedings{zhu2019deformable,
        title={Deformable convnets v2: More deformable, better results},
        author={Zhu, Xizhou and Hu, Han and Lin, Stephen and Dai, Jifeng},
        booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
        pages={9308--9316},
        year={2019}
}
[3] @inproceedings{su2019pixel,
        title={Pixel-adaptive convolutional neural networks},
        author={Su, Hang and Jampani, Varun and Sun, Deqing and Gallo, Orazio and Learned-Miller, Erik and Kautz, Jan},
        booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
        pages={11166--11175},
        year={2019}
}
[4] @article{molina2019pad,
        title={Pad$\backslash$'e Activation Units: End-to-end Learning of Flexible Activation Functions in Deep Networks},
        author={Molina, Alejandro and Schramowski, Patrick and Kersting, Kristian},
        journal={arXiv preprint arXiv:1907.06732},
        year={2019}
}
[5] @inproceedings{vaswani2017attention,
        title={Attention is all you need},
        author={Vaswani, Ashish and Shazeer, Noam and Parmar, Niki and Uszkoreit, Jakob and Jones, Llion and Gomez, Aidan N and Kaiser, {\L}ukasz and Polosukhin, Illia},
        booktitle={Advances in neural information processing systems},
        pages={5998--6008},
        year={2017}
}

cell-detr's People

Contributors

christophreich1996 avatar tim-prangemeier avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

cell-detr's Issues

Training without segmentation on my own dataset

Hi, I would like to ask if I can set the weight of segmentation loss and classification loss to 0?

My own dataset don't have segmentation label, and only have 1 class

does it work on your model?

Issues with Airspeed during installation

On the last line of the installation code
(venv) maxs@doraemon1:~/Cell-DETR/pade_activation_unit/cuda$ python setup.py build install

I get:
Traceback (most recent call last): File "setup.py", line 301, in <module> generate_cpp_module(fname='pau_cuda.cpp') File "setup.py", line 51, in generate_cpp_module content = file_content.merge(locals()) File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 95, in merge self.merge_to(namespace, output, loader) File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 105, in merge_to self.ensure_compiled() File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 100, in ensure_compiled self.root_element = TemplateBody(self.filename, self.content) File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 270, in __init__ self.parse() File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 1245, in parse self.block = self.next_element(Block) File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 314, in next_element element = element_spec(self.filename, self._full_text, self.end) File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 270, in __init__ self.parse() File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 1277, in parse FallthroughHashText))) File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 321, in next_element self.end) File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 270, in __init__ self.parse() File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 1033, in parse self.require_match(self.OPEN_PAREN, '(') File "/home/maxs/.virtualenvs/venv/lib/python3.6/site-packages/airspeed/__init__.py", line 308, in require_match raise self.syntax_error(expected) airspeed.TemplateSyntaxError: line 6, column 8: expected ( in define definition, got: CHECK_CUDA(x) AT_ASSERTM(x.type().i ... ...

I thought the issue was with the Airspeed library / cuda combination, but I've updated both. The setup.py file has an airspeed template that doesn't match the expectations of airspeed itself. Have you any idea where I should look for the problem? I can't find a single thing about this problem anywhere. Do I need a specific older version of Airspeed?

Thanks in advance for reading this message,
Max

RuntimeWarning: Mean of empty slice. return _methods._mean(a, axis=axis, dtype=dtype,

Epoch 5/10 Best val. mIoU=0.0000 Loss C.=0.2363 Loss BB.=0.4493 Loss Seg.=1.7465: 50% 1480/2960 [01:15<01:05, 22.48it/s]/usr/local/lib/python3.8/site-packages/numpy/core/fromnumeric.py:3432: RuntimeWarning: Mean of empty slice.
return _methods._mean(a, axis=axis, dtype=dtype,
/usr/local/lib/python3.8/site-packages/numpy/core/_methods.py:190: RuntimeWarning: invalid value encountered in double_scalars
ret = ret.dtype.type(ret / rcount)
Epoch 10/10 Best val. mIoU=0.0000 Loss C.=0.2336 Loss BB.=0.4569 Loss Seg.=1.2564: 100% 2960/2960 [02:31<00:00, 23.09it/s]/usr/local/lib/python3.8/site-packages/numpy/core/fromnumeric.py:3432: RuntimeWarning: Mean of empty slice.
return _methods._mean(a, axis=axis, dtype=dtype,
/usr/local/lib/python3.8/site-packages/numpy/core/_methods.py:190: RuntimeWarning: invalid value encountered in double_scalars
ret = ret.dtype.type(ret / rcount)
Epoch 10/10 Best val. mIoU=0.0000 Loss C.=0.2336 Loss BB.=0.4569 Loss Seg.=1.2564: 100% 2960/2960 [02:40<00:00, 23.09it/s]/usr/local/lib/python3.8/site-packages/numpy/core/fromnumeric.py:3432: RuntimeWarning: Mean of empty slice.
return _methods._mean(a, axis=axis, dtype=dtype,
/usr/local/lib/python3.8/site-packages/numpy/core/_methods.py:190: RuntimeWarning: invalid value encountered in double_scalars
ret = ret.dtype.type(ret / rcount)
Epoch 10/10 Best val. mIoU=0.0000 Loss C.=0.2336 Loss BB.=0.4569 Loss Seg.=1.2564: 100% 2960/2960 [02:41<00:00, 18.29it/s]
/usr/local/lib/python3.8/site-packages/numpy/core/fromnumeric.py:3432: RuntimeWarning: Mean of empty slice.
return _methods._mean(a, axis=axis, dtype=dtype,

about the test of the network

Hello, when I was following your guidance to do tests on the functions of the network, I met such a problem:
2022-10-19 20-41-06 的屏幕截图
2022-10-19 20-41-11 的屏幕截图
It seems that we still need the data for training when applying the trained models. So are there any problems with our test, or we should download the whole dataset before conducting this test?
Meanwhile, since all of the experiments in your essay were conducted on images of 128128 pixels, however, the images we achieved in our experiments had a size of 20482048 pixels, which were much larger than those of the dataset. When adopting Cell-DETR to our further research, are the images too large for the network to deal with, as the complexity of a Transformer corresponds to O(n^2d)? Were there any tests or experiments on the images with larger sizes?
Thank you!

IndexError: tuple index out of range

Traceback (most recent call last):
File "main.py", line 150, in
model_wrapper.test()
File "/usr/local/lib/python3.8/site-packages/torch/autograd/grad_mode.py", line 49, in decorate_no_grad
return func(*args, **kwargs)
File "/content/drive/MyDrive/Cell-DETR/model_wrapper.py", line 421, in test
misc.plot_instance_segmentation_labels(
File "/content/drive/MyDrive/Cell-DETR/misc.py", line 435, in plot_instance_segmentation_labels
instances_map += np.array(colors_traps[min(counter_trap_instance, len(colors_cells) - 1)]).reshape(1, 1, 3)
IndexError: tuple index out of range

nvcc fatal : Unsupported gpu architecture 'compute_75'

python setup.py build install
something error in this step


gcc -pthread -B /home/szp/anaconda3/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -DWITH_CUDA -I/home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/TH -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/THC -I/home/szp/anaconda3/include/python3.8 -c /home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src/cpu/deform_cpu.cpp -o build/temp.linux-x86_64-3.8/home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src/cpu/deform_cpu.o -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="gcc" -DPYBIND11_STDLIB="libstdcpp" -DPYBIND11_BUILD_ABI="cxxabi1011" -DTORCH_EXTENSION_NAME=DCN -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++14
cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++
gcc -pthread -B /home/szp/anaconda3/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -DWITH_CUDA -I/home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/TH -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/THC -I/home/szp/anaconda3/include/python3.8 -c /home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src/cpu/deform_psroi_pooling_cpu.cpp -o build/temp.linux-x86_64-3.8/home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src/cpu/deform_psroi_pooling_cpu.o -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="gcc" -DPYBIND11_STDLIB="libstdcpp" -DPYBIND11_BUILD_ABI="cxxabi1011" -DTORCH_EXTENSION_NAME=DCN -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++14
cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++
gcc -pthread -B /home/szp/anaconda3/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -DWITH_CUDA -I/home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/TH -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/THC -I/home/szp/anaconda3/include/python3.8 -c /home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src/cpu/modulated_deform_cpu.cpp -o build/temp.linux-x86_64-3.8/home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src/cpu/modulated_deform_cpu.o -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="gcc" -DPYBIND11_STDLIB="libstdcpp" -DPYBIND11_BUILD_ABI="cxxabi1011" -DTORCH_EXTENSION_NAME=DCN -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++14
cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++
/usr/bin/nvcc -DWITH_CUDA -I/home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/TH -I/home/szp/anaconda3/lib/python3.8/site-packages/torch/include/THC -I/home/szp/anaconda3/include/python3.8 -c /home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src/cuda/deform_psroi_pooling_cuda.cu -o build/temp.linux-x86_64-3.8/home/szp/Cell-DETR/Deformable-Convolution-V2-PyTorch/src/cuda/deform_psroi_pooling_cuda.o -D__CUDA_NO_HALF_OPERATORS
-D__CUDA_NO_HALF_CONVERSIONS
-D__CUDA_NO_HALF2_OPERATORS
--expt-relaxed-constexpr --compiler-options '-fPIC' -DCUDA_HAS_FP16=1 -D__CUDA_NO_HALF_OPERATORS
-D__CUDA_NO_HALF_CONVERSIONS
_ -D__CUDA_NO_HALF2_OPERATORS__ -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=DCN -D_GLIBCXX_USE_CXX11_ABI=0 -gencode=arch=compute_75,code=sm_75 -std=c++14
nvcc fatal : Unsupported gpu architecture 'compute_75'
error: command '/usr/bin/nvcc' failed with exit status 1

Preparing the train and validation data in .pt format from coco format json file

Hi Christoph,

Thank you so much for providing the code and the explanations. It is really a brilliant work.
I would like to use the Cell-DETR model for the detection and segmentation of a single object in order to study the DETR.
However, I met some problem when I prepared the train and validation data.

First step
I completed image annotations by the tool Labelme. This tool can convert to VOC-format or to COCO-format.(Labelme convert format details)

Second step
After complete annotations, I convert all the related files into a single json file with COCO json format.
(View the example json file here)

However, I'm really confused about how to prepare the data structured in bounding boxes, classes, inputs, and instances.
As for the input images, the following method may work.

import torchvision.transforms
pil_to_tensor = transforms.ToTensor()(img).unsqueeze_(0)

As for the bounding boxes, classes, and instances, should I manually edit and save as tensors in .pt files?
Could you please elaborate the above process?
Do you have any idea how to complete preparing the train and validation data in batch?
Would you please share the details how you prepare the data structured in bounding boxes, classes, inputs, and instances?
Thanks in advance.

Best regards,
Simon

ModulatedDeformConvPack not found!

from modules.modulated_deform_conv import ModulatedDeformConvPack

I wasn't able to import this package. I tried build from [https://github.com/chengdazhi/Deformable-Convolution-V2-PyTorch]
and the modules used to build DCN is deprecated.
Please help with this issue and it would be nice if you upload that modules directory along with latest stable build script for DCN.

Adapting model to more labels

Hi there,
first of all thank you so much for providing the code to your paper and for the clear explanations.

I want to use the Cell-DETR model for a use-case with more than 20 labels per image.
I already got everything running, but it seems like that the model only predicts 12 bounding boxes and masks and the predicted boxes are also all located on one side of the image.

Do you have any idea where that might come from? Is there an additional hardcoded parameter in the architecture that I need to change?

Best,
Lisa

No module named 'modules'

Hello,I found an issue when running:
File "main.py", line 70, in
from modules.modulated_deform_conv import ModulatedDeformConvPack
ModuleNotFoundError: No module named 'modules'

Enquiry about validate and test methods in ModelWrapper

Hello,

Thanks for your interesting paper and for sharing the code.

Could I ask for test method in (model_wrapper.py), there is a conversion of the bounding box coordinates format from (xc yc w h)-> (x0 y0 x1 y1) (lines 351- 354) when calculating the test metrics for the bounding boxes but why is it not the case for the validate method (lines 222-224)?

Thank you!
Charlene

Failed to setup Pau_cuda

I failed at setup.py in cuda/

error message:

gcc-6.3.0: No such file or directory
nvcc fatal : Failed to preprocess host compiler properties.
error: command '/usr/local/cuda/bin/nvcc' failed with exit status 1

How to prepare data in .pt format!

Hi,
Please help me to prepare the data in .pt format.
My data is in coco format and I run maskRCNN on it.
I need to try with your model also.
Can you please support converting the data to .pt format?

Thanks
Mostafa

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.