Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

A novel region proposal network for more general object detection ( including scene text detection ).

NotificationsYou must be signed in to change notification settings

HCIILAB/DeRPN

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

34 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DeRPN is a novel region proposal network which concentrates on improving the adaptivity of current detectors.The paper is availablehere.

Recent Update

· Mar. 13, 2019: The DeRPN pretrained models are added.

· Jan. 25, 2019: The code is released.

Contact Us

Welcome to improve DeRPN together. For any questions, please feel free to contact Lele Xie (xie.lele@mail.scut.edu.cn) or Prof. Jin (eelwjin@scut.edu.cn).

Citation

If you find DeRPN useful to your research, please consider citing our paper as follow:

@article{xie2019DeRPN,  title     = {DeRPN: Taking a further step toward more general object detection},  author    = {Lele Xie, Yuliang Liu, Lianwen Jin*, Zecheng Xie}  joural    = {AAAI}  year      = {2019}}

Main Results

Note: The reimplemented results are slightly different from those presented in the paper for different training settings, but the conclusions are still consistent. For example, this code doesn't use multi-scale training which should boost the results for both DeRPN and RPN.

COCO-Text

training data: COCO-Text train

test data: COCO-Text test

networkAP@0.5recall@0.5AP@0.75recall@0.75
RPN+Faster R-CNNVGG1632.4852.547.4017.59
DeRPN+Faster R-CNNVGG1647.3970.4611.0525.12
RPN+R-FCNResNet-10137.7154.3513.1722.21
DeRPN+R-FCNResNet-10148.6271.3013.3727.57

Pascal VOC

training data: VOC 07+12 trainval

test data: VOC 07 test

Inference time is evaluated on one TITAN XP GPU.

networkinference timeAP@0.5AP@0.75AP
RPN+Faster R-CNNVGG1664 ms75.5342.0842.60
DeRPN+Faster R-CNNVGG1665 ms76.1744.9743.84
RPN+R-FCNResNet-10185 ms78.8754.3050.04
DeRPN+R-FCN (900) *ResNet-10184 ms79.2154.4350.28

( "*": On Pascal VOC dataset, we found that it is more suitable to train the DeRPN+R-FCN model with 900 proposals. For other experiments, we use the default proposal number to train the models, i.e., 2000 proposals fro Faster R-CNN, 300 proposals for R-FCN. )

MS COCO

training data: COCO 2017 train

test data: COCO 2017 test/val

test setnetworkAPAP50AP75APSAPMAPL
RPN+Faster R-CNNVGG1624.245.423.77.626.637.3
DeRPN+Faster R-CNNVGG1625.547.225.210.327.936.7
RPN+R-FCNResNet-10127.747.929.010.130.240.1
DeRPN+R-FCNResNet-10128.449.029.511.131.740.5
val setnetworkAPAP50AP75APSAPMAPL
RPN+Faster R-CNNVGG1624.145.023.87.627.837.8
DeRPN+Faster R-CNNVGG1625.547.325.09.928.837.8
RPN+R-FCNResNet-10127.848.128.810.431.242.5
DeRPN+R-FCNResNet-10128.448.529.511.532.942.0

Getting Started

  1. Requirements
  2. Installation
  3. Preparation for Training & Testing
  4. Usage

Requirements

  1. Cuda 8.0 and cudnn 5.1.
  2. Some python packages: cython, opencv-python, easydict et. al. Simply install them if your system misses these packages.
  3. Configure the caffe according to your environment (Caffe installation instructions). As the code requires pycaffe, caffe should be built with python layers. In Makefile.config, make sure to uncomment this line:
WITH_PYTHON_LAYER := 1
  1. An NVIDIA GPU with more than 6GB is required for ResNet-101.

Installation

  1. Clone the DeRPN repository

    git clone https://github.com/HCIILAB/DeRPN.git
  2. Build the Cython modules

    cd$DeRPN_ROOT/libmake
  3. Build caffe and pycaffe

    cd$DeRPN_ROOT/caffemake -j8&& make pycaffe

Preparation for Training & Testing

Dataset

  1. Download the datasets ofPascal VOC 2007 & 2012,MS COCO 2017 andCOCO-Text.

  2. You need to put these datasets under the $DeRPN_ROOT/data folder (with symlinks).

  3. For COCO-Text, the folder structure is as follow:

    $DeRPN_ROOT/data/coco_text/images/train2014$DeRPN_ROOT/data/coco_text/images/val2014$DeRPN_ROOT/data/coco_text/annotations# train2014, val2014, and annotations are symlinks from /pth_to_coco2014/train2014,# /pth_to_coco2014/val2014 and /pth_to_coco2014/annotations2014/, respectively.
  4. For COCO, the folder structure is as follow:

    $DeRPN_ROOT/data/coco/images/train2017$DeRPN_ROOT/data/coco/images/val2017$DeRPN_ROOT/data/coco/images/test-dev2017$DeRPN_ROOT/data/coco/annotations# the symlinks are similar to COCO-Text
  5. For Pascal VOC, the folder structure is as follow:

    $DeRPN_ROOT/data/VOCdevkit2007$DeRPN_ROOT/data/VOCdevkit2012#VOCdevkit2007 and VOCdevkit2012 are symlinks from $VOCdevkit whcich contains VOC2007 and VOC2012.

Pretrained models

Please download the ImageNet pretrained models (VGG16 andResNet-101, password: k4z1), and put them under

$DeRPN_ROOT/data/imagenet_models

We also provide the DeRPN pretrained modelshere (password: fsd8).

Usage

cd$DeRPN_ROOT./experiments/scripts/faster_rcnn_derpn_end2end.sh [GPU_ID] [NET] [DATASET]# e.g., ./experiments/scripts/faster_rcnn_derpn_end2end.sh 0 VGG16 coco_text

Copyright

This code is free to the academic community for research purpose only. For commercial purpose usage, please contact Dr. Lianwen Jin:lianwen.jin@gmail.com.

About

A novel region proposal network for more general object detection ( including scene text detection ).

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors2

  •  
  •  

[8]ページ先頭

©2009-2026 Movatter.jp