Code Monkey home page Code Monkey logo

Comments (40)

daixiaogang avatar daixiaogang commented on August 25, 2024 2

@SeokjuLee ,What are the news about releasing dataset?

from vpgnet.

BAILOOL avatar BAILOOL commented on August 25, 2024

@sunpeng1996 , Usually the segmentation is performed pixel-wise, meaning that every pixel has a specific class. In this paper, we have utilized grid-wise annotation which is simply a group of pixels in a grid cell. Specifically, the grid cell size is 8x8 pixels. Thus, the segmentation is performed on grid cells rather than individual pixels.

Regarding the dataset, please check Readme file ("Dataset contact" section) for the dataset questions. It turns out that Samsung wants to have an exclusive control who gets the dataset. Good luck!

from vpgnet.

hexiangquan avatar hexiangquan commented on August 25, 2024

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@hexiangquan The forward pass time for whole networks (4-task: grid, objectness, multi-label, vpp) is about 30ms on a single Titan X.

from vpgnet.

wsz912 avatar wsz912 commented on August 25, 2024

Sorry, Sir. I can't send the email to [email protected]. There is a problem that

Could not connect to SMTP host: 143.248.39.8, port: 25;
nested exception is:
java.net.ConnectException: Connection timed out

I want to apply for the dataset. What materials should I provide? Where and how can I answer the dataset questions?

I am looking forward to your reply.

Best wishes.

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@wsz1029 Sorry for the mail problem. Our lab server was attacked and currently repairing.. The dataset is preserved by Samsung. Please contact Tae-Hee Lee ([email protected]), Hyun Seok Hong ([email protected]), and Seung-Hoon Han ([email protected]).

from vpgnet.

daixiaogang avatar daixiaogang commented on August 25, 2024

@BAILOOL ,how did you do the grid-wise annotation?How long it takes to annotation one picture?

from vpgnet.

BAILOOL avatar BAILOOL commented on August 25, 2024

@daixiaogang, the detailed explanation about "Data Collection and Annotation" is provided in Section 3.1 in the paper. Simply speaking, you mark corner points of lane and road markings to form a closed polygon. This closed region (when it is filled) is essentially the same as the pixel-wise annotation (traditional annotation of segmentation datasets). Further, we wrote a script to transform this annotation to a grid-wise (each grid cell is 8x8 pixels) annotation. The reason we did the annotation in such way is that the lanes and road markings can be easily bounded given a polygon and it is much faster technique than performing pixel-wise annotation.

Given the that we have an internal tool ( unfortunately, not available for disclosure) polygon annotation takes about 1-5 minutes per image depending on the complexity. Subsequently, pixel-wise to grid-wise annotation is done automatically almost instantaneously.

from vpgnet.

daixiaogang avatar daixiaogang commented on August 25, 2024

@BAILOOL ,thanks for your explanation. In your project,different lane is regarded as a object? And the datasets will not be open?

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@daixiaogang Could you explain the meaning of 'different lane as a object'? Each lane and road marking has different shape, so they are hard to defined by a single box (rectangle). That's the main reason why we use the grid annotation. Specifically, we define each object by a list of grids, such as [(x1 y1 x1' y1' c1), (x2 y2 x2' y2' c2), ...] where {x,y,x',y'} localize the position of each grid and {c} is the corresponding class. There are discontinuous lanes such as dashed lines, in this case, only painted markings are regarded as one of lane classes. The dataset will be available, but currently is being reviewed by Samsung company. They are undergoing a merger and restructuring, which is delaying the work.

from vpgnet.

YaoLing13 avatar YaoLing13 commented on August 25, 2024

Is the output of grid box in VPGNet used for lane detection?

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@xumi13 We didn't use the grids for the lane detection. We elaborated the post-processing of lane detection in our paper :)

from vpgnet.

daixiaogang avatar daixiaogang commented on August 25, 2024

@SeokjuLee ,I am sorry,my English is very poor,thanks for your explanation.

from vpgnet.

qinghuizhao avatar qinghuizhao commented on August 25, 2024

Could you explain the meaning of “Vanishing Point”?is there no public datasets?

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@qinghuizhao Hello, the geometric meaning of a vanishing point is a point where parallel lines in a three-dimensional space converge to a two-dimensional plane by a graphical perspective. In our work, we define our "Vanishing Point (VP)" as the nearest point on the horizon where lanes converge and disappear predictively around the farthest point of the visible lane. You can see some examples of our "VP" annotation in this link (https://youtu.be/jnewRlt6UbI?t=16). There are several public vanishing point dataset such as The Toulouse Vanishing Points Dataset, but the meaning our "VP" is slightly different and ours assumes the case of "VP" in driving road scenario.

from vpgnet.

gexahedron avatar gexahedron commented on August 25, 2024

What are the news about releasing dataset?

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@gexahedron Sorry for the delay in publishing our dataset. We keep asking Samsung to publish dataset, and hope to be released soon.

from vpgnet.

YangShuoAI avatar YangShuoAI commented on August 25, 2024

@SeokjuLee can u please tell me the details of train_list.txt and test_list.txt in make_lmdb.sh?

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@KeplerCV The list files can be generated by 'caltech-lanes-dataset/vpg_annot_v1.m'. The format should be [Img path] [num. of grids] [x_1 y_1 x'_1 y'_1] ... [x_n y_n x'_n' y'_n] line by line. There is also a visualizing part (commented) for better understanding. Because at this moment, our dataset is being reviewed by Samsung, so we provide a baseline network that doesn't need VP annotations.

from vpgnet.

daixiaogang avatar daixiaogang commented on August 25, 2024

@SeokjuLee ,I have run caltech-lanes-dataset/vpg_annot_v1.m and get four *.txt file which is your label ,can I use this to train your network? Do you have the pretrained model?

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@daixiaogang Yes, converted list files are for the toy example that how we trained the VPGNet. You can train with them (after LMDB parsing), but they are small sets so the performance is not guaranteed. It is best to finetune, but for the pretrained model and dataset, we need to get a permission from Samsung. The reason I share this provisional code is that they are delaying the permissions..

from vpgnet.

qinghuizhao avatar qinghuizhao commented on August 25, 2024

can u please share some train pictures?

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@daixiaogang This issue is caused by the path of list file. please try "../../build/tools/convert_driving_data /home/swjtu/daixiaogang/VPGNet/caltech-lanes-dataset /home/swjtu/daixiaogang/VPGNet/caltech-lanes-dataset/cordova1.txt LMDB_train" after delete the previous LMDB files.

from vpgnet.

daixiaogang avatar daixiaogang commented on August 25, 2024

@SeokjuLee ,I have solve this problem and run the code with caltech-lanes datasets. If I used may dataset which only label the right and left lane boundary with (x1,y1,x1',y1'),(x2,y2,x2',y2'),what should I do to make label to run on your code?

from vpgnet.

daixiaogang avatar daixiaogang commented on August 25, 2024

@SeokjuLee ,your caffe is run on CPU mode? your makefile.config have comment USECUDNN=1.

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@daixiaogang In this case, you need to change your annotation format same as Caltech to VPGNet annotation method. Please refer our matlab code 'vpg_annot_v1.m' to make grid annotations. It would be easy to follow with our visualization lines. About the CUDNN option, we trained on GPU mode, but not used cudnn.

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

Please leave code/training questions in the CODE ISSUES pannel (the 3rd one).

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@qinghuizhao The training curve depends on how you've arranged the dataset. I uploaded one example code, 'outputs/sj_show_plot.py'. Please refer it.

from vpgnet.

llmettyll avatar llmettyll commented on August 25, 2024

Hi, I tested VPGNet toy example with Caltech lane DB using your caffe project.
And I checked that validation accuracy is around 98.9 % after default solver setting when I divided the dataset into train_set(8) and val_set(2) with shuffled list.
However, the Caltech DB doesn't have 'bb-label', 'pixel-label', 'size-label', and 'norm-label' information, which you used in 'train_val.prototxt'. So, I am very curious about how to annotate these labels.

According to 'drive_data_layer.cpp' and 'caffe.proto' in your project,
'norm-label' and 'size-label' are assigned to '1', '1/w', or '1/h' at first.
And they are partially changed to '-1' or '0' in the later code.
On the other hand, I couldn't find how the 'bb-label' is annotated in 'output.txt' and is deliverd to 'drive_data_layer.cpp'.

Can you explain on these~???

Thanks :)

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

@llmettyll Hi, the drive-data-layer handles every label. If you see the network diagram (train_val.png), you could follow the splitted outputs (type, data, and label). Therefore to generate those information, only the list file is needed.

from vpgnet.

gexahedron avatar gexahedron commented on August 25, 2024

Could you send me your dataset, please, if it's possible? My email - [email protected]
Thanks!

from vpgnet.

aisichonzi007 avatar aisichonzi007 commented on August 25, 2024

@SeokjuLee hello, how to open the .ccvl file

from vpgnet.

xinping12345678 avatar xinping12345678 commented on August 25, 2024

Could you send me your dataset, please, if it's possible? My email - [email protected]
Thanks!

hello, Did you get the data set? If you get the dataset, I would be very grateful if you could send it to me.

from vpgnet.

chenchaohui avatar chenchaohui commented on August 25, 2024

Hi, SeokjuLee, from your paper,we can know the lane is first annotated with spline ,then convert it into grid cell annotation use a matlab script you provided. But how should we annotate the zebra line,stop line and arrows on the surface of the road. Waiting for your help.

from vpgnet.

yurenwei avatar yurenwei commented on August 25, 2024

Excuse me, I want to ask about the labeling format of your dataset. This means that I want to use your network, how do I label the data set.

from vpgnet.

leicao-me avatar leicao-me commented on August 25, 2024

Dear Authors,

Is there any update on releasing your dataset, please? Thank you

from vpgnet.

CountofMont avatar CountofMont commented on August 25, 2024

Could you send me your dataset, please, if it's possible? My email ----- [email protected]
Thank you very much!

from vpgnet.

hakillha avatar hakillha commented on August 25, 2024

Hi could anyone send me the access to the dataset? My email is [email protected]. Rly appreciate it if someone could share the data!

from vpgnet.

aisichonzi007 avatar aisichonzi007 commented on August 25, 2024

from vpgnet.

SeokjuLee avatar SeokjuLee commented on August 25, 2024

Hi, the dataset is now available.
Please fill out a form for the download link.

from vpgnet.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.