Image Annotation for Large Dataset
up vote
0
down vote
favorite
I have this huge data set for which I have only taken a sample from that data to show you,now as you can see it has two class cat and dog where in the training data i have to label it manually since the cat and dog images are mixed, so is there any alternative way to do it.I have to annotate this then only i can train as to if whether cat or dog.
annotations computer-vision transfer-learning
add a comment |
up vote
0
down vote
favorite
I have this huge data set for which I have only taken a sample from that data to show you,now as you can see it has two class cat and dog where in the training data i have to label it manually since the cat and dog images are mixed, so is there any alternative way to do it.I have to annotate this then only i can train as to if whether cat or dog.
annotations computer-vision transfer-learning
How accurate must the labelling be? If you can have some error rate, then you could always label automatically with an available deep network; many deep networks are quite accurate for dog and cat images, since these are often well represented in common training datasets.
– Mozglubov
Nov 15 at 21:35
Hi Mozglubov, in order to model as from the training set i have to split this as cat and dog first and for larger images in a directory i cannot manually annotate each one as cat.1,cat.2 or dog.1,dog.2,so is this there any other alternative for this or i have to manually do this?
– S L SREEJITH
2 days ago
Well, what I was trying to get at is whether or not you need that exact annotation. Depending on what you plan to do with this dataset, it is sometimes sufficient to have labels which are mostly correct. This is sometimes referred to as "soft" or "noisy" labelling. This paper by Reed et al. (arxiv.org/abs/1412.6596) provides a good example of achieving strong results despite not having fully accurate training labels. If you absolutely do need a human-accurate set of annotations but cannot do it yourself, then you may want to use services like Amazon's Mechanical Turk.
– Mozglubov
2 days ago
add a comment |
up vote
0
down vote
favorite
up vote
0
down vote
favorite
I have this huge data set for which I have only taken a sample from that data to show you,now as you can see it has two class cat and dog where in the training data i have to label it manually since the cat and dog images are mixed, so is there any alternative way to do it.I have to annotate this then only i can train as to if whether cat or dog.
annotations computer-vision transfer-learning
I have this huge data set for which I have only taken a sample from that data to show you,now as you can see it has two class cat and dog where in the training data i have to label it manually since the cat and dog images are mixed, so is there any alternative way to do it.I have to annotate this then only i can train as to if whether cat or dog.
annotations computer-vision transfer-learning
annotations computer-vision transfer-learning
edited Nov 8 at 13:20
Dave
1,91051322
1,91051322
asked Nov 8 at 11:46
S L SREEJITH
12
12
How accurate must the labelling be? If you can have some error rate, then you could always label automatically with an available deep network; many deep networks are quite accurate for dog and cat images, since these are often well represented in common training datasets.
– Mozglubov
Nov 15 at 21:35
Hi Mozglubov, in order to model as from the training set i have to split this as cat and dog first and for larger images in a directory i cannot manually annotate each one as cat.1,cat.2 or dog.1,dog.2,so is this there any other alternative for this or i have to manually do this?
– S L SREEJITH
2 days ago
Well, what I was trying to get at is whether or not you need that exact annotation. Depending on what you plan to do with this dataset, it is sometimes sufficient to have labels which are mostly correct. This is sometimes referred to as "soft" or "noisy" labelling. This paper by Reed et al. (arxiv.org/abs/1412.6596) provides a good example of achieving strong results despite not having fully accurate training labels. If you absolutely do need a human-accurate set of annotations but cannot do it yourself, then you may want to use services like Amazon's Mechanical Turk.
– Mozglubov
2 days ago
add a comment |
How accurate must the labelling be? If you can have some error rate, then you could always label automatically with an available deep network; many deep networks are quite accurate for dog and cat images, since these are often well represented in common training datasets.
– Mozglubov
Nov 15 at 21:35
Hi Mozglubov, in order to model as from the training set i have to split this as cat and dog first and for larger images in a directory i cannot manually annotate each one as cat.1,cat.2 or dog.1,dog.2,so is this there any other alternative for this or i have to manually do this?
– S L SREEJITH
2 days ago
Well, what I was trying to get at is whether or not you need that exact annotation. Depending on what you plan to do with this dataset, it is sometimes sufficient to have labels which are mostly correct. This is sometimes referred to as "soft" or "noisy" labelling. This paper by Reed et al. (arxiv.org/abs/1412.6596) provides a good example of achieving strong results despite not having fully accurate training labels. If you absolutely do need a human-accurate set of annotations but cannot do it yourself, then you may want to use services like Amazon's Mechanical Turk.
– Mozglubov
2 days ago
How accurate must the labelling be? If you can have some error rate, then you could always label automatically with an available deep network; many deep networks are quite accurate for dog and cat images, since these are often well represented in common training datasets.
– Mozglubov
Nov 15 at 21:35
How accurate must the labelling be? If you can have some error rate, then you could always label automatically with an available deep network; many deep networks are quite accurate for dog and cat images, since these are often well represented in common training datasets.
– Mozglubov
Nov 15 at 21:35
Hi Mozglubov, in order to model as from the training set i have to split this as cat and dog first and for larger images in a directory i cannot manually annotate each one as cat.1,cat.2 or dog.1,dog.2,so is this there any other alternative for this or i have to manually do this?
– S L SREEJITH
2 days ago
Hi Mozglubov, in order to model as from the training set i have to split this as cat and dog first and for larger images in a directory i cannot manually annotate each one as cat.1,cat.2 or dog.1,dog.2,so is this there any other alternative for this or i have to manually do this?
– S L SREEJITH
2 days ago
Well, what I was trying to get at is whether or not you need that exact annotation. Depending on what you plan to do with this dataset, it is sometimes sufficient to have labels which are mostly correct. This is sometimes referred to as "soft" or "noisy" labelling. This paper by Reed et al. (arxiv.org/abs/1412.6596) provides a good example of achieving strong results despite not having fully accurate training labels. If you absolutely do need a human-accurate set of annotations but cannot do it yourself, then you may want to use services like Amazon's Mechanical Turk.
– Mozglubov
2 days ago
Well, what I was trying to get at is whether or not you need that exact annotation. Depending on what you plan to do with this dataset, it is sometimes sufficient to have labels which are mostly correct. This is sometimes referred to as "soft" or "noisy" labelling. This paper by Reed et al. (arxiv.org/abs/1412.6596) provides a good example of achieving strong results despite not having fully accurate training labels. If you absolutely do need a human-accurate set of annotations but cannot do it yourself, then you may want to use services like Amazon's Mechanical Turk.
– Mozglubov
2 days ago
add a comment |
1 Answer
1
active
oldest
votes
up vote
0
down vote
One possible solution is to upload your dataset to labelbox (link: https://www.labelbox.com/) there you are able to annotate your dataset and then download the results for instance as a JSON file. The web page correlates your images with the labels and then you can use those informations for your work.
add a comment |
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
0
down vote
One possible solution is to upload your dataset to labelbox (link: https://www.labelbox.com/) there you are able to annotate your dataset and then download the results for instance as a JSON file. The web page correlates your images with the labels and then you can use those informations for your work.
add a comment |
up vote
0
down vote
One possible solution is to upload your dataset to labelbox (link: https://www.labelbox.com/) there you are able to annotate your dataset and then download the results for instance as a JSON file. The web page correlates your images with the labels and then you can use those informations for your work.
add a comment |
up vote
0
down vote
up vote
0
down vote
One possible solution is to upload your dataset to labelbox (link: https://www.labelbox.com/) there you are able to annotate your dataset and then download the results for instance as a JSON file. The web page correlates your images with the labels and then you can use those informations for your work.
One possible solution is to upload your dataset to labelbox (link: https://www.labelbox.com/) there you are able to annotate your dataset and then download the results for instance as a JSON file. The web page correlates your images with the labels and then you can use those informations for your work.
answered Nov 8 at 16:58
Niccolò Cacciotti
1124
1124
add a comment |
add a comment |
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53207102%2fimage-annotation-for-large-dataset%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
How accurate must the labelling be? If you can have some error rate, then you could always label automatically with an available deep network; many deep networks are quite accurate for dog and cat images, since these are often well represented in common training datasets.
– Mozglubov
Nov 15 at 21:35
Hi Mozglubov, in order to model as from the training set i have to split this as cat and dog first and for larger images in a directory i cannot manually annotate each one as cat.1,cat.2 or dog.1,dog.2,so is this there any other alternative for this or i have to manually do this?
– S L SREEJITH
2 days ago
Well, what I was trying to get at is whether or not you need that exact annotation. Depending on what you plan to do with this dataset, it is sometimes sufficient to have labels which are mostly correct. This is sometimes referred to as "soft" or "noisy" labelling. This paper by Reed et al. (arxiv.org/abs/1412.6596) provides a good example of achieving strong results despite not having fully accurate training labels. If you absolutely do need a human-accurate set of annotations but cannot do it yourself, then you may want to use services like Amazon's Mechanical Turk.
– Mozglubov
2 days ago