Powered by OpenAIRE graph
Found an issue? Give us feedback
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/ ZENODOarrow_drop_down
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Software . 2022
Data sources: Datacite
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Software . 2022
Data sources: Datacite
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Software . 2022
Data sources: ZENODO
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Software . 2022
Data sources: Datacite
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Software . 2022
Data sources: ZENODO
versions View all 3 versions
addClaim

Common guillemots in the Baltic Sea studied with video surveillance and object detection: raw data, annotations, model, and model outputs

Authors: Hentati-Sundberg, Jonas; Olin, Agnes;

Common guillemots in the Baltic Sea studied with video surveillance and object detection: raw data, annotations, model, and model outputs

Abstract

The dataset contains: Annotations (images and bounding box coordinates). The format is "YOLO Darknet TXT", described in detail here: https://roboflow.com/formats/yolo-darknet-txt The YOLOV5 model, 2 files: .cfg and .weights Inference from the YOLO model (bounding boxes) in SQlite format (.db) The videos themselves are several Tb (approx 1 Gb per H of film) and therefore not possible to share at this moment at a reasonable cost. Sample videos (downsampled to 1 FPS) are provided as .avi files. Python code for training and running Yolov5 is available at: https://github.com/BalticSeabird/SeabirdDetections R code for post-processing the model outputs including the generation of figures is available at https://github.com/BalticSeabird/ObjectDetectionInferencesFunding provided by: Svenska Forskningsrådet FormasCrossref Funder Registry ID: http://dx.doi.org/10.13039/501100001862Award Number: 2021-02639Funding provided by: VetenskapsrådetCrossref Funder Registry ID: http://dx.doi.org/10.13039/501100004359Award Number: 2021-03892Funding provided by: Marcus och Amalia Wallenbergs minnesfondCrossref Funder Registry ID: http://dx.doi.org/10.13039/501100011898Award Number: 2018-0093

The raw data is videos in .avi format, recorded with a Avtech AVH8516 Network Video Recorder (NVR) connected to a 2 megapixel IP camera (Avtech AVM543P) operating at 25 frames per second (FPS). From the videos, single frames have been picked out and annotated with bounding boxes with three classes: adult bird (0), chick (1) and egg (2). Using these annotations, a YOLOv5 model has been trained. The videos have then been downsampled to 1 FPS and Object detection has been run on the whole material, and the output has been stored in a SQLite database coordinates for each detected object.

The data comes from common guillemots studied at Stora Karlsö, Sweden between 2019 and 2021. The common guillemots breed at an artificial cliff, and has been filmed continusly from above over three breeding seasons. Using the video material, a YOLOv5 model has been trained to detect adult birds, chicks and eggs. The dataset contains annotations (bounding boxes) used for training the model, the model itself, and outputs from the model (object detections). The data can be used and shared freely.

Keywords

Big Data, Ecological velocity, machine learning, Monitoring, deep learning, object detection, artificial intelligence, seabirds

  • BIP!
    Impact byBIP!
    selected citations
    These citations are derived from selected sources.
    This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    0
    popularity
    This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
    Average
    influence
    This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    Average
    impulse
    This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
    Average
    OpenAIRE UsageCounts
    Usage byUsageCounts
    visibility views 3
  • 3
    views
    Powered byOpenAIRE UsageCounts
Powered by OpenAIRE graph
Found an issue? Give us feedback
visibility
selected citations
These citations are derived from selected sources.
This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Citations provided by BIP!
popularity
This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
BIP!Popularity provided by BIP!
influence
This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Influence provided by BIP!
impulse
This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
BIP!Impulse provided by BIP!
views
OpenAIRE UsageCountsViews provided by UsageCounts
0
Average
Average
Average
3
Related to Research communities