Powered by OpenAIRE graph
Found an issue? Give us feedback
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/ ZENODOarrow_drop_down
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Dataset . 2022
License: CC BY
Data sources: Datacite
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Dataset . 2022
License: CC BY
Data sources: ZENODO
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Dataset . 2022
License: CC BY
Data sources: Datacite
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
https://doi.org/10.5281/zenodo...
Dataset . 2022
License: CC BY
Data sources: Sygma
versions View all 3 versions
addClaim

sustAGE User Postures & Actions Monitoring dataset

Authors: Papoutsakis, Konstantinos; Papadopoulos, Georgios; Maniadakis, Michail; Papadopoulos, Thodoris; Lourakis, Manolis; Pateraki, Maria; Varlamis, Iraklis;

sustAGE User Postures & Actions Monitoring dataset

Abstract

The sustAGE User Postures & Actions Monitoring dataset contains videos, data related to occurrences of ergonomic (body straining) postures as well as heart rate measurements of line workers during work activities in a realistic manufacturing environment. The time-synchronized data streams were captured during assembly tasks performed by real line workers in a realistic car door assembly production line in the premises of Stellantis --- Centro Ricerche FIAT (CRF)/SPW Research \& Innovation department in Melfi, Italy in the context of the sustAGE project. The dataset can be used by methods that target the following research tasks: a) vision-based detection of worker’s ergonomically unhealthy (body straining) postures during assembly tasks according to the MURI risk analysis tool, b) vision-based recognition of human assembly actions, c) multi-modal analysis and forecasting of worker heart rate and physical fatigue based on heart rate measurements and the detection (annotated occurrences) of body straining postures during work activities. Each video shows a single line worker that performs a series of car door assembly activities, noted as a task cycle execution, for a specific workstation of the production line. A set of twelve task cycle executions performed by three different workers were captured using static StereoLabs ZED or ZED2 sensors. Each task cycle execution has a mean duration of 4 minutes. Moreover, time-synchronised data for the occurrences of body straining postures and worker heart rate are available for 8 work sessions. Each work session was recorded during morning or afternoon time of the work shift and regards 3 to 5 consecutive task cycle executions performed by the same line worker. Heart rate data was captured using a Garmin Vivoactive 3 smartwatch. Time annotations are available for the samples collected during the work sessions based on the Unix time (Epoch time) format. Annotation data are available for twelve task cycle executions and contain the semantic labels and temporal boundaries of each assembly action and for the ergonomic risk scores of body straining postures performed by the line worker during each task cycle execution. Data was provided by experts in manufacturing and ergonomics based on the MURI risk analysis tool. Documentation of this dataset can be found in [1]. If you use the dataset in your research work, you are kindly asked to cite [1] in your publications. [1] Papoutsakis K, Papadopoulos G, Maniadakis M, Papadopoulos T, Lourakis M, Pateraki M, Varlamis I. Detection of Physical Strain and Fatigue in Industrial Environments Using Visual and Non-Visual Low-Cost Sensors. MDPI Technologies. 2022; 10(2):42. https://doi.org/10.3390/technologies10020042

{"references": ["https://doi.org/10.3390/technologies10020042"]}

Keywords

physical fatigue, ergonomic body postures, visual recognition, heart rate analysis, dataset, manufacturing activities, computer vision

  • BIP!
    Impact byBIP!
    selected citations
    These citations are derived from selected sources.
    This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    0
    popularity
    This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
    Average
    influence
    This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    Average
    impulse
    This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
    Average
    OpenAIRE UsageCounts
    Usage byUsageCounts
    visibility views 27
    download downloads 17
  • 27
    views
    17
    downloads
    Powered byOpenAIRE UsageCounts
Powered by OpenAIRE graph
Found an issue? Give us feedback
visibility
download
selected citations
These citations are derived from selected sources.
This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Citations provided by BIP!
popularity
This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
BIP!Popularity provided by BIP!
influence
This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Influence provided by BIP!
impulse
This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
BIP!Impulse provided by BIP!
views
OpenAIRE UsageCountsViews provided by UsageCounts
downloads
OpenAIRE UsageCountsDownloads provided by UsageCounts
0
Average
Average
Average
27
17