Items related to Spatio-Temporal Networks for Human Activity Recognition...

Spatio-Temporal Networks for Human Activity Recognition based on Optical Flow in Omnidirectional Image Scenes - Softcover

 
9783961002054: Spatio-Temporal Networks for Human Activity Recognition based on Optical Flow in Omnidirectional Image Scenes
  • PublisherTU Chemnitz
  • Publication date2024
  • ISBN 10 3961002053
  • ISBN 13 9783961002054
  • BindingPaperback
  • LanguageEnglish
  • Number of pages212

Search results for Spatio-Temporal Networks for Human Activity Recognition...

Seller Image

Roman Seidel
ISBN 10: 3961002053 ISBN 13: 9783961002054
New Taschenbuch
Print on Demand

Seller: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Taschenbuch. Condition: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -The property of human motion perception is used in this dissertation to infer human activity from data using artificial neural networks. One of the main aims of this thesis is to discover which modalities, namely RGB images, optical flow and human keypoints, are best suited for HAR in omnidirectional data. Since these modalities are not yet available for omnidirectional cameras, they are synthetically generated with a 3D indoor simulation with the result of a large-scale dataset, called OmniFlow. Due to the lack of omnidirectional optical flow data, the OmniFlow dataset is validated using Test-Time Augmentation. Compared to the baseline, which contains Recurrent All-Pairs Field Transforms trained on the FlyingChairs and FlyingThings3D datasets, it was found that only about 1000 images need to be used for fine-tuning to obtain a very low End-point Error. For an evaluation on activity-level, two state-of-the-art convolutional neural networks (CNNs), namely the Temporal Segment Network (TSN) for the modalities RGB images and optical flow and the PoseC3D for the modality human keypoints, were used. Both CNNs were trained and validated on OmniFlow and on the real-world dataset OmniLab. For both networks, TSN and PoseC3D, three hyperparameters were varied and the top-1, top-5 and mean accuracies were reported. In addition, confusion matrices indicating the class-wise accuracy of the 15 activity classes have been given for the modalities RGB images, optical flow and human keypoints. 180 pp. Englisch. Seller Inventory # 9783961002054

Contact seller

Buy New

US$ 24.72
Convert currency
Shipping: US$ 25.20
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 2 available

Add to basket

Seller Image

Roman Seidel
ISBN 10: 3961002053 ISBN 13: 9783961002054
New Taschenbuch
Print on Demand

Seller: AHA-BUCH GmbH, Einbeck, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Taschenbuch. Condition: Neu. nach der Bestellung gedruckt Neuware - Printed after ordering - The property of human motion perception is used in this dissertation to infer human activity from data using artificial neural networks. One of the main aims of this thesis is to discover which modalities, namely RGB images, optical flow and human keypoints, are best suited for HAR in omnidirectional data. Since these modalities are not yet available for omnidirectional cameras, they are synthetically generated with a 3D indoor simulation with the result of a large-scale dataset, called OmniFlow. Due to the lack of omnidirectional optical flow data, the OmniFlow dataset is validated using Test-Time Augmentation. Compared to the baseline, which contains Recurrent All-Pairs Field Transforms trained on the FlyingChairs and FlyingThings3D datasets, it was found that only about 1000 images need to be used for fine-tuning to obtain a very low End-point Error. For an evaluation on activity-level, two state-of-the-art convolutional neural networks (CNNs), namely the Temporal Segment Network (TSN) for the modalities RGB images and optical flow and the PoseC3D for the modality human keypoints, were used. Both CNNs were trained and validated on OmniFlow and on the real-world dataset OmniLab. For both networks, TSN and PoseC3D, three hyperparameters were varied and the top-1, top-5 and mean accuracies were reported. In addition, confusion matrices indicating the class-wise accuracy of the 15 activity classes have been given for the modalities RGB images, optical flow and human keypoints. Seller Inventory # 9783961002054

Contact seller

Buy New

US$ 24.72
Convert currency
Shipping: US$ 32.39
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 2 available

Add to basket

Seller Image

Roman Seidel
Published by Universitätsverlag Chemnitz, 2024
ISBN 10: 3961002053 ISBN 13: 9783961002054
New Taschenbuch

Seller: preigu, Osnabrück, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Taschenbuch. Condition: Neu. Spatio-Temporal Networks for Human Activity Recognition based on Optical Flow in Omnidirectional Image Scenes | Roman Seidel | Taschenbuch | Englisch | Universitätsverlag Chemnitz | EAN 9783961002054 | Verantwortliche Person für die EU: preigu, Ansas Meyer, Lengericher Landstr. 19, 49078 Osnabrück, mail[at]preigu[dot]de | Anbieter: preigu. Seller Inventory # 128685799

Contact seller

Buy New

US$ 24.72
Convert currency
Shipping: US$ 76.71
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 5 available

Add to basket