Biodiversity Data Journal 9: e60548 OO) doi: 10.3897/BDJ.9.e60548 open access Software Description An open-source, citizen science and machine learning approach to analyse subsea movies Victor Anton*, Jannes Germishuys§, Per Bergstrém!, Mats Lindegarth!, Matthias Obst! $ Wildlife.ai, New Plymouth, New Zealand § Combine AB, Gothenburg, Sweden | Department of Marine Sciences, Géteborg University, Gothenburg, Sweden | SeAnalytics AB, Gothenburg, Sweden Corresponding author: Victor Anton (victor@wildlife.ai), Matthias Obst (matthias.obst@marine.gu.se) Academic editor: Danwei Huang Received: 09 Nov 2020 | Accepted: 11 Feb 2021 | Published: 24 Feb 2021 Citation: Anton V, Germishuys J, Bergstr6ém P, Lindegarth M, Obst M (2021) An open-source, citizen science and machine learning approach to analyse subsea movies. Biodiversity Data Journal 9: e60548. https://doi.org/10.3897/BDJ.9.e60548 Abstract Background The increasing access to autonomously-operated technologies offer vast opportunities to sample large volumes of biological data. However, these technologies also impose novel demands on ecologists who need to apply tools for data management and processing that are efficient, publicly available and easy to use. Such tools are starting to be developed for a wider community and here we present an approach to combine essential analytical functions for analysing large volumes of image data in marine ecological research. New information This paper describes the Koster Seafloor Observatory, an open-source approach to analysing large amounts of subsea movie data for marine ecological research. The approach incorporates three distinct modules to: manage and archive the subsea movies, involve citizen scientists to accurately classify the footage and, finally, train and test machine learning algorithms for detection of biological objects. This modular approach is © Anton V et al. This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. 2 Anton V etal based on open-source code and allows researchers to customise and further develop the presented functionalities to various types of data and questions related to analysis of marine imagery. We tested our approach for monitoring cold water corals in a Marine Protected Area in Sweden using videos from remotely-operated vehicles (ROVs). Our study resulted in a machine learning model with an adequate performance, which was entirely trained with classifications provided by citizen scientists. We illustrate the application of machine learning models for automated inventories and monitoring of cold water corals. Our approach shows how citizen science can be used to effectively extract occurrence and abundance data for key ecological species and habitats from underwater footage. We conclude that the combination of open-source tools, citizen science systems, machine learning and high performance computational resources are key to successfully analyse large amounts of underwater imagery in the future. Keywords marine biodiversity, autonomous underwater vehicles, remotely-operated vehicles, artificial intelligence, big data, image analysis, participatory science, Essential Biodiversity Variables, research infrastructure, biodiversity monitoring Introduction Biological observation techniques in the marine environment need to improve radically to serve our understanding of marine ecosystems under the influence of multiple stressors including long-term global change (Benedetti-Cecchi et al. 2018). Over the last decade, biologists have gained an increased access to autonomously operated technologies for data collection, offering the opportunity to generate enormous volumes of data. This is especially the case for high-definition optical imagery recorded by ROV’s (remotely- operated vehicles), AUVs (autonomous underwater vehicles), drop-cameras, video plankton recorders and drones (Bean et al. 2017, Danovaro et al. 2016). Although such image-based observations may revolutionise the fields of marine biology and biodiversity monitoring, these methods also impose completely new demands for data management and processing on researchers. In-situ monitoring systems need to be coupled to data services that allow for swift exploration, processing and long-term storage (Guidi et al. 2020). Some of these services already exist, for example, the Global Reef Record and CoralNet, which allow researchers to host and analyse images of coral reefs (Beijoom et al. 2015), Ecolaxa that offers analysis of large amounts of plankton imagery (Picheral et al. 2017) and FathomNet, which offers machine learning algorithms and training data to analyse deep-sea footage. Although these platforms have pioneered the daily use of image analysis tools in marine science, they may not be able to provide all the functionalities needed by the fast-growing community of users. Some of these sought-after functions include seamless connectivity with project-specific data archives, the involvement of non-scientific audiences in environmental research, modules that can be easily updated to include state-of-the-art An open-source, citizen science and machine learning approach to analyse ... 3 analytical tools and versatile systems that researchers can easily adapt to fit to different types of data and purposes. Here, we present the Koster Seafloor Observatory, an open-source modular approach for managing, processing, and analysing large amounts of subsea movie data for marine ecological research. The Koster Seafloor Observatory allows scientists to upload underwater footage to a customised citizen science website and then train machine learning algorithms with those classifications provided by citizen scientists. These algorithms can be accessed through an Application Programming Interface (API) allowing researchers to test the performance of the algorithms under different confidence and overlapping thresholds, share their models with a wider audience and extract species observations from new footage. Project description Title: Mapping cold water corals in Sweden's first marine national park Study area description: We piloted the Koster Seafloor Observatory to extract data on spatiotemporal distribution and relative abundance of habitat-building species from deep- water recordings in a Marine Protected Area, the Kosterhavets National Park in Sweden. The Park, established in 2009, contains a highly diverse and unique marine ecosystem. The seafloor in the deeper waters of the Park has oceanic connections and hence contains much of the bottom-dwelling fauna, which is otherwise only found in deep oceanic waters (Lavaleye et al. 2009). This fauna includes large habitat-building species (Costello et al. 2005), such as sponges (e.g. Geodia baretti, Phakellia ventilabrum) and cold water corals (e.g. Desmophyllum pertusum), as well as other large species which can be easily identified from camera footage (e.g. starfish Porania pulvillus, Crossaster papposus, Echinus esculentus). Design description: The Koster Seafloor Observatory is divided into three main modules: data management, citizen science and machine learning with high performance computing (Fig. 1). Module 1: Data management (Anton et al. 2019) In the data management module, researchers store and process the data in a way that maximises efficiency, convenience and opportunities for sharing and collaboration. To store and access the raw data, we use long-term and short-term storage servers. The long-term storage server, or cold storage, archives large amounts of files that need not be accessed frequently. In our case, these include recordings from Remotely-Operated Vehicles (ROVs) managed by the University of Gothenburg, Sweden. The movies (mp4 and mov formats) are on average 1-2 hours long and have been systematically collected from all expeditions since the late 1990s (Fig. 1). The metadata associated with these movies is regularly published in the Swedish National Data Archive. 4 Anton V et al The short-term storage server, or hot storage, stores a small proportion of files that are frequently used for analysis. Here, we transferred 60 movies from the cold storage to a project-specific short-term storage server (Suppl. material 2). The number of movies we selected was a compromise between selecting a representative sample and efficiently using the limited storage of our server. This "hot server" was Linux-based and hosted by C halmers University of Technology, Gothenburg. The specifications of this High Performance Computing server consisted of a GTX2080Ti GPU with 2 x 8 core Intel(R) Core(TM) i9-9900 CPU @ 3.10GHz (total 16 cores) and 2GB DDR4 RAM. Data Management Citizen Science a= Pe rant nie ry | Vv | le >| | Workflow 2 a v | Aggregat wnnota Machine Learning and High-performance Computing Results Figure 1. EESl High-level overview of the three main modules and the components of the Koster Seafloor Observatory. We created a SQLite database to link all information related to the movies and the classifications provided by both citizen scientists and machine learning algorithms (Fig. 1). The database has seven interconnected tables (Fig. 2). The “movies”, “sites” and “species” tables have project-specific information from the underwater movie metadata, as well as the species choices available for citizen scientists to annotate the clips, retrieved from Zooniverse. The “agg_annotations frame” and “agg _annotations clip” tables contain information related to the annotations provided by citizen scientists. The “subjects” table has information related to the clips and frames uploaded to the Koster Seafloor Observatory. The "model_annotations" table holds information related to the annotations provided by the machine learning algorithms. The database followed the Darwin Core (DwC) standards to maximise the sharing, use and reuse of open-access biodiversity data. An open-source, citizen science and machine learning approach to analyse ... 5 movies model_annotations species varchar text id integer id integer filename text frame_number integer created_on datetime version_ml integer integer species_id integer datetime movie_id integer text created_at datetime integer confidence integer text fi agg_annotations_frame subjects id integer id integer species_id integer subject_type varchar x_position integer filename text y_position integer clip_start_time datetime width integer clip_end_time datetime height integer J frame_exp_sp_id integer subject_id integer ‘ frame_number integer workflow_id varchar subject_set_id varchar classifications_count integer retired_at datetime retirement_reason text id integer created_at datetime d species_id integer movie_id integer how_many integer first_seen integer id integer name text coord_lat varchar coord_lon varchar protected varchar agg_annotations_clip subject_id integer Figure 2. EES Entity relationship diagram of the SQLite database used by the Koster Seafloor Observatory. Module 2: Citizen science (Anton et al. 2019) In the citizen science module, researchers and citizen scientists work together to efficiently and accurately annotate raw data. To identify the species recorded in our footage, we created a citizen science website. The site is hosted in Zooniverse, the largest citizen science platform in the world. The website contains rich supporting material (e.g. background, tutorials, field guides) and features two workflows that help citizen scientists to Classify biological objects in video (workflow 1) and locate these objects in still images (workflow 2). Workflow 1 (species identification): Citizen scientists are presented with 10-second clips of underwater footage and need to select at least one of the 27 available choices (Fig. 3). The choices include species of scientific importance, animals grouped at different taxonomic levels (e.g. “gastropods” or “fish”), as well as a few miscellaneous options (“Nothing here’, “Human objects’). If citizen scientists select a species or animal, they also need to specify the number of individuals of the taxon selected and the time (in seconds) when any of the individuals fully appears on the screen. We compared the classifications provided by an expert to those provided by citizen scientists to estimate the accuracy of citizen scientists to identify cold water corals (Table 1). A total of 2,594 clips were classified both by an expert and by eight different citizen 6 Anton V etal scientists. We aggregated the classifications provided by citizen scientists on a per-clip basis and retained the classifications of cold water corals and grouped the rest of classifications into "Other". For this case study, we chose cold water corals (Desmophyllum pertusum) because this species has a crucial ecological role in the study site (Costello et al. 2005). We used a confusion matrix to understand how agreement amongst citizen scientists correlates to the accuracy of their aggregated classifications (e.g. an agreement threshold of 80% corresponds to an agreement on the classifications of at least seven of the eight citizen scientists who annotated the clip). "Adequate" accuracy of citizen scientists with respect to experts depends on multiple parameters, including the type of data classified, the classification tool and the research questions (Aceves-Bueno et al. 2017). In our study, we decided that at least 80% of agreement amongst citizen scientists was an appropriate accuracy threshold as it minimised the number of false positives citizen scientists provide. TASK TUTORIAL Uke z Fan-shaped g Northern | Sugar sponge shrimp starfish yy Football gE Great spider la Rosy feather crab star sponge a Sponge (any Ez Deep sea E Echinoderm species) king crab (any species) Dead man's re Norway 2 File clam lobster «sl fingers FIELD GUIDE mr Dooplet sea | Crustacean | Bivalves (any anemone (any species) species) Common sea Black brittle o Gastropods is) a 5 pen star (any species) Deep water mo BM cushin star coral | be | Coral (any ol Common sea a | Nothing species) urchin here Showing 27 of 27 , {* oU8 Figure 3. EES] Screenshot of the Zooniverse annotation interface. On the left, display of the clips. On the right, species choices available. 0008 : : am 808 Figure 4. EES] Example of a frame containing cold water coral displayed to the citizen scientists (left) and the same frame with annotated rectangles provided by a citizen scientist (right). An open-source, citizen science and machine learning approach to analyse ... T Table 1. Confusion matrices derived from applying different citizen scientists agreement thresholds (Cit.Sci. Agr.) when comparing expert classifications to citizen scientist classifications of 2,594 underwater videos. Each video was classified by an expert and eight different citizen scientists. Classifications of cold water coral were retained and all other classifications were grouped as "Other". Expert classifications were compared to citizen scientist classifications with at least 80%, 60% and 40% of agreement amongst their responses (i.e. an agreement threshold of 80% corresponds to an agreement on the classifications of at least seven of the eight citizen scientists who annotated the clip). Cit. Sci. Agr. 2 80% Cit. Sci. Agr. 2 60% Cit. Sci. Agr. 2 40% Coral Other Coral Other Coral Other Expert Coral 111 467 315 263 475 103 Other 2 2014 22 1994 84 1932 Workflow 2 (object location): Citizen scientists are presented with a still image of the species of interest. To annotate the image, citizen scientists need to draw rectangles around the individuals of the species (Fig. 4). If citizen scientists are not able to identify any individual of the species of interest in the frame, they will not draw any rectangle. Each still image is annotated by at least five different citizen scientists before it is “retired” from the website. We used a four-stage video processing framework to upload clips and still images to the Koster Seafloor Observatory and download the annotations provided by citizen scientists (Fig. 5). Stage 1: Generate and upload clips (Fig. 5, circle a). In this stage, we split the +1 hour long movies into 10-second clips. After the clips were created, we randomly selected 5,702 clips from the original 60 movies and uploaded them to workflow 1 of the Koster Seafloor Observatory. Stage 2: Process clip annotations (Fig. 5, circle b). We retrieved the annotations provided by citizen scientists in workflow 1 and aggregated them on a per-clip basis. To aggregate workflow 1 annotations, we grouped the annotations each clip received and retained only those choices that were selected by at least 80% of the citizen scientists who annotated the clip. In our study, there were 194 clips for which cold-water coral was identified at least by 80% of the citizen scientists. We also averaged the answers from citizen scientists to the question "When is the first time the species appears fully in the video?". Stage 3: Generate and upload frames (Fig. 5, circle c). We extracted up to three frames per clip from the 194 clips containing cold water corals and extracted one frame per second after the first time the species fully appeared in the clip. After extracting 533 frames, we then uploaded them to workflow 2 of the Koster Seafloor Observatory. Five different citizen scientists per frame annotated the location of cold water corals in the still images. 8 Anton V etal ° Stage 4: Process frame annotations (Fig. 5, circle d). We retrieved workflow 2 annotations provided by citizen scientists and aggregated them on a per-frame basis. To aggregate workflow 2 annotations, we retained the area of overlapping between those rectangles drawn by 80% of the citizen scientists who annotated the frame. A total of 409 of the 533 frames had matching rectangles drawn by 80% of the citizen scientists. We formatted the aggregated annotations appropriately to train YOLOv3 algorithms (Redmon and Farhadi 2018) Figure 5. EES Four-stage video processing framework used to identify species of interest. Module 3: Machine learning and High Performance Computing (Germishuys et al. 2019) In the machine learning and High Performance Computing module, researchers train, test and expose state-of-the-art machine learning models. The aggregated citizen scientist annotations are used to train object-detection models that track and identify the species of interest. In our case study, we used 409 user-annotated ground-truth frames obtained from workflow 2 (Suppl. material 1) to train an algorithm to identify cold water corals. We augmented this data by using a frame tracker which filled subsequent movie frames with bounding boxes with the highest probability of containing the object of interest. This typically increased the amount of data by a factor of 10. The frames were then pre- processed to remove background distortion because colours often lose_ intensity underwater, mainly due to poor visibility. Three datasets were then created, one for training the model, another for validation (which is used to tune the model hyperparameters) and, finally, a testing set. Once the data were prepared, the model training was done until satisfactory metrics were achieved on evaluation measures (i.e. F1 = 0.970, Recall = 0.962, Precision = 0.979 and mMAP@0.5 = 0.962). An open-source, citizen science and machine learning approach to analyse ... 9 We made the trained model available through an application programming interface (API), where it can be used by researchers to run predictions of the species of interest in new recordings (Fig. 1). To this end, we used FastAPI (Ramirez 2020) as it provides the speed, scalability and reliability required to have multiple users making use of the service at the same time. The API was also supplied with a user-friendly front-end, using the Streamlit (Teixeira 2020) framework, allowing a broader audience of scientific users (i.e. ecologists, ROV and AUV-pilots, students) to access the service through a web application. The interface allows researchers to browse through already-classified footage or to upload their own footage as either images or video. Once the media has been uploaded/selected, users are able to manipulate hyperparameter thresholds (IOU threshold, confidence threshold) and interactively see the impact on the model output. The API is described by Germishuys et al. (2019). We compared manual observations of cold water corals provided by an expert to those provided by our machine learning model to estimate the accuracy of the model under different confident thresholds (Table 2). Both expert and model classified movies corresponding to 132 squares of a spatial grid within the National Park into "Coral" and "No coral" (i.e. presence/absence of cold water corals). To estimate the final classifications of the machine learning model, we aggregated the raw model output, containing coral observations for each frame under 0.5, 0.7 and 0.9 confidence thresholds, into periods in which the species was continuously observed with > 50% overlap between consecutive bounding boxes. These aggregated observation periods described the first and last frame in which coral was visible (Suppl. material 3). If aggregated observation periods were within the footage corresponding to one square, the square was classified as Coral. We used confusion matrices to estimate the accuracy of the machine-based classifications under the different thresholds. The best accuracy for our case study was achieved with a confidence threshold of 0.7. Table 2. Confusion matrices derived from applying different confidence thresholds (ML confidence) when overlaying manual with machine-based observations in movies corresponding to 132 squares of a spatial grid within the Kosterhavets National Park, Sweden. Detailed metadata for these recordings are provided in Suppl. material 3. ML confidence = 0.5 ML confidence = 0.7 ML confidence = 0.9 Coral Nocoral Coral Nocoral Coral No coral Expert Coral 54 15 52 17 28 4 No coral 13 50 5 58 1 62 The last component of this module is a data visualisation toolkit that enables researchers to explore and visualise the ecological data extracted from the outputs of the machine learning model. In our case, we mapped the cold water coral annotations provided by the expert and the machine learning model with a 0.7 confidence threshold (Fig. 6). Our results 10 Anton V etal highlight that machine learning models with a relatively high confidence threshold are well- suited for automated monitoring of cold water coral over large areas. Al vs Manual, Conf = 0.7 No data mi me oll Grid size 5m Figure 6. EESI Comparison of manual and machine learning model-based spatial distribution of cold water coral in the reef area Sacken in Kosterhavets National Park, Sweden. Spatial distribution is based on coral observations in ROV movies corresponding to 132 squares of the spatial grid. Confidence threshold (Conf) for the model is set to 0.7. Grid size 5 m. Discussion The functionalities of the Koster Seafloor Observatory have been tested in the present case study, which illustrates the scientific potential of this open-source and modular approach. Our approach can be used to extract ecological data on abundance and distribution for many benthic species from underwater recordings. Underwater footage is today routinely collected by many research institutes, which may allow for a concerted analysis of such data over broad spatial and temporal scales in the future. Such analyses may calculate data products for biological state variables on regional or even global level, so-called Essential Biodiversity Variables or EBVs (Pereira et al. 2013, Hardisty et al. 2019). A recent study by Kissling et al. (2018) suggests that image-based sensor networks are promising candidates for EBVs, while many other studies highlight the potential of these methods for marine monitoring programmes (Mack et al. 2020, Lopez-Vazquez et al. 2020). Our case study provides empirical support that these methods are ready for implementation in national monitoring programmes and that useful data products can be derived from image-based sensors, especially in marine environments which are particularly difficult to access and survey. An open-source, citizen science and machine learning approach to analyse ... 11 In order to scale up analysis of underwater imagery in the future to extract ecological data for larger regions, longer time periods and more species, several technical bottlenecks have to be addressed. Data archiving functions can fall under organisational or governmental responsibilities and may not be fulfilled by a single global system. Consequently, most underwater recordings are currently locally archived and cannot be discovered. Here, further work is needed to promote the use of open interoperable archives and data portals (e.g. European Marine Data Archive, EMODnet portal) that enable researchers to adequately publish metadata associated with underwater recordings. Another important technical bottleneck is the disconnection between many essential data services that need to interact to successfully analyse image data. We suggest that seamless links should be developed especially between citizen science platforms (for training of machine learning models) and high-performance computation services (for extracting ecological data from large amounts of imagery). Regional, national and global research infrastructures should take a leading role in this development to overcome current technical challenges. Funding: The project was funded by Ocean Data Factory, an expert network supported by grants from Sweden’s Innovation Agency (grant agreement no. 2019-02256), the Swedish Agency for Marine and Water Management (grant agreement no. 956-19) and the Swedish Research Council (through Swedish LifeWatch grant agreement no. 829-2009-6278). The presented work was furthermore supported by the NelC programme DeepDive and the Horizon 2020 project ENVRIplus (grant agreement no. 654182). Web location (URIs) Homepage: https://www.zooniverse.org/projects/victorav/the-koster-seafloor-observatory/ about/results Usage licence Usage licence: Creative Commons Public Domain Waiver (CC-Zero) IP rights notes: Our approach is open for use in research, as well as public and academic education for analysis of community composition in marine ecosystems. Acknowledgements We thank the data providers who allowed us to use movie material. These include ROV- pilots (especially Tomas Lundalv, Lisbeth Jonsson and Roger A. Johansson), as well as the data curator at the University of Gothenburg (Lars Ove Loo) and Chalmers Technical University (Ola Benderius). We acknowledge the tremendous help from taxonomic experts Thomas Dahlgren, Kennet Lundin and Bjérn Kallstr6m who actively curated the citizen science platform, as well as the ROV pilots who offered their material for use (Tomas Lunddlv, Lisbeth Jonsson and Roger A. Johansson), while Emil Burman helped with the 12 Anton V etal translation of the site. We also thank the Zooniverse team and the 2,451 citizen scientists who helped us classify the footage. We also thank the two reviewers of this manuscript for their comments and suggestions. Finally we are grateful for support by the Center for Sea and Society and the Gothenburg Global Biodiversity Center. Author contributions VA, MO, and JG conceived and designed the study. VA and MO set up and continue to maintain the Zooniverse site. VA and JG wrote the code for both Github projects (data processing workflow and model). MO worked with the public contributions to the Zooniverse site. PB and ML contributed with the data management and archiving of the original movies, the manual annotations of movies, as well as the analysis of the model results. VA, MO and JG contributed equally to the writing and revision of the manuscript. References ° Aceves-Bueno E, Adeleye A, Feraud M, Huang Y, Tao M, Yang Y, Anderson S (2017) The Accuracy of Citizen Science Data: A Quantitative Review. The Bulletin of the Ecological Society of America 98 (4): 278-290. https://doi.org/10.1002/bes2.1336 ° Anton V, Germishuys J, Obst M (2019) Koster Data Management Repository. 0.1. Github. URL: https://github.com/ocean-data-factory-sweden/koster data _ management ° Bean TP, Greenwood N, Beckett R, Biermann L, Bignell JP, Brant JL, Copp GH, Devlin MJ, Dye S, Feist SW, Fernand L, Foden D, Hyder K, Jenkins CM, van der Kooij J, Kroger S, Kupschus S, Leech C, Leonard KS, Lynam CP, Lyons BP, Maes T, Nicolaus EEM, Malcolm SJ, Mcllwaine P, Merchant ND, Paltriguera L, Pearce DJ, Pitois SG, Stebbing PD, Townhill B, Ware S, Williams O, Righton D, et al. (2017) A review of the tools used for marine monitoring in the UK: Combining historic and contemporary methods with modeling and socioeconomics to fulfill legislative needs and scientific ambitions. Frontiers in Marine Science 4 https://doi.org/10.3389/fmars.2017.00263 ° Beijoom O, Edmunds P, Roelfsema C, Smith J, Kline D, Neal B, Dunlap M, Moriarty V, Fan T, Tan C, Chan §S, Treibitz T, Gamst A, Mitchell BG, Kriegman D (2015) Towards automated annotation of benthic survey images: variability of human experts and operational modes of automation. PLOS ONE 10 (7): e0130312. https://doi.org/10.1371/ journal.pone.0130312 ° Benedetti-Cecchi L, Crowe T, Boehme L, Boero F, Christensen A, Gréemare A, Hernandez F, Kromkamp JC, Nogueira Garcia E, Petinakis G, Robidart J, Sousa Pinto |, Zingone A (2018) Strengthening Europe's capability in biological ocean observations. European Marine Board Future Science Brief 3, Ostend, Belgium. [ISBN 9789492043559] ° Costello MJ, McCrea M, Freiwald A, Lundalv T, Jonsson L, Bett BJ, van Weering TCE, de Haas H, Roberts JM, Allen D (2005) Role of cold-water Lophelia pertusa coral reefs as fish habitat in the NE Atlantic. Springer, 34 pp. [ISBN 3-540-241 36-1] https://doi.org/10.1007/3-540-27673-4 41 ° Danovaro R, Carugati L, Berzano M, Cahill AE, Carvalho S, Chenuil A, Corinaldesi C, Cristina S, David R, Dell'Anno A, Dzhembekova N, Garces E, Gasol JM, Goela P, Feral An open-source, citizen science and machine learning approach to analyse ... 13 JP, Ferrera |, Forster RM, Kurekin AA, Rastelli E, Marinova V, Miller Pl, Moncheva S, Newton A, Pearman JK, Pitois SG, Rene A, Rodriguez-Ezpeleta N, Saggiomo V, Simis SGH, Stefanova K, Wilson C, Lo Martire M, Greco S, Cochrane SKJ, Mangoni O, Borja A (2016) Implementing and innovating marine monitoring approaches for assessing marine environmental status. Frontiers in Marine Science 3: 213. https://doi.org/ 10.3389/fmars.2016.00213 Germishuys J, Anton V, Obst M (2019) Koster YOLO object detection model. 0.1. Github. URL: https://github.com/ocean-data-factory-sweden/koster_ml Guidi L, Fernandez-Guerra A, Bakker D, Canchaya C, Curry E, Foglini F, Irission JO, Malde K, Marshall CT, Obst M, Ribeiro RP, Tjiputra J (2020) Big data in marine science. Future Science Brief 6. European Marine Board, Ostend, Belgium. [ISBN 9789492043931] https://doi.org/10.5281/zenodo.3755793 Hardisty AR, Belbin L, Hobern D, McGeoch MA, Pirzl R, Williams KJ, Kissling WD (2019) Research infrastructure challenges in preparing essential biodiversity variables data products for alien invasive species. Environmental Research Letters 14 (2). https://doi.org/10.1088/1748-9326/aaf5db Kissling WD, Ahumada JA, Bowser A, Fernandez M, Fernandez N, Garcia EA, Guralnick RP, Isaac NJB, Kelling S, Los W, McRae L, Mihoub JB, Obst M, Santamaria M, Skidmore AK, Williams KJ, Agosti D, Amariles D, Arvanitidis C, Bastin L, De Leo F, Egloff W, Elith J, Hobern D, Martin D, Pereira HM, Pesole G, Peterseil J, Saarenmaa H, Schigel D, Schmeller DS, Segata N, Turak E, Uhlir PF, Wee B, Hardisty AR (2018) Building Essential Biodiversity Variables (EBVs) of species distribution and abundance at a global scale. Biological Reviews 93 (1): 600-625. https://doi.org/10.1111/brv.12359 Lavaleye M, Duineveld G, Lundalv T, White M, Guihen D, Kiriakoulakis K, Wolff GA (2009) Cold-water corals in the Tilser reef. Preliminary observations on the dynamic reef environment. Oceanography 22 (1): 76-84. httos://doi.org/10.5670/oceanog.2009.08 Lopez-Vazquez V, Lopez-Guede JM, Marini S, Fanelli E, Jonnsen E, Aguzzi J (2020) Video Image Enhancement and Machine Learning Pipeline for Underwater Animal Detection and Classification at Cabled Observatories. Sensors 20 (3). https://doi.org/ 10.3390/s20030726 Mack L, Attila J, Aylagas E, BeermannA, Borja A, Hering D, Kahlert M, Leese F, Lenz R, Lehtiniemi M, Liess A, Lips U, Mattila O, Meissner K, Pyhdalahti T, Setala O, Strehse J, Uusitalo L, Willstrand Wranne A, Birk S (2020) A Synthesis of Marine Monitoring Methods With the Potential to Enhance the Status Assessment of the Baltic Sea. Frontiers in Marine Science 7 https://doi.org/10.3389/fmars.2020.552047 Pereira HM, Ferrier S, Walters M, Geller GN, Jongman RHG, Scholes RJ, Bruford MW, Brummitt N, Butchart SHM, Cardoso AC, Coops NC, Dulloo E, Faith DP, Freyhof J, Gregory RD, Heip C, Hoft R, Hurtt G, Jetz W, Karp DS, McGeoch MA, Obura D, Onoda Y, Pettorelli N, Reyers B, Sayre R, Scharlemann JPW, Stuart SN, Turak E, Walpole M, Wegmann M (2013) Essential Biodiversity Variables. Science 339 (6117): 277-278. https://doi.org/10.1126/science.1229931 Picheral M, Colin S, Irisson J (2017) EcoTaxa, a tool for the taxonomic classification of images. http://ecotaxa.obs-vlfr.fr. Ramirez S, et al. (2020) FastAPI! framework. 0.61.1. Github. URL: https://github.com/ tiangolo/fastapi Redmon J, Farhadi A (2018) YOLOv3: An Incremental Improvement. arXiv URL: https://arxiv.org/abs/1804.02767 14 Anton V etal ° Teixeira T, et al. (2020) Streamlit. 0.69.0. Github. URL: https://github.com/streamilit/ streamlit Supplementary materials Suppl. material 1: Dataset of underwater images of Desmophyllum pertusum E{) Authors: Victor Anton, Jannes Germishuys, Per Bergstr6m, Mats Lindegarth, Matthias Obst Data type: images, zipped Brief description: Instances of Desmophyllum pertusum used to train Koster YOLO machine learning model. Download file (33.28 MB) Suppl. material 2: Metadata for movies used in the case study EU Authors: Victor Anton, Jannes Germishuys, Per Bergstr6m, Mats Lindegarth, Matthias Obst Data type: table with occurrences Brief description: This file contains metadata from the movies used to test the model and illustrate its application. To access the movie data files, contact the authors or search the filenames in the Swedish National Data Service: https://snd.gu.se/en/catalogue/study/snd1069. Download file (21.34 kb) Suppl. material 3: model results EE} Authors: Victor Anton , Jannes Germishuys , Per Bergstrém , Mats Lindegarth , Matthias Obst Data type: table Brief description: Model output from analysis of the selected movies in Supplementary material 2. Explanation of variables: FilenamelnThisStudy (movielD), frame_no_start (frame number when the object was detected for the first time), frame_no_end (frame number when the object was detected for the last time), max_conf (highest confidence value achieved by the object throughout the consecutive frames), x (x-position of the upper-left corner of the bounding box with the highest confidence value), y (y-position of the upper-left corner of the bounding box with the highest confidence value), w (width of the bounding box with the highest confidence value), h (height of the bounding box with the highest confidence value). Download file (9.54 kb)