Our findings revealed that citizen scientists, failed to achieve comparable levels of accuracy to those of experts when recording presence/absence data for various marine taxa and threats. Nevertheless, volunteer divers provided abundance estimates for threats and motile species that were aligned with expert evaluations. Significant discrepancies were observed in the volunteers’ estimations of sessile species abundance, and they were frequently susceptible to both false-positive and false-negative reporting, often misidentifying fish species that were either absent from or present at the dive sites. These outcomes underscore the importance of rigorously assessing data quality in citizen science projects or setting limits under which divers can be reliably involved in complementing professional scientific data gathering.
Divers demonstrated limited effectiveness in recognizing environmental stressors commonly included in several citizen science programs, such as instances of tissue necrosis linked to marine heatwaves and broader impacts of climate change—parameters that are critical for assessing ecosystem health and resilience (Starko et al. 2024; Trégarot et al. 2024). Conversely, volunteer divers demonstrated an improved performance in recording marine litter, anchoring, or discarded fishing gear, as these are relatively straightforward to detect and do not require specialized expertise. This likely highlights a limitation in the ability of citizen scientists to detect more complex environmental stressors or impacts that are not immediately apparent. These findings raise critical concerns regarding the potential boundaries and limitations of citizen science programmes in capturing complex ecological phenomena.
Disparities in accuracy emerged between sessile and motile species. For sessile taxa, volunteers reporting deviated significantly from those of experts in four out of the 15 taxa assessed. In contrast, discrepancies in motile taxa were primarily confined to “other sea urchins”. Abundance estimates for motile species were generally in agreement with expert scientists, whereas estimates for sessile taxa showed greater variability. These findings align with previous studies, which demonstrated that volunteer performance often varies by species type and may be influenced by individual interests (Branchini et al. 2015). For instance, macro photography enthusiasts might focus on benthic organisms, while those interested in larger species may overlook smaller or less conspicuous taxa (Meschini et al. 2021). Species detectability further complicates this dynamic, as less common or elusive species are inherently more challenging to identify, leading to more accurate data for common and easily recognizable taxa (Cox et al. 2012; Kosmala et al. 2016; Farr et al. 2023).
Our analyses revealed a significant dependence of volunteer data accuracy on participants’ diving experience and familiarity with dive sites. According to various studies (Hermoso et al. 2021; Lucrezi et al. 2018; Cerrano et al. 2017; Martin et al. 2016) experienced divers demonstrated greater precision, potentially due to enhanced confidence in their equipment and underwater skills, enabling them to concentrate more effectively on their surroundings. Additionally, familiarity with specific sites likely facilitated the recognition of local marine species and environmental features, thereby reducing errors and enhancing data reliability. We, therefore, encourage that specific diving skills and experience with the study sites should be used as the initial criteria for filtering citizen science collected data in demanding marine contexts.
Another important concern arises from the observation that divers were more prone to false-positive species identifications than to false negative ones. This tendency suggests a potential bias toward overreporting, which may reflect an eagerness to contribute or a lack of taxonomic certainty. Such patterns could compromise data reliability by inflating perceived biodiversity, thereby underscoring the need for improved training protocols and validation mechanisms within citizen science frameworks. Without rigorous validation methods, inconsistencies and observer bias can undermine the scientific credibility and usefulness of citizen science data, limiting its potential contribution to further applications such as ecological modelling, conservation planning, and policy development. Systematic evaluation not only helps identify potential sources of error but also guides the development of training protocols, data verification tools, and quality control mechanisms. This ultimately strengthens the role of citizen science in advancing scientific data production (Lukyanenko et al. 2016; Stevenson 2018; Anhalt-Depies et al. 2019).
For citizen science to make a meaningful contribution to ecological research, both the scientific community and the public must have confidence in the accuracy of the data. In fact, the proportion of published studies relying on citizen science data does not reflect the abundance and diversity of active citizen science programmes (Kullenberg and Kasperowski 2016; Davis et al. 2023), potentially due to concerns about data quality among peer reviewers (Theobald et al. 2015; Davis et al. 2023). Verification processes, whether applied to entire datasets or selected subsets, can enhance confidence in the reliability of citizen-collected data. However, implementing verification is not straightforward, often requiring means of comparison (e.g., images, videos or even the presence of experts along with citizens). Therefore, while verifying subsets of data may enable researchers to estimate error rates and identify additional sampling needs for hypothesis testing, the associated costs (time, effort, professional availability) raise questions about whether verification or comprehensive training for citizens prior to their involvement is the more effective approach. Ultimately, citizen science offers considerable prospects for advancing ecological and conservation research. Understanding, quantifying and eliminating biases in these data is an essential step towards their widespread application in addressing ecological questions and monitoring biodiversity.
Enhancing data quality in citizen science initiatives involves several strategies, including targeted training programmes, skill-based prequalification and on-going feedback for long-term participants (Kosmala et al. 2016; van der Wal et al. 2016). However, intensive training may inadvertently reduce participation, potentially limiting the educational and engagement benefits of such programs. Albeit, citizen scientists often produce data comparable to marine professionals (Forrester et al. 2015; van der Velde et al. 2017), they face greater challenges with specialized tasks, such as difficulty in species identification (Farr et al. 2023; Díaz-Calafat et al. 2024) or abundance estimation (Gillett et al. 2012; Done et al. 2017). Certain attributes are inherently more subjective or complex, further complicating data collection. Therefore, in light of the findings of our study, we caution that citizen science projects, and the users of the data collected under such projects, should carefully consider the trade-off between higher data reliability and broader public engagement.