This article provides a comprehensive guide for researchers and professionals on designing and executing effective multiple database search strategies for rigorous environmental evidence synthesis.
This article provides a comprehensive guide for researchers and professionals on designing and executing effective multiple database search strategies for rigorous environmental evidence synthesis. It addresses the critical need to minimize bias and maximize recall of relevant studies, covering foundational principles, practical methodology, common troubleshooting, and validation techniques. Tailored for those conducting systematic reviews and maps in environmental management and health, the content explores structuring searches with PICO/PECO, selecting diverse bibliographic sources, leveraging supplementary search methods, and implementing peer review to ensure transparency and reproducibility, ultimately supporting reliable and defensible research conclusions.
In environmental evidence research, the validity and reliability of any synthesis are fundamentally dependent on the comprehensiveness and rigor of the literature search process. A well-designed search strategy serves as the critical foundation for identifying all relevant studies, thereby minimizing selection bias and ensuring that subsequent conclusions are based on a complete representation of the available evidence. Research demonstrates that search strategy quality directly impacts mapping outcomes, particularly when dealing with large bodies of research where terminology may not be standardized [1]. The challenge lies in balancing sensitivity (retrieving all relevant studies) with precision (excluding irrelevant ones) while navigating practical constraints of time and resources, especially in multidisciplinary fields like environmental science where evidence may be scattered across diverse sources [2] [3].
Within the context of a broader thesis on multiple database search strategies, this article establishes the critical importance of systematic search approaches for minimizing bias in evidence synthesis. The comprehensive identification of relevant literature through robust search methods ensures that synthesis findings accurately reflect the true state of knowledge rather than representing a skewed subset of available evidence. As publication rates continue to increase across research fields, the methodological rigor applied to literature searching becomes increasingly vital for valid research synthesis [1].
Table 1: Database performance in retrieving unique references in systematic reviews
| Database | Unique References Retrieved | Percentage of Total Unique References | Key Strengths |
|---|---|---|---|
| Embase | 132 | 45.4% | Comprehensive biomedical literature, strong international coverage |
| MEDLINE | 67 | 23.0% | Premier biomedical database, strong subject indexing |
| Web of Science Core Collection | 53 | 18.2% | Multidisciplinary coverage, citation indexing |
| Google Scholar | 39 | 13.4% | Grey literature, theses, conference proceedings |
| Other specialized databases | Varies by topic | Varies | Topic-specific coverage (e.g., CINAHL, PsycINFO) |
Data adapted from a prospective exploratory study analyzing 58 published systematic reviews totaling 1,746 relevant references identified through database searches [4]. The study found that 16% of included references (291 articles) were uniquely found in a single database, highlighting the importance of searching multiple sources.
Table 2: Overlap of studies in evidence bases on nutrient recovery research
| Evidence Base | Primary Research Focus | Number of Studies (2013-2017) | Overlap with Other Evidence Bases | Key Methodological Differences |
|---|---|---|---|---|
| SA | Distinct nutrient recovery options | Lower coverage | Limited | Screening stopped once no new options emerged |
| BR | Domestic wastewater broadly | Moderate coverage | Approximately 10% with similar scope | Required explicit mention of intended reuse |
| UM | Human urine only | Higher coverage for specific stream | Variable | Covered conceptual studies beyond technologies |
| EW | Domestic wastewater broadly | Higher coverage | Core component for comparison | Single compound search string |
| EB | Expanded coverage | Highest coverage | Benchmark for comparison | Additional targeted searches by subdomain |
Data synthesized from comparison of five evidence bases on recovery and reuse of nutrients found in human excreta and domestic wastewater [1]. The analysis revealed surprisingly low overlap between evidence bases compiled through different search methodologies, with only about 10% of studies appearing in both of two major evidence bases even after correcting for differences in scope and time period.
The process of developing an effective search strategy requires careful consideration of multiple factors that influence both comprehensiveness and efficiency. According to evidence synthesis experts, search strategies must consider: "The scope of the research question/topic and the intended use of the synthesis; The time constraints for conducting the review; The inclusion and exclusion criteria; The expected data types to be synthesized; The study types that will be considered" [2]. This comprehensive approach ensures the search strategy is appropriately tailored to the specific synthesis objectives while remaining feasible within resource constraints.
The iterative development of search strings represents a critical success factor in minimizing bias. This process involves continuous testing and refinement of search terms to achieve optimal balance between sensitivity and precision [2]. In environmental research where terminology is often not standardized, the development of a sensitive and specific search strategy becomes particularly challenging yet essential [1]. Research has identified an issue described as "differential search term sensitivity and specificity," where compound search terms do not perform equally well across all subdomains of a research topic, necessitating tailored approaches for different aspects of complex environmental questions [1].
Empirical evidence indicates that optimal literature searches for systematic reviews should include at least Embase, MEDLINE, Web of Science Core Collection, and Google Scholar as a minimum requirement to achieve adequate coverage [4]. This combination achieved an overall recall of 98.3% and 100% recall in 72% of systematic reviews analyzed in a prospective study. The research further suggests that approximately 60% of published systematic reviews may fail to retrieve 95% of all available relevant references due to insufficient database searching [4].
Specialized databases should be added when the review topic aligns with their focus, as they contribute unique references not found in major multidisciplinary databases [4]. For public health and environmental topics, this may require searching a wider range of databases due to the multidisciplinary nature of the evidence [3]. The database combination should be carefully selected based on topic specificity, with recognition that a "one size fits all" approach is not appropriate for complex environmental questions [3].
Objective: To minimize selection bias in evidence synthesis through systematic and comprehensive literature retrieval across multiple databases.
Materials and Equipment:
Procedure:
Quality Control Measures:
Objective: To minimize publication bias by identifying relevant studies not published in traditional commercial journals.
Materials and Equipment:
Procedure:
Quality Control Measures:
Search Workflow Diagram: This flowchart illustrates the comprehensive literature search process from question definition to final study set identification, highlighting critical stages for minimizing bias.
Table 3: Essential research tools and platforms for comprehensive literature searching
| Tool Category | Specific Examples | Primary Function | Role in Minimizing Bias |
|---|---|---|---|
| Bibliographic Databases | Embase, MEDLINE, Web of Science Core Collection | Comprehensive publication indexing | Ensure broad coverage of peer-reviewed literature |
| Multidisciplinary Platforms | Google Scholar, Scopus | Cross-disciplinary search | Capture research outside core subject databases |
| Citation Management Software | EndNote, Zotero, Mendeley | Reference organization and deduplication | Enable efficient management of large result sets |
| Systematic Review Tools | Rayyan, Covidence, EPPI-Reviewer | Screening and data extraction workflow | Standardize review process and reduce screening errors |
| Grey Literature Resources | Government databases, institutional repositories, clinical trial registries | Unpublished and hard-to-locate evidence | Counter publication bias and location bias |
| Text Mining Applications | Voyant Tools, AntConc, VOSviewer | Terminology analysis and expansion | Improve search term sensitivity through text analysis |
These essential tools form the foundation of a comprehensive search strategy capable of minimizing various forms of bias in evidence synthesis. Proper utilization of these resources, adapted from recommendations across multiple sources [1] [4] [2], enables researchers to achieve the balance between sensitivity and specificity required for valid and reliable evidence synthesis.
In the realm of evidence-based research, particularly in fields like environmental evidence and drug development, the efficiency and comprehensiveness of literature searching are paramount. A foundational understanding of core search terminologyâsearch terms, search strings, and search strategiesâis the first critical step in ensuring that research is built upon a complete and unbiased body of evidence. This is especially true for systematic reviews and other rigorous research syntheses, which require meticulous documentation and reproducible methodologies [5]. Misunderstanding these concepts can lead to incomplete searches, biased results, and ultimately, flawed conclusions. This document provides detailed application notes and protocols for defining and employing these key elements within the context of searching multiple databases for environmental evidence research.
The following table summarizes the key terminology that forms the foundation of effective database searching.
Table 1: Core Search Terminology
| Term | Definition | Core Function | Example |
|---|---|---|---|
| Search Term [6] [7] | A single word or short phrase entered into a search engine or database to retrieve information. | The basic building block of a search; represents a key concept. | pesticide, degradation |
| Search String [8] | A combination of search terms, numbers, and special characters (e.g., Boolean operators, truncation) submitted to a search engine. | Translates a single search concept into a formal query the database can execute. | degrad* OR breakdown OR decomposition |
| Search Strategy [9] [10] | An organised structure of key terms and protocols used to search a database. It accounts for all possible search terms, keywords, phrases, and their variations. | The master plan that ensures a comprehensive, systematic, and reproducible search process. | The complete protocol for a systematic review, including databases searched, all search strings for all concepts, and limits applied. |
The relationship between these components is hierarchical and integrated, as shown in the following workflow.
This protocol outlines the step-by-step process for constructing a robust search strategy suitable for systematic reviews and other in-depth research syntheses in environmental evidence.
3.1.1 Objective: To create a comprehensive, transparent, and reproducible search strategy for retrieving relevant literature from multiple bibliographic databases.
3.1.2 Research Reagent Solutions:
Table 2: Essential Tools for Search Strategy Development
| Tool / Resource | Function | Example / Application in Environmental Evidence |
|---|---|---|
| Bibliographic Databases | Host scholarly literature and use specific indexing rules. | Web of Science, Scopus, PubMed, EMBASE, Environment Complete, GreenFILE [11] [5]. |
| Thesauri & Controlled Vocabularies | Provide standardized subject headings to find articles by topic, not just author words. | MeSH (for MEDLINE/PubMed), Emtree (for EMBASE). Use to find "Conservation of Natural Resources" instead of "nature management" [9] [10]. |
| Search Syntax Tools | Symbols that enable searching for word variations and phrases. | Truncation (pesticid* for pesticide, pesticides), Wildcards (behavio?r for behavior, behaviour), Phrase searching ("climate change") [9] [12] [10]. |
| Boolean Operators | Logical connectors (AND, OR, NOT) that combine search terms to broaden or narrow results. | (agriculture OR farming) AND (water quality) [8] [12]. |
| Reference Management Software | Software to store, organize, and deduplicate search results. | EndNote, Zotero, Mendeley. |
3.1.3 Methodology:
riparian buffers, nitrate, pollution reduction, and rivers.riparian buffer, riparian zone, streamside vegetation, buffer strip, "riparian forest".nitrate, NO3, nitrogen.("riparian buffer" OR "riparian zone" OR "buffer strip" OR "streamside vegetation")(nitrat* OR NO3 OR nitrogen)(String for Concept 1) AND (String for Concept 2) AND (String for Concept 3)...3.2.1 Objective: To adapt a core search strategy for effective use across different bibliographic databases, ensuring comprehensiveness while respecting the unique features of each platform.
3.2.2 Methodology:
*, ?, #) via the database's "Help" section. [9]ADJ3 in Ovid) to search for terms near each other. [9] [10][tiab] for title/abstract in PubMed).Riparian Buffers("riparian buffer"[tiab] OR "riparian zone"[tiab] OR "buffer strip"[tiab] OR "streamside vegetation"[tiab]) OR "Riparian Zones"[Mesh]The following table summarizes the key operators and symbols used in constructing search strings and their quantitative impact on search results.
Table 3: Search Syntax and Their Effects on Results
| Operator/Symbol | Function | Syntax Example | Effect on Result Set Size | Notes & Database Variability |
|---|---|---|---|---|
| Boolean AND [8] [12] | Narrows search; finds records with ALL terms. | nitrate AND groundwater |
Decreases | Fundamental for combining distinct concepts. |
| Boolean OR [8] [12] | Broadens search; finds records with ANY term. | (river OR stream OR watershed) |
Increases | Used to group synonyms and related terms for a single concept. |
| Boolean NOT [8] [12] | Excludes records containing a term. | aquatic NOT marine |
Decreases | Use with caution; can inadvertently exclude relevant records. |
| Truncation (*) [9] [12] | Finds multiple word endings. | pesticid* finds pesticide, pesticides. |
Increases | Symbol may vary; check database guide. |
| Wildcard (?) [9] [10] | Replaces a single character. | behavio?r finds behavior, behaviour. |
Increases | Useful for British/American spellings. |
| Phrase Search (" ") [8] [10] | Finds exact phrase. | "invasive species" |
Decreases | Increases relevance by preventing term separation. |
| Proximity (ADJ#) [9] [10] | Finds terms within a specified number of words. | (soil ADJ3 contamination) |
Varies (typically decreases) | Highly database-specific. Powerful for improving relevance. |
The following diagram deconstructs a sample search string to illustrate how Boolean operators and grouping create the final logical query executed by the database.
In environmental evidence research, which often involves complex, interdisciplinary topics, applying these protocols with precision is critical. Systematic maps and reviews, as published in Environmental Evidence, require searches that are comprehensive to minimize bias and fully capture the available literature on a topic like "impacts of airborne anthropogenic noise on wildlife." [13]
Furthermore, the use of multiple databases is non-negotiable. Relying on a single database risks missing a significant portion of the evidence base. [5] After retrieval, evidence reviews in this field are often appraised for reliability using tools like the CEESAT (Collaboration for Environmental Evidence Synthesis Assessment Tool), which evaluates the transparency and comprehensiveness of the search strategy itself. A "Gold" rating requires a search strategy that meets the highest standards of conduct and reporting, including the thoughtful application of the terms and strategies defined in this document. [14]
In environmental evidence research, the integrity of systematic reviews and meta-analyses is fundamentally dependent on the comprehensiveness and impartiality of the literature search process. Searches conducted for evidence synthesis must be transparent, reproducible, and designed to minimise biases, as failing to include relevant information can lead to inaccurate or skewed conclusions [15]. This application note examines three pervasive search biasesâpublication, language, and temporal biasâthat threaten the validity of evidence syntheses. We frame this discussion within the context of multiple database search strategies, providing environmental evidence researchers and drug development professionals with structured protocols to identify, quantify, and mitigate these biases throughout the research lifecycle.
Search biases represent systematic errors in literature identification and selection that can significantly affect evidence synthesis outcomes. The Collaboration for Environmental Evidence (CEE) Guidelines emphasize that biases linked to the search itself must be minimized and/or highlighted as they may affect synthesis outputs [15]. Within environmental evidence research, several distinct bias types require specific consideration:
Publication Bias: An asymmetry in the likelihood of publishing results where statistically significant (positive) results are more likely to be accepted for publication than non-significant ones (negative results) [15] [16]. This bias has been a source of major concern for systematic reviews and meta-analysis as it might lead to overestimating an effect/impact of an intervention or exposure on a population.
Language Bias: Occurs when studies with significant or 'interesting' results are more likely to be published in English and are easier to access than results published in other languages [15] [16]. Recent evidence demonstrates that excluding non-English-language studies may significantly bias ecological meta-analyses, sometimes changing the direction of mean effect sizes [17].
Temporal Bias: Encompasses the risk that studies supporting a hypothesis are more likely to be published first, with results potentially not supported by later studies [15]. This bias also manifests when researchers overlook older publications due to a 'latest is best' culture, potentially perpetuating misinterpretations.
Table 1: Characteristics of Major Search Biases in Environmental Evidence Research
| Bias Type | Primary Mechanism | Impact on Evidence Synthesis | Common Research Contexts |
|---|---|---|---|
| Publication Bias | Selective publication of statistically significant results | Overestimation of effect sizes; skewed summary estimates | Clinical trials; intervention studies; experimental ecology |
| Language Bias | Systematic differences in study characteristics and results between languages | Altered direction or magnitude of overall mean effect sizes | Biodiversity conservation; ecosystem management; social ecology |
| Temporal Bias | Time-dependent publication patterns and preference for recent studies | Perpetuation of early findings without validation; loss of historical context | Long-term ecological monitoring; climate change research; emerging contaminants |
A systematic assessment of language bias in ecological meta-analyses revealed substantial differences in effect-size estimates between English- and Japanese-language studies. In half of the eligible meta-analyses examined, effect sizes differed significantly between language groups, causing considerable changes in overall mean effect sizes and even their direction when non-English-language studies were excluded [17]. These differences were attributable to systematic variations in reported statistical results and associated study characteristics, particularly taxa and ecosystems, between language groups.
Table 2: Impact of Language Bias on Meta-Analysis Findings
| Meta-Analysis Topic | Percentage Change in Effect Size | Change in Statistical Significance | Key Differing Study Characteristics |
|---|---|---|---|
| Freshwater pollution impacts | +20.3% when excluding Japanese studies | Non-significant â Significant | Wastewater source; measurement methods |
| Marine reserve effectiveness | -32.7% when excluding Japanese studies | Significant â Non-significant | Governance type; monitoring duration |
| Forest management effects | +15.1% when excluding Japanese studies | Maintained significance | Tree species; silvicultural methods |
| Agricultural intervention outcomes | -8.9% when excluding Japanese studies | Maintained non-significance | Crop types; farm size; soil properties |
Statistical assessments of publication bias indicate its persistent presence across environmental research domains. While comprehensive quantitative data specific to environmental sciences remains limited, methodological studies suggest that publication bias may affect 20-40% of meta-analyses in ecological and environmental research, potentially inflating effect size estimates by 15-30% compared to unbiased estimates [15].
A rigorous search strategy forms the foundation for minimizing biases in evidence synthesis. The process should be carefully planned and documented to ensure transparency and reproducibility [15] [16].
Protocol 4.1: Structured Search Strategy Development
Objective: To develop a comprehensive search strategy that minimizes publication, language, and temporal biases through systematic planning and documentation.
Materials:
Procedure:
Troubleshooting:
Relying on a single database introduces significant bias risk, as no single resource provides comprehensive coverage of relevant literature [18]. Multiple database searching is essential for minimizing bias.
Diagram 1: Multi-database search workflow for comprehensive coverage
Protocol 4.2: Implementation of Multiple Database Searches
Objective: To execute comprehensive searches across multiple information sources while minimizing introduction of database-specific biases.
Materials:
Procedure:
Quality Control:
Grey literature addressing includes crucial evidence that mitigates publication bias by including studies with non-significant results [18].
Protocol 4.3: Systematic Grey Literature Search
Objective: To identify and incorporate relevant grey literature that may contain non-significant results or context-specific evidence missing from commercial databases.
Materials:
Procedure:
Considerations:
Table 3: Research Reagent Solutions for Search Bias Mitigation
| Tool Category | Specific Resources | Primary Function | Bias Addressed |
|---|---|---|---|
| Bibliographic Databases | Scopus, Web of Science, MEDLINE, GreenFILE, AGRICOLA | Comprehensive literature identification | Publication, Temporal |
| Regional/Language Databases | CiNii (Japanese), SciELO (Regional), CNKI (Chinese) | Non-English literature retrieval | Language |
| Grey Literature Sources | OpenGrey, Institutional Repositories, Government Databases | Unpublished/less formally published evidence | Publication |
| Citation Tracking Tools | Citationchaser, Google Scholar, Connected Papers | Forward and backward citation searching | Publication, Temporal |
| Search Validation Tools | PRISMA-S Checklist, Search Summary Tables | Search methodology documentation and evaluation | All Biases |
| Test-list Development | Benchmark articles, Expert consultation | Search strategy validation | All Biases |
A systematic approach to bias assessment throughout the evidence synthesis process enables researchers to identify, quantify, and account for potential biases.
Diagram 2: Integrated bias assessment throughout evidence synthesis
Effective management of publication, language, and temporal biases requires deliberate, systematic approaches throughout the evidence synthesis process. By implementing the protocols and utilizing the toolkit presented in this application note, environmental evidence researchers can significantly enhance the reliability and validity of their findings. The integration of multiple database strategies, comprehensive grey literature searching, and intentional inclusion of non-English language sources represents a minimum standard for rigorous evidence synthesis in environmental research. Future methodological developments should focus on standardized bias assessment metrics and more efficient approaches to managing the increasing volume of relevant evidence across languages and publication types.
Formulating a precise and structured research question is a critical first step in any systematic review or evidence-based research project. A well-framed question defines the scope of the research, guides the search strategy, and determines the inclusion and exclusion criteria for studies. Within environmental evidence research and related fields, the PICO (Population, Intervention, Comparator, Outcome) and PECO (Population, Exposure, Comparator, Outcome) frameworks provide systematic approaches for breaking down clinical or research questions into searchable components [19]. While PICO was originally developed for clinical intervention questions, PECO has been adapted specifically for environmental and occupational health research where investigators examine associations between exposures and health outcomes rather than planned interventions [20] [19]. This application note provides detailed protocols for effectively translating research questions into PICO/PECO elements within the context of multiple database search strategies for environmental evidence research.
The PICO framework, introduced in 1995 by Richardson et al., breaks down clinical questions into searchable keywords [19]. The mnemonic represents:
PICO is most suitable for intervention studies, particularly randomized controlled trials, and forms the foundation for many systematic reviews in clinical medicine [19].
The PECO framework adapts the PICO structure for questions exploring associations between exposures and health outcomes:
PECO is particularly valuable in environmental health research, where the focus is often on evaluating whether an exposure is associated with a health outcome rather than testing a planned intervention [20].
The choice between PICO and PECO depends primarily on the nature of the research question:
Table 1: Framework Selection Guide Based on Research Question Type
| Research Question Type | Recommended Framework | Key Applications |
|---|---|---|
| Clinical Interventions | PICO | Therapy efficacy, treatment comparisons, clinical decision-making |
| Environmental Exposures | PECO | Chemical risk assessment, occupational health, environmental epidemiology |
| Context-Dependent Questions | PICOC | Health policy, service delivery, economic evaluations |
| Qualitative Research | SPICE | Patient experiences, attitudes, opinions |
In environmental evidence synthesis, precisely defining each PECO element is essential for creating reproducible search strategies and study inclusion criteria:
Population Specification: Define human populations by demographic characteristics, health status, occupational categories, or sensitive subpopulations. For animal studies, specify species, strain, sex, and developmental stages [21]. Population definitions should align with the research question's scope while remaining practical for literature searching.
Exposure Characterization: Specify the chemical or physical agent using standard identifiers (CAS numbers), forms (salts, metabolites), exposure routes (oral, inhalation, dermal), duration patterns (acute, chronic), and exposure metrics (environmental concentrations, biomonitoring data) [21]. The exposure definition should encompass all relevant pathways by which the population encounters the agent.
Comparator Formulation: Define appropriate comparison groups based on exposure levels (e.g., low vs. high exposure, exposed vs. unpopulations), exposure timing, or case-control designs [21]. The comparator should provide a meaningful reference for assessing exposure-outcome associations.
Outcome Measurement: Specify clinically meaningful or biologically relevant endpoints, including disease incidence, mortality, functional impairments, or subclinical markers [21]. Consider measurement methods, timing, and validation when defining outcomes.
Research context influences how PECO questions are structured. The framework can be applied across different scenarios that reflect varying levels of prior knowledge about exposure-outcome relationships [20]:
Table 2: PECO Formulation Scenarios with Examples from Environmental Health
| Scenario | Research Context | PECO Approach | Example |
|---|---|---|---|
| 1 | Exploring exposure-outcome relationships | Examine incremental effects across exposure range | Among newborns, what is the effect of a 10 dB increase in noise during gestation on postnatal hearing impairment? [20] |
| 2 | Evaluating exposure cut-offs using data-derived values | Compare highest vs. lowest exposure groups (tertiles, quartiles) | Among newborns, what is the effect of the highest dB exposure compared to the lowest dB exposure during pregnancy on postnatal hearing impairment? [20] |
| 3 | Assessing known exposure standards | Apply externally-defined cut-offs from other populations | Among commercial pilots, what is the effect of occupational noise exposure compared to noise exposure in other occupations on hearing impairment? [20] |
| 4 | Evaluating regulatory thresholds | Use established exposure limits associated with health outcomes | Among industrial workers, what is the effect of exposure to <80 dB compared to â¥80 dB on hearing impairment? [20] |
| 5 | Assessing exposure reduction interventions | Select comparators based on achievable exposure reductions | Among the general population, what is the effect of an intervention that reduces noise levels by 20 dB compared to no intervention on hearing impairment? [20] |
Purpose: To translate PECO elements into comprehensive, reproducible search strategies across multiple bibliographic databases.
Materials:
Methodology:
Element Deconstruction:
Search String Formulation:
Database-Specific Adaptation:
Search Documentation:
Validation:
Purpose: To establish a systematic, reproducible process for screening and selecting studies based on PECO-defined criteria.
Materials:
Methodology:
Criteria Specification:
Pilot Screening:
Formal Screening Process:
Selection Documentation:
Quality Assurance:
The following diagram illustrates the systematic process of translating a PECO research question into executable database search strategies:
Diagram 1: PECO to Search Strategy Translation Workflow
The following diagram visualizes the structure of a comprehensive multi-database search strategy based on PECO elements:
Diagram 2: Database Search Strategy Architecture
Table 3: Essential Research Tools for Systematic Evidence Synthesis
| Tool Category | Specific Solutions | Function in PECO-Based Research |
|---|---|---|
| Bibliographic Databases | PubMed/MEDLINE, Embase, Web of Science, Scopus, TOXLINE | Comprehensive literature retrieval across multiple sources with specialized indexing [20] |
| Search Translation Tools | Polyglot Search Translator, SR-Accelerator | Adaptation of search strategies across database interfaces and syntax requirements |
| Systematic Review Software | Covidence, Rayyan, DistillerSR | Streamlined screening, selection, and data extraction processes with dual-reviewer functionality |
| Reference Management | EndNote, Zotero, Mendeley | Deduplication, citation organization, and bibliography generation |
| Chemical Identification | CAS Registry, PubChem, ChemIDplus | Standardized chemical nomenclature for precise exposure terminology [21] |
| Data Extraction Tools | Systematic Review Data Repository (SRDR), CADIMA | Structured data collection from included studies with custom form creation |
| Quality Assessment Instruments | ROBINS-I, Cochrane Risk of Bias, OHAT toolkits | Critical appraisal of individual studies for risk of bias evaluation [20] |
| Evidence Integration Platforms | Health Assessment Workspace Collaborative (HAWC), IRIS Submit | Organization and synthesis of evidence streams for hazard assessment |
Environmental health research often involves complex exposure scenarios that require special consideration in PECO formulation:
Mixed Exposures: When populations experience multiple concurrent exposures, consider whether to focus on the specific exposure of interest while accounting for potential confounding by other exposures, or to include studies of relevant mixtures with appropriate extraction strategies [21].
Exposure Biomarkers: When using biomonitoring data (e.g., chemical levels in blood or urine), clearly specify whether the biomarker represents recent exposure, cumulative burden, or susceptibility, as this affects comparator group definition [21].
Temporal Relationships: Consider the timing of exposure relative to outcome development, including critical exposure windows, latency periods, and acute versus chronic effects when defining PECO elements.
PECO formulation directly supports subsequent evidence assessment processes in systematic reviews:
Risk of Bias Assessment: Well-defined PECO elements facilitate appropriate application of risk of bias tools specific to different study designs (e.g., ROBINS-I for non-randomized studies, Cochrane tool for randomized trials) [20].
Evidence Grading: Clear PECO specifications enable consistent evaluation of domains used in evidence grading systems (e.g., GRADE), including assessment of indirectness, imprecision, and inconsistency across studies.
Hazard Identification: In environmental health assessments, precise PECO definitions support transparent determinations about the strength of evidence for causal relationships between exposures and outcomes [20] [21].
The protocols and application notes presented here provide a foundation for implementing PICO/PECO frameworks in environmental evidence research. Proper application of these structured approaches enhances the reproducibility, comprehensiveness, and validity of literature searches and evidence syntheses, ultimately supporting more reliable public health and regulatory decisions.
In the field of environmental evidence research, the volume of scientific literature is increasing annually, making comprehensive research synthesis both more critical and more challenging [1]. Within this context, the role of an information specialist becomes indispensable for navigating large bodies of research effectively. These professionals bring specialized expertise in developing systematic search strategies, managing complex datasets, and ensuring methodological rigor throughout the evidence synthesis process [2]. This application note details the essential protocols for integrating an information specialist into environmental evidence research teams, with a specific focus on executing multiple database search strategies that balance sensitivity with precision.
The value of methodological rigor in evidence synthesis is quantitatively demonstrated through comparative analysis of mapping outcomes. Research indicates that different approaches to evidence mapping on similar topics can yield surprisingly low overlap in included studies, with one analysis finding only approximately 10% of studies featured in both evidence bases despite similar scope and time periods [1]. The following table summarizes key quantitative findings from evidence synthesis comparisons:
Table 1: Comparative Outcomes of Evidence Synthesis Approaches
| Evidence Base | Studies Screened | Studies Coded | Screening Approach | Key Limitations |
|---|---|---|---|---|
| BR | Not specified | Not specified | Team of 4 reviewers | Lower study coverage |
| SA | Not specified | Not specified | Subset screening stopped at saturation | Fewer studies covered |
| EW | >150,000 | >15,000 | Single reviewer | Potential consistency issues |
| EB | >150,000 | >15,000 | Single reviewer | Limited consistency checking |
The substantial resource requirements for comprehensive evidence synthesis are further highlighted by projects involving screening of over 150,000 records and coding of over 15,000 studies [1]. These findings underscore the critical need for specialized search expertise to ensure comprehensive and unbiased evidence coverage.
Information specialists provide essential competencies that function as "research reagents" within the evidence synthesis process. The following table details these key specialized skills and their functions:
Table 2: Essential Research Reagent Solutions in Evidence Synthesis
| Research Reagent | Function in Evidence Synthesis |
|---|---|
| Search Strategy Development | Formulates comprehensive search strings balancing sensitivity and precision [2] |
| Database Schema Knowledge | Navigates platform-specific functionalities and metadata fields effectively [2] |
| Gray Literature Sourcing | Identifies and retrieves non-traditional publications from governmental, NGO, and research sources [2] |
| Citation Management | Manages and deduplicates large volumes of bibliographic data (10,000-100,000+ records) [2] |
| Metadata Enhancement | Cleans and enhances variable-quality bibliographic data for screening and analysis [2] |
| Terminology Mapping | Addresses challenges of non-standardized terminology across subdomains [1] |
These specialized competencies enable information specialists to mitigate common pitfalls in evidence synthesis, including differential search term sensitivity where compound search terms perform unevenly across different subdomains of research [1].
The protocol employs a structured approach to search string development using population, intervention, and outcome terms in the format: ãpopulation termsã AND ãintervention termsã AND ãoutcome termsã [1]. Each conceptual component incorporates multiple synonymous terms to enhance sensitivity.
Figure 1: Search Strategy Development Workflow
Information specialists implement rigorous quality control measures throughout the screening and coding process. The protocol includes:
Comprehensive documentation creates an audit trail for the entire search process:
With the emergence of AI tools in evidence synthesis, information specialists play a critical role in their responsible implementation. The protocol requires:
Figure 2: AI Integration Quality Control Process
The integration of an information specialist as a core member of environmental evidence research teams provides methodological rigor essential for trustworthy syntheses. Through development of comprehensive search strategies, management of complex data workflows, and implementation of quality control measures, these professionals directly address the challenges posed by increasingly large bodies of research. The protocols outlined herein provide a framework for leveraging their specialized expertise to enhance the comprehensiveness, efficiency, and reliability of evidence synthesis in environmental research. As the field continues to evolve with new technological capabilities, the human expertise of information specialists remains essential for navigating the complexities of multiple database search strategies while maintaining the integrity of the synthesis process.
The rigor and comprehensiveness of environmental evidence research, including systematic reviews and maps, are fundamentally dependent on a well-considered database search strategy. Library databases provide access to well-organized, carefully selected, and often peer-reviewed content, offering researchers far greater control than general search engines like Google, which only skim the surface of the open web [23]. A transparent and reproducible search plan, detailed in a research protocol, is a cornerstone of the systematic review process, minimizing bias and ensuring consistency [24] [25]. This document outlines core environmental and multidisciplinary databases, provides structured search methodologies, and integrates this process within the broader context of developing a robust research protocol for environmental evidence synthesis.
For research focused on environmental topics, several specialized databases offer deep coverage of the relevant literature. The following table summarizes key resources.
Table 1: Core Databases for Environmental Science Research
| Database Name | Primary Focus | Key Features |
|---|---|---|
| Environment Complete [26] | Environmental Science | A comprehensive database offering deep coverage in environmental disciplines. |
| GreenFILE [26] [27] | Environmental Topics | Focuses on the relationship between humans and the environment. |
| AGRICOLA [27] | Agriculture & Related Fields | Covers literature related to agriculture, forestry, and allied disciplines. |
When conducting a thorough literature review, it is essential to search multiple databases because each differs in its coverage of journals and other publication types [26]. For a preliminary search, such as for a discussion paper, one or two databases may suffice. However, for a comprehensive systematic review, searching across multiple specialized and multidisciplinary sources is necessary to ensure all relevant journals are covered [26]. Platforms like EBSCOhost allow simultaneous searching across multiple databases it hosts, such as Environment Complete, GreenFILE, and Academic Search Complete, providing a more efficient way to cover a wider selection of literature [26].
Multidisciplinary databases are critical for environmental evidence research because many related topics intersect with fields such as public health, sociology, economics, and engineering. The following table lists pivotal multidisciplinary resources.
Table 2: Key Multidisciplinary Databases for Environmental Research
| Database Name | Scope | Notable Strengths |
|---|---|---|
| Web of Science Core Collection [26] [28] | Multidisciplinary | A trusted, publisher-neutral citation database covering over 22,000 peer-reviewed journals across 254 subjects, with powerful citation analysis tools [28]. |
| Academic Search Complete [26] | Multidisciplinary | A large multidisciplinary database providing peer-reviewed full-text journals, magazines, and newspapers. |
| CORE [29] | Open Access Research | The world's largest collection of open access research papers, aggregating data from global repositories and journals. |
Databases like Web of Science Core Collection are particularly valuable for their comprehensive citation data, which allows researchers to explore connections between ideas, track the influence of research, and identify emerging fields [28]. Its consistent and accurate indexing over decades makes it a reliable foundation for research discovery and impact assessment [28].
A structured approach to searching, documented in a pre-established protocol, is essential for minimizing bias and ensuring the review is systematic and reproducible [25]. The workflow below outlines the key stages from protocol development to search execution.
The search process begins long before entering terms into a database. As shown in the workflow, the initial step involves developing a detailed protocol that defines the research question and outlines the methodology [24] [25]. This protocol should be registered in a dedicated registry, such as PROCEED for environmental sciences, to promote transparency, prevent duplication, and allow for peer review of the methods [25]. The protocol explicitly describes how the search will be executed, including the databases that will be searched, the search terms and strings, and any planned search limits [24].
Effective database searching requires specific techniques that differ from internet search engines. Mastering these techniques allows for precise control over search results.
AND to narrow results by requiring all connected terms (e.g., Diabetes AND exercise), OR to broaden results by including any of the connected terms (e.g., Dementia OR Alzheimer's), and NOT to exclude terms (e.g., "alternative energy" AND infrastructure NOT solar) [23] [30] [27]. Use NOT with caution, as it can inadvertently exclude relevant records [26]."climate change", "environmental studies") to instruct the database to search for the words in that exact order [30] [27].*) to find variant endings of a word. For example, ecolog* will retrieve records containing ecology, ecologist, and ecological [27].( ) to group related terms combined with OR and then combine them with other concepts using AND. Example: ("climate change" OR "global warming") AND policy [30].feminist W/4 ecology finds feminist within four words of ecology in that order; in Agricola, feminist N/4 ecology finds the words within four words of each other in any order) [27].The following table details essential "research reagents" â the key tools and concepts required for executing a successful database search in environmental evidence research.
Table 3: Essential Research Reagents for Database Searching
| Tool/Concept | Function | Application Example |
|---|---|---|
| Boolean Operators (AND, OR, NOT) [23] [27] | Logically combines search terms to broaden or narrow results. | ("water quality" OR "water pollution") AND agriculture |
| Phrase Searching (" ") [30] [27] | Ensures terms are searched as an exact phrase, increasing relevance. | "coral bleaching", "Aldo Leopold" |
| Truncation (*) [citaton:8] | Retrieves various word endings from a root word, expanding search coverage. | sustain* finds sustainable, sustainability, sustaining. |
| Subject Headings [31] | Uses the database's controlled vocabulary to find articles on a topic, improving precision. | Using official database subject terms instead of keywords. |
| Citation Chaining [30] | Uses reference lists and "cited by" data to find relevant literature forwards and backwards in time. | Using Google Scholar's "Cited by" feature to find newer related papers. |
| Fluphenazine-d8 | Fluphenazine-d8, MF:C22H26F3N3OS, MW:445.6 g/mol | Chemical Reagent |
| hCAIX-IN-8 | hCAIX-IN-8, MF:C19H16N4O6, MW:396.4 g/mol | Chemical Reagent |
Citation chaining is a powerful technique to expand your literature base. It involves moving backwards by reviewing the reference list of a key article and moving forwards by using tools like Google Scholar's "Cited by" feature to find newer publications that have cited the original work [30]. This is particularly useful when initial database searches yield insufficient results.
The database selection and search strategy are integral components of a systematic review protocol. Adhering to established guidelines ensures the review's credibility and utility for policymakers.
A robust protocol must detail the eligibility criteria (inclusion/exclusion), the search strategy (databases, keywords, date limits), the screening process, and the data extraction and synthesis plans [24] [25]. For environmental evidence synthesis, the Collaboration for Environmental Evidence (CEE) guidelines are a key organizing body [32] [25]. When reporting the review, follow standards like ROSES (Reporting standards for Systematic Evidence Syntheses), which are required by journals like Environmental Evidence [24]. Registering the protocol in a repository like PROCEED is considered a best practice and is often a prerequisite for journal publication [25].
In the realm of environmental evidence research, the ability to conduct comprehensive, unbiased literature searches across multiple databases is fundamental to robust evidence synthesis. Boolean operatorsâAND, OR, and NOTâform the cornerstone of systematic search strategies, enabling researchers to navigate vast quantities of scientific literature with precision [33]. These operators, based on a system of logic developed by mathematician George Boole, function as specific commands that expand or narrow search parameters when using databases or search engines [33] [34]. For complex reviews, such as those mapping large bodies of research on topics like nutrient recovery from wastewater, a well-crafted Boolean search strategy is paramount, particularly when terminology lacks standardization and resources are limited [1]. This guide provides detailed protocols for constructing effective search strings that ensure both comprehensive coverage and methodological efficiency in environmental evidence synthesis.
The three primary Boolean operators serve distinct functions in refining search results. Understanding their individual and combined effects is crucial for developing effective search strategies.
Table 1: Core Boolean Operators and Their Functions
| Operator | Function | Effect on Results | Example | Use Case |
|---|---|---|---|---|
| AND | Narrows search by requiring all specified terms to be present in the results [33] [34]. | Decreases the number of results, increasing specificity [35]. | paradigm AND syntagm [33] |
Use when you need results containing two or more specific keywords [33]. |
| OR | Broadens search by retrieving results containing any of the specified terms [33] [34]. | Increases the number of results, improving sensitivity and recall [35]. | meteor OR meteorite [33] |
Use to include synonyms, acronyms, or related concepts [33] [18]. |
| NOT (or AND NOT) | Excludes results containing a specific term or concept [33] [34]. | Decreases the number of results by removing irrelevant records. Use with caution [35]. | football NOT soccer [33] |
Use to filter out a clearly defined, unwanted concept that is likely to cause noise [33]. |
The following diagram illustrates the logical relationships created by these operators when searching a database.
Beyond the core operators, specific modifiers add layers of control and precision to search strings, which is critical for handling complex research questions in environmental science.
Table 2: Advanced Search Modifiers and Proximity Operators
| Modifier/ Operator | Function | Example | Application in Evidence Synthesis |
|---|---|---|---|
| Parentheses ( ) | Groups concepts and controls the order of search execution, a process known as "nesting" [33] [34]. | (rural OR urban) AND sociology [33] |
Ensures synonyms are grouped logically before being combined with other concepts [34]. |
| Quotation Marks " " | Searches for an exact phrase [33] [35]. | "Newtonian mechanics" [33] |
Crucial for capturing specific technical terms or multi-word concepts accurately. |
| Asterisk * | Serves as a truncation wildcard, finding variations of a root word [33] [12]. | Develop* returns develop, developer, developing, development [33]. |
Captures plural forms, different tenses, and related terms, improving search sensitivity [18]. |
| Proximity (NEAR, WITHIN) | Finds terms within a specified number of words of each other [33]. | Solar N5 energy finds "solar" and "energy" within 5 words [33]. |
Highly useful for locating concepts that are discussed in relation to each other without being a fixed phrase. |
Table 3: Research Reagent Solutions for Systematic Searching
| Tool or Component | Function | Brief Explanation & Best Practice |
|---|---|---|
| Bibliographic Databases | Primary containers for peer-reviewed literature. | Search multiple databases (e.g., Scopus, Web of Science) as no single source contains all literature [2] [18]. |
| Controlled Vocabulary | Pre-defined "subject headings" or "keywords" used by databases. | Using thesaurus terms (e.g., MeSH in MEDLINE) ensures a wide net is cast [18]. |
| Plain Text Keywords | Free-text words and phrases. | Include synonyms, acronyms, outdated terms, and alternate spellings to be comprehensive [18]. |
| Search Field Tags | Commands that restrict searching to specific metadata fields. | Tags like TI (Title) or AB (Abstract) help balance precision and sensitivity [18]. |
| Grey Literature Sources | Non-traditionally published evidence. | Includes reports, theses, and conference proceedings; crucial for reducing publication bias in environmental sciences [2] [18]. |
| Citation Management Software | Tools for managing and deduplicating results. | Essential for handling the large volume of records (10,000-100,000+) typical in systematic reviews [2]. |
| TH-Z145 | TH-Z145, MF:C16H28O7P2, MW:394.34 g/mol | Chemical Reagent |
| 1-Octanol-d2 | 1-Octanol-d2, MF:C8H18O, MW:132.24 g/mol | Chemical Reagent |
This protocol provides a step-by-step methodology for developing, testing, and executing a comprehensive search strategy for an evidence synthesis project.
OR operator.
(human excreta OR wastewater OR "sewage sludge")AND operator.
(human excreta OR wastewater) AND (recover* OR recycl* OR reus*) AND (nutrient* OR nitrogen OR phosphorus)"struvite precipitation").recover* to find recover, recovers, recovery, etc.).site:.gov), institutional repositories, and professional organization websites [2] [18] [12].The following workflow diagram summarizes this multi-stage protocol.
Systematic documentation of the search process and its outcomes is a mandatory step in evidence synthesis. The following tables provide a framework for presenting quantitative data related to the search strategy and results, ensuring transparency and reproducibility.
Table 4: Search Strategy and Yield by Database
| Database / Source | Platform / Interface | Search Date | Search Syntax (Translated) | Results Captured |
|---|---|---|---|---|
| Scopus | Elsevier | 2025-11-25 | ( TITLE-ABS-KEY ( ( human AND excreta OR wastewater ) AND ( recover* OR recycl* OR reus* ) AND ( nutrient* OR nitrogen OR phosphorus ) ) ) |
4,250 |
| Web of Science Core Collection | Clarivate | 2025-11-25 | TS=((human excreta OR wastewater) AND (recover* OR recycl* OR reus*) AND (nutrient* OR nitrogen OR phosphorus)) |
3,880 |
| Google Scholar | 2025-11-25 | "human excreta" wastewater recover recycle reuse nutrient (first 200 relevant) |
200 | |
| Organizational Website (e.g., US EPA) | site:.epa.gov |
2025-11-26 | site:.epa.gov nutrient recovery wastewater |
45 |
Table 5: Search Results and Screening Flow
| Process Stage | Number of Records | Cumulative Total | Notes / Actions Taken |
|---|---|---|---|
| Records Identified from Databases | 8,130 | 8,130 | From structured database searches. |
| Records Identified from Other Sources | 245 | 8,375 | Grey literature, citation chasing, etc. |
| Records After Duplicates Removed | 6,500 | 6,500 | Using reference management software. |
| Records Screened by Title/Abstract | 6,500 | 6,500 | 5,200 records excluded. |
| Full-Text Articles Assessed for Eligibility | 1,300 | 1,300 | 850 records excluded with reasons. |
| Studies Included in Final Synthesis | 450 | 450 |
The critical importance of a well-designed search strategy is vividly illustrated in the field of environmental evidence, where terminology can be diverse and poorly standardized. A comparison of five evidence maps on the topic of nutrient recovery from human excreta and domestic wastewater revealed a surprisingly low overlap in the studies they included [1]. Even after correcting for differences in scope, only about a tenth of the studies were common to both evidence bases derived from two major reviews [1].
This highlights the challenge of "differential search term sensitivity and specificity," where compound search terms are not equally effective across all subdomains of a research topic [1]. A search string that is highly sensitive for one technology (e.g., struvite precipitation from urine) might lack the specific terms needed to capture studies on another (e.g., vermicomposting of feces) [1]. To mitigate this, the compilation of the evidence platform Egestabaseâwhich involved screening over 150,000 recordsâemployed a strategy of additional targeted searches for individual subdomains (e.g., 'urine AND struvite precipitation', 'feces AND vermicomposting') to ensure comprehensive coverage beyond a single, compound search string [1]. This case underscores that in complex environmental domains, a single search string is often insufficient, and a modular, multi-pronged search approach is necessary for a balanced and comprehensive mapping outcome.
In the field of environmental evidence research, the effectiveness of a study is fundamentally dependent on the quality and comprehensiveness of the literature search. Multiple database search strategies are paramount to minimize bias and ensure all relevant evidence is captured. This necessitates mastery of advanced search techniques, including truncation, wildcards, and phrase searching, to construct sensitive and precise search strategies across diverse bibliographic databases. These techniques allow researchers to effectively account for linguistic variations, such as plurals, different spellings, and synonymous phrases, which are common challenges in scientific literature. This document provides detailed application notes and experimental protocols for implementing these techniques within a robust search methodology, specifically tailored for complex evidence syntheses in environmental health and related domains.
Table 1: Core Advanced Search Techniques
| Technique | Primary Function | Common Symbol(s) | Key Consideration |
|---|---|---|---|
| Truncation [36] [37] | Searches for multiple word endings from a common root. | Asterisk (*), sometimes !, ?, or # |
Can retrieve irrelevant results if the root is too short (e.g., cat* finds catapult). |
| Wildcard [36] [38] | Represents a single or multiple unknown characters within a word. | Question mark (?), asterisk (*), or hash (#) |
Useful for accounting for internal spelling variations (e.g., wom!n for woman/women). |
| Phrase Searching [36] [39] | Forces a search for an exact sequence of words. | Quotation marks (" ") |
Overly rigid; may exclude relevant studies that use the same words in a slightly different order. |
| Proximity Searching [37] [39] | Finds terms within a specified number of words of each other, in any or specified order. | Varies by database (e.g., NEAR/n, ADJn, Nn, Wn) |
Increases recall compared to phrase searching while maintaining higher precision than a simple AND. |
The implementation of advanced search techniques is not standardized and varies significantly across database platforms. The following protocols outline the specific syntax required for major databases used in scientific research.
Table 2: Database-Specific Search Operators
| Database / Platform | Truncation Symbol | Wildcard Symbols | Proximity Operator | Phrase Search |
|---|---|---|---|---|
| Ovid (MEDLINE, Embase) [37] [39] | * or $ |
$ (single character) |
ADJn (e.g., soil ADJ3 pollut*) |
" " |
| EBSCOhost (CINAHL, Academic Search) [39] | * |
? (single character), # (zero/one character) |
Nn (any order), Wn (specified order) |
" " |
| Web of Science [37] | * (any position) |
? (single character) |
NEAR/n |
" " |
| Scopus [37] [39] | * |
W/n (any order), PRE/n (specified order) |
" " |
|
| Cochrane Library [39] | * |
? (zero/one character) |
NEAR or NEAR/x |
" " or NEXT (for terms with wildcards) |
| PubMed | * |
Limited support; relies on Automatic Term Mapping. | " " |
This protocol provides a step-by-step methodology for constructing a search strategy for a systematic review or evidence synthesis.
Step 1: Concept Identification and Keyword Generation
Step 2: Apply Truncation and Wildcards
invertebrate* (to capture invertebrate, invertebrates)arthropod*worm? (to capture worm, worms)insect*Step 3: Formulate Phrase and Proximity Searches
"heavy metal", "climate change")."drug resistance" would be: (drug* N3 resist*) AND bacter*.Step 4: Combine Concepts with Boolean Operators
AND operator.OR operator.( ) to nest terms and control the order of execution [39].(concept A terms) AND (concept B terms) AND (concept C terms)Step 5: Translate and Execute Across Databases
Step 6: Peer Review of Search Strategy
The following diagram illustrates the logical workflow for building a complex search string using the advanced techniques discussed, from initial keyword generation to the final, executable query.
Table 3: Key Resources for Systematic Search Development
| Item | Function & Application in Evidence Synthesis |
|---|---|
| Boolean Operators (AND, OR, NOT) [36] | The logical foundation for combining search terms. OR broadens a search (synonyms), AND narrows it (different concepts), and NOT excludes terms (use with caution). |
| Subject Headings (MeSH, Emtree, CINAHL Headings) [37] [39] | Controlled vocabulary terms assigned by databases to describe content. Using exploded subject headings ensures comprehensive retrieval of all articles indexed under a concept and its more specific terms. |
| PRESS Checklist [37] | An evidence-based checklist used for the peer review of electronic search strategies, critical for minimizing errors and ensuring the search strategy is of high quality. |
| Yale MeSH Analyzer [37] | A web-based tool that analyzes the MeSH terms assigned to a set of known, relevant articles. This helps identify potentially missing subject headings or keywords for one's search strategy. |
| Database Syntax Guide | Reference documentation for the specific database platform (e.g., Ovid, EBSCOhost). Essential for correctly implementing truncation, wildcards, and proximity operators, as syntax varies. |
| MRTX-1257-d6 | MRTX-1257-d6, MF:C33H39N7O2, MW:571.7 g/mol |
| NSC81111 | NSC81111, CAS:1678-14-4, MF:C19H16O4, MW:308.3 g/mol |
In the context of environmental evidence research, developing a robust, multiple-database search strategy is fundamental to the integrity of any systematic review or systematic map. A comprehensive and unbiased literature search minimizes the risk of missing relevant studies, thereby protecting the review's conclusions from potential biases [15]. A critical, yet often overlooked, step in validating this process is the creation and use of a test-list of known articles, often called a "gold set" [40] or "benchmarking articles" [41]. This application note provides detailed protocols for developing and utilizing such a test-list to empirically assess the performance of search strategies within environmental evidence synthesis.
The principle is analogous to a diagnostic test in clinical practice; just as a test must correctly identify patients with a disease, a search strategy must correctly retrieve studies relevant to the review question [42]. By using a pre-defined set of known relevant articles, reviewers can move beyond theoretical checks to a quantitative evaluation of their search strategy's sensitivity (ability to retrieve all relevant items) and precision (ability to retrieve only relevant items) [43].
A test-list serves multiple crucial functions throughout the search development process, ensuring the search strategy is both comprehensive and efficient.
The following protocol outlines the steps for creating a robust and representative test-list of known articles.
The first step is to gather a pool of potentially relevant articles from diverse sources to ensure the test-list is well-rounded. A minimum of 10-20 key articles is a recommended starting point.
citationchaser to find more recent studies that cite the key articles) [18] [44].Not all candidate articles are equally suitable for the test-list. The goal is to create a final set that is both highly relevant and diverse.
Maintain a structured record of the test-list. Table 1 provides a template for documenting the test-list and its key characteristics.
Table 1: Template for Documenting the Test-List of Known Articles
| Article ID | Citation | PECO Elements Represented | Source in Test-List | Notes (e.g., type of grey literature, non-English language) |
|---|---|---|---|---|
| GL-01 | Author, A. (Year). Title... | Population: X, Exposure: Y | Expert Recommendation | Government report |
| GL-02 | Author, B. (Year). Title... | Population: Z, Exposure: Y | Scoping Search | Conference abstract |
| ... | ... | ... | ... | ... |
Once a draft search strategy has been developed for a specific database (e.g., MEDLINE via Ovid), the following experimental protocol should be used to test its performance.
Run the final, translated search strategy in the target database. Export the results into a citation manager or screening tool, ensuring the date of the search and the exact search string are recorded for reproducibility [41].
Within the screening environment, check whether each article from the test-list is present in the search results. It is critical to confirm that the article is indexed in the database being tested; an article cannot be retrieved if it is not present in the database.
Calculate the following key metrics to quantify the search strategy's performance. Table 2 provides a structure for recording these results across multiple databases.
Recall = (Number of Retrieved Test-List Articles / Number of Indexed Test-List Articles) * 100Absolute Recall = (Number of Retrieved Test-List Articles / Total Number of Test-List Articles) * 100Table 2: Template for Recording Search Performance Test Results
| Database & Platform | Total Test-List Articles | Indexed Test-List Articles | Retrieved Test-List Articles | Recall (%) | Absolute Recall (%) | Action Taken |
|---|---|---|---|---|---|---|
| Scopus | 20 | 18 | 16 | 88.9 | 80.0 | Strategy accepted |
| Web of Science | 20 | 17 | 14 | 82.4 | 70.0 | Terms "X" and "Y" added |
| Global Index Medicus | 20 | 12 | 10 | 83.3 | 50.0 | Noted lower coverage |
If the recall is unacceptably low for a database where the test-list articles are known to be indexed, the search strategy requires refinement.
The following workflow diagram illustrates the complete process of developing and using the test-list.
This section details the essential "research reagents" and tools required to implement the protocols described in this application note.
Table 3: Essential Research Reagents and Tools for Search Assessment
| Item/Tool | Function/Description | Example/Reference |
|---|---|---|
| Benchmarking Articles | A pre-identified set of known relevant articles that form the test-list against which search performance is measured. | [41] |
| Citation Management Software | Software for storing, organizing, and deduplicating bibliographic records exported from database searches. Essential for managing the test-list and search results. | EndNote, Zotero, Mendeley [2] |
| Systematic Review Management Platform | Web-based platforms that facilitate the screening process, allowing teams to efficiently check for the presence of test-list articles within large result sets. | Covidence, Rayyan [41] |
| Citation Chasing Tools | Tools that automate the process of forward and backward citation chasing to help identify candidate articles for the test-list. | citationchaser [18] |
| Bibliographic Databases | Disciplinary and multidisciplinary databases that are searched. The choice of databases should be justified by the topic of the review. | Scopus, Web of Science, MEDLINE, Embase, Global Index Medicus [41] [15] |
| Grey Literature Sources | Repositories for non-commercially published material, crucial for reducing publication bias and for finding relevant reports for the test-list. | Government websites, institutional repositories, clinical trials registries [2] [44] |
| Reporting Guidelines | Checklists to ensure the search process, including the use of a test-list, is fully and transparently reported. | PRISMA-S [41] |
| SIAIS164018 | SIAIS164018, MF:C43H48ClN10O7P, MW:883.3 g/mol | Chemical Reagent |
| TSC25 | TSC25, MF:C14H18Cl2N4OS, MW:361.3 g/mol | Chemical Reagent |
Systematic evidence synthesis in environmental research requires comprehensive search strategies that transcend single databases and languages to minimize bias and ensure global relevance. Effective planning for multiple languages and grey literature sources is fundamental to constructing a valid evidence base, particularly in environmental sciences where relevant data is often distributed across non-traditional publication channels and multilingual sources. A well-structured protocol mitigates the risk of overlooking significant evidence by systematically addressing database selection, language barriers, and grey literature integration.
The core principle of this protocol is systematic transparency, ensuring every search step is documented, reproducible, and justifiable [18]. This involves a strategic balance between sensitivity (retrieving all potentially relevant records) and precision (retrieving a high proportion of relevant records) [2]. Furthermore, the protocol acknowledges the resource-intensive nature of comprehensive searching and provides guidance for prioritizing resources when a full systematic review is not viable [1].
Searching multiple databases is critical because no single database provides comprehensive coverage of the literature. A metaresearch study confirmed that searching two or more databases significantly decreases the risk of missing relevant studies [45]. The selection of databases should be informed by their specific scope and the research topic.
Table 1: Performance Metrics of Key Bibliographic Databases in Systematic Reviews
| Database | Median Recall (%) | Unique Contribution of Included References (n) | Key Strengths and Subject Focus |
|---|---|---|---|
| Embase | 82.1 | 132 | Biomedical and pharmacological literature; strong European coverage [4]. |
| MEDLINE/PubMed | 73.6 | 63 | Life sciences and biomedicine; includes "ahead of print" publisher content [4]. |
| Web of Science Core Collection | 86.5 | 102 | Multidisciplinary science, social sciences, and arts & humanities; allows cited reference searching [4] [46]. |
| Scopus | Information Missing | Information Missing | Multidisciplinary; includes conference proceedings and cited reference searching [46]. |
| Google Scholar | Information Missing | 109 | Broad coverage including grey literature; requires structured screening of top results [4]. |
| Global Index Medicus | Information Missing | Information Missing | Biomedical and public health literature from low- and middle-income countries [46]. |
Data adapted from a prospective exploratory study of 58 systematic reviews [4].
Research indicates that a combination of Embase, MEDLINE, Web of Science Core Collection, and Google Scholar achieves a recall of 98.3% and 100% recall in 72% of systematic reviews, establishing this as a minimum baseline for comprehensive searching [4]. For multidisciplinary environmental topics, supplementary databases like Scopus and subject-specific databases (e.g., Avery Index to Architectural Periodicals for built environment topics) should be considered [46].
This protocol provides a detailed, sequential methodology for executing a comprehensive, multilingual search that incorporates grey literature.
site:.gov "search terms") and dedicated resources like DiscoverGov for U.S. government literature and Policy Commons for global think tank reports [46].
Systematic Search Workflow for Evidence Synthesis
Table 2: Research Reagent Solutions for Comprehensive Evidence Searching
| Resource Name | Type | Primary Function | Access |
|---|---|---|---|
| Embase | Bibliographic Database | Comprehensive biomedical and pharmacological literature coverage; crucial for minimizing missed studies [4]. | Subscription |
| Web of Science Core Collection | Bibliographic Database | Multidisciplinary coverage with powerful cited reference searching capabilities [4] [46]. | Subscription |
| Global Index Medicus | Bibliographic Database | Provides access to literature from low- and middle-income countries, addressing language and geographic biases [46]. | Free |
| CABI: CAB Abstracts | Bibliographic Database | Focuses on applied life sciences, including agriculture, environment, and public health. Essential for environmental topics. | Subscription |
| Grey Matters | Grey Literature Tool | A practical checklist and source guide for systematic searching of health-related grey literature [46]. | Free |
| Policy Commons | Grey Literature Repository | Search engine for policy reports, working papers, and publications from think tanks, IGOs, and NGOs globally [46]. | Free/Registration |
| Citationchaser | Software Tool | Facilitates efficient forward and backward citation chasing in systematic reviews [18]. | Free (R package/web tool) |
| CADIMA | Systematic Review Tool | An open-access tool supporting the entire systematic review process, including search planning and documentation. | Free |
| EndNote / Zotero | Citation Manager | Manages, deduplicates, and organizes large volumes of bibliographic data from multiple database searches [2]. | Subscription / Freemium |
| Rayyan / Covidence | Screening Tool | Web-based tools that facilitate collaborative title/abstract and full-text screening among review team members. | Freemium / Subscription |
In the context of environmental evidence research, systematic reviews and evidence syntheses are fundamental for integrating knowledge and informing policy [47]. A central challenge in this process is the development and execution of effective search strategies across multiple bibliographic databases. The objective is to balance search sensitivity (recall), the ability to capture all relevant records, with search precision, the ability to exclude irrelevant records [48]. Searches with high sensitivity tend to have low precision, resulting in an unmanageably large volume of results for screening. Conversely, highly precise searches risk missing critical evidence, potentially biasing the review's findings [48]. This Application Note provides detailed protocols and tools for researchers to systematically manage this trade-off, ensuring their literature searches in environmental studies are both comprehensive and efficient.
The following table summarizes typical performance characteristics and outcomes for search strategies with different balances of recall and precision, based on reported practices in evidence synthesis [48].
Table 1: Characteristics and Outcomes of Search Strategy Approaches
| Search Strategy Approach | Estimated Relative Recall (%) | Estimated Precision (%) | Typical Outcome for a Systematic Review | Primary Risk |
|---|---|---|---|---|
| High-Sensitivity Search | ~90-100 | ~1-5 | Very large volume of records to screen (e.g., 10,000+); high workload. | Low feasibility; reviewer fatigue. |
| Balanced Search | ~80-90 | ~5-15 | Manageable volume of records (e.g., 2,000-5,000); sustainable workload. | Potential to miss some relevant studies. |
| High-Precision Search | ~50-80 | ~15-30 | Low volume of records to screen (e.g., <1,000); fast screening process. | High probability of missing relevant evidence; introduction of bias. |
A survey of recent systematic reviews indicates that the evaluation of search string performance is rarely reported, underscoring the need for more rigorous and transparent methodologies [48]. Furthermore, the adoption of machine learning (ML) tools to assist with screening remains limited, with only about 5% of studies explicitly reporting their use; when applied, ML is primarily focused on the screening phase to manage large result volumes [47].
This protocol adapts the PSALSAR method (Protocol, Search, Appraisal, Synthesis, Analysis, Reporting) for systematic literature reviews in environmental science, with a specific focus on the search and appraisal stages [49].
OR.AND.TITLE-ABS-KEY in Scopus) and truncation/wildcards as appropriate for each database.(terma OR termb OR termc) AND (termx OR termy OR termz) [50] [48].(Number of benchmark records found / Total number of benchmark records) * 100 [48].This workflow for the search and appraisal stages can be visualized as follows:
This protocol provides a detailed, objective method for estimating the sensitivity of a search string, a process identified as critical yet underutilized [48].
Table 2: Essential Research Toolkit for Search Strategy Evaluation
| Tool / Resource | Type | Primary Function in Protocol |
|---|---|---|
| Benchmark Publication Set | Research Material | Serves as the known-relevant "gold standard" for objective performance testing [48]. |
| Power Thesaurus | Online Tool | Assists in identifying synonyms and related terms to improve search term coverage [50]. |
| Bibliographic Databases (Scopus, Web of Science, etc.) | Platform | Hosts academic literature and provides interfaces for executing and testing search strings [50] [48]. |
| Reference Manager (Mendeley, Zotero) | Software | Manages search results, removes duplicate records, and stores the benchmark set [50]. |
| Relative Recall Calculator (Spreadsheet) | Analytical Tool | Calculates the sensitivity metric (Relative Recall %) for the evaluated search string [48]. |
The logical relationship and workflow for this objective evaluation is shown below:
The methodologies described above are particularly pertinent for environmental evidence research, where data is often extensive, heterogeneous, and sourced from diverse disciplines [50]. Applying a structured framework like PSALSAR ensures a reproducible and transparent process [49]. Furthermore, the integration of the FAIR principles (Findable, Accessible, Interoperable, Reusable) and a focus on data life cycle management into the research data management plan are emerging as critical themes for enhancing the value and impact of environmental syntheses [50]. By adopting these rigorous protocols for search strategy development and validation, researchers in environmental science, drug development, and public health can strengthen the reliability and comprehensiveness of their evidence syntheses, thereby providing a more robust foundation for decision-making.
In the rigorous field of environmental evidence research, the integrity of a systematic review or meta-analysis is fundamentally dependent on the quality of the literature search. A comprehensive, transparent, and reproducible search strategy forms the bedrock of a reliable evidence base. However, this process is susceptible to specific, common errors in syntax and spelling that can systematically bias results, leading to incomplete or flawed conclusions. Within the context of a broader thesis on multiple database search strategies, this article details these frequent pitfalls, provides protocols for their identification and correction, and offers practical tools to enhance search quality for researchers, scientists, and drug development professionals.
Errors in electronic search strategies are not merely clerical; they have a direct and significant impact on the recall and precision of a literature search. Recall (or sensitivity) refers to the proportion of relevant studies successfully retrieved, while precision refers to the proportion of retrieved studies that are relevant. Syntax and spelling errors predominantly reduce recall, meaning relevant studies are missed, potentially introducing bias and undermining the validity of the entire synthesis [51].
Evidence from assessments of systematic reviews highlights the prevalence of this issue. An evaluation of reviews from the Cochrane Database of Systematic Reviews (CDSR) found that among the search strategies that could be assessed, a striking 91% contained at least one error [51]. These errors can distort the perceived utility of bibliographic databases and may inflate the importance of less systematic search methods [51].
Table 1: Frequency and Impact of Common Search Errors
| Error Type | Example | Potential Consequence | Reported Frequency in Assessable Cochrane Reviews |
|---|---|---|---|
| Spelling & Typographical Errors | Searching for elipseSize instead of ellipseSize [52] |
Failure to retrieve relevant records containing the correct spelling. | Common, though specific frequency not isolated [51]. |
| Boolean Operator Misuse | Incorrect nesting of terms using AND/OR [51] |
Retrieves an illogical set of records, either too broad or too narrow. | Among the most common errors identified [51]. |
| Insufficient Search Reporting | Failing to report the full search strategy for replication [51] | Makes the search irreproducible and the review's validity unverifiable. | 63% of reviews had strategies that could not be assessed [51]. |
Boolean operators (AND, OR, NOT) and parentheses are the fundamental syntax for constructing database queries. Misuse can completely alter the meaning of a search.
1. Objective: To systematically verify the logical structure of a search string and ensure it accurately represents the research question's concepts.
2. Materials:
3. Methodology:
* Deconstruct the PICO/S: Break down your research question (e.g., Population, Intervention, Comparator, Outcome for health; Subject, Phenomenon of Interest, Context for environment) into discrete concepts [51] [53].
* Map Concepts to Search Syntax: For each concept, list all synonymous text words and controlled vocabulary terms (e.g., MeSH, Emtree) combined with the OR operator. This creates a conceptual "block" [18].
* Combine Conceptual Blocks: Join these conceptual blocks with the AND operator to ensure the search results must contain at least one term from each block [51].
* Validate Nesting with Parentheses: Use parentheses to group terms unambiguously. The protocol should involve checking that every opening parenthesis has a corresponding closing parenthesis and that the logical order of operations is correct. For example: (salmon OR trout) AND (population decline OR abundance).
* Peer Review: Employ the Peer Review of Electronic Search Strategies (PRESS) checklist [54]. Have an information specialist or experienced colleague review the entire strategy for logical errors and completeness.
The following diagram outlines a robust workflow for developing and validating a search strategy, incorporating checks for both syntax and spelling errors.
Misspelled identifiers and keywords are valid lexemes to a database's lexical analyzer and will not trigger an error message [52]. Instead, they silently fail to retrieve relevant records.
1. Objective: To minimize the risk of missing relevant studies due to spelling variations, typos, or terminological errors.
2. Materials:
3. Methodology:
* Pre-Search Term Validation: Use online thesauri (e.g., Power Thesaurus) and subject heading databases (MeSH for MEDLINE) to identify all variant spellings and synonyms for each key concept during the initial search development [50]. Actively consider:
* British vs. American English: e.g., behaviour vs. behavior.
* Plurals and Word Endings: Use database wildcards (e.g., forest* to find forest, forestry, forests) [53].
* Common Misspellings: Manually check for typos in your search strings.
* Benchmark Testing ("Gold Standard" Validation): Compile a list of 5-10 key articles that are known to be relevant to your review topic. Run your final search strategy and confirm that it retrieves these benchmark articles. A failure to retrieve one or more articles indicates a problem with the search terminology, which may be due to spelling or synonym coverage [18].
* Iterative Search and Screening: Be prepared to refine your search terms based on the language and terminology encountered in the titles and abstracts of articles retrieved during preliminary searches. If you see a relevant synonym you missed, incorporate it.
Table 2: Research Reagent Solutions for Robust Searching
| Tool / Reagent | Function in the Search Process | Example / Application |
|---|---|---|
| Boolean Operators (AND, OR, NOT) | Combines search terms logically to broaden or narrow results [51] [55]. | (conservation OR preservation) AND (biodiversity OR "species richness") |
| Controlled Vocabulary (Thesauri) | Uses a database's standardized subject headings to tag content, ensuring comprehensive retrieval regardless of the author's chosen wording [18]. | Using MeSH term "Environmental Monitoring" in MEDLINE instead of text words like "environmental assessment" or "ecosystem tracking". |
| Wildcards and Truncation | Accounts for variations in spelling, word endings, and plurals [53]. | forest* finds forest, forestry, forests. col?r finds color and colour. |
| Search Strategy Template | A pre-formatted document to track and document search strategies across multiple databases, ensuring transparency and reproducibility [18]. | Recording the database, platform, date searched, and full search string for every database used. |
| Reference Management Software | Assembles a library of search results, combines results from multiple databases, and removes duplicate records [53] [50]. | Using Mendeley, Zotero, or EndNote to manage thousands of citations from Scopus, Web of Science, etc. |
Environmental evidence research requires searching multiple databases (e.g., Scopus, Web of Science, specialist indexes) to capture the interdisciplinary literature [18] [50]. Each database has unique search syntax and controlled vocabularies, multiplying the risk of errors.
The following diagram illustrates the process of translating and executing a search across multiple databases while maintaining consistency and accuracy.
Key Considerations for Multiple Databases:
In the context of complex, multi-database search strategies for environmental evidence, vigilance against syntax and spelling errors is not a minor detail but a core methodological imperative. By adopting structured protocols for Boolean logic verification and comprehensive spelling checks, researchers can significantly enhance the recall and precision of their searches. Integrating tools such as controlled vocabularies, search templates, and benchmark testing, along with rigorous peer review, creates a robust defense against the common errors that compromise systematic reviews. Ultimately, a meticulously constructed and documented search strategy is the first and most critical step in ensuring the reliability and authority of the synthesized evidence.
Search Limits: Pre-indexed database features that instantly restrict results by specific criteria (e.g., publication year, language) through interface controls [56]. These rely on database indexing which may be incomplete or inconsistent, particularly for newly added records.
Search Filters (Hedges): Validated search strings designed to retrieve specific study types or categories (e.g., randomized controlled trials, human studies) [56]. Unlike limits, filters are transparent, reproducible strings that can be peer-reviewed and cited.
Sensitivity: The ability of a search to identify all relevant records within a source, calculated as the proportion of relevant records successfully retrieved [57].
Precision: The proportion of retrieved records that are relevant to the research question [2].
Table 1: Comparative analysis of limitation approaches across evidence syntheses in environmental research
| Evidence Base | Date Restrictions | Language Restrictions | Source Type Considerations | Reported Impact on Results |
|---|---|---|---|---|
| SA Review [1] | Not specified | Not specified | Focus on distinct recovery options rather than all evidence | Covered considerably fewer studies than less restricted evidence bases |
| UM Evidence Base [1] | 2013-2017 period analyzed | Not specified | Covered only human urine versus broader wastewater fractions | Limited scope to specific nutrient source |
| EW/EB Evidence Platforms [1] | Comprehensive search with date documentation | Not explicitly restricted | Covered domestic/municipal wastewater broadly including multiple fractions | Identified substantially more studies than restricted searches |
| CEEDER Database [14] | Continuously updated | Not specified | Includes both commercially published journals and grey literature | Provides comprehensive evidence overview across environmental sector |
Table 2: Risk assessment of common limitation types in environmental evidence synthesis
| Limitation Type | Potential Benefits | Methodological Risks | Recommended Mitigation Strategies |
|---|---|---|---|
| Publication Date | Focus on current evidence; Manageable result sets | Missing foundational studies; Temporal bias | Document rationale; Search backwards until saturation; Consider key historical periods |
| Language | Reduced translation costs; Focus on major research languages | Geographic bias; Exclusion of regionally important evidence | Provide clear justification; Consider regional languages relevant to topic |
| Source Type | Increased efficiency; Focus on peer-reviewed literature | Publication bias; Exclusion of grey literature critical to environmental topics | Use comprehensive grey literature search protocols [2]; Document sources |
Purpose: To establish a transparent methodology for applying temporal boundaries while minimizing the risk of excluding historically important evidence.
Materials: Bibliographic databases (e.g., Web of Science, Scopus), reference management software, protocol documentation template.
Procedure:
Validation: Test retrieval of benchmark studies; Report number of pre-restriction era studies identified through citation chasing.
Purpose: To establish ethically and methodologically defensible language boundaries while acknowledging potential geographic biases.
Materials: Translation resources, multilingual team members when possible, regional database access.
Procedure:
Validation: Report results of pilot analysis; Document number of non-inclusion language records identified through supplementary methods.
Purpose: To balance comprehensive evidence collection with practical resource constraints through strategic source selection.
Materials: Multiple bibliographic databases, grey literature sources, specialized repositories.
Procedure:
Validation: Test retrieval of known relevant studies across source types; Report grey literature yield percentage.
Decision Framework for Applying Search Limitations
Table 3: Essential research reagents and solutions for implementing search limitations
| Tool/Resource | Function | Application Notes |
|---|---|---|
| Benchmark Study Set | Validation of search strategy sensitivity | Curate 3-5 known relevant studies; Test retrieval with applied limits [57] |
| ROSES Reporting Template | Standardized methodology reporting | Ensure transparent reporting of limitations and justification [53] |
| Citation Chasing Tools | Identification of seminal works outside restrictions | Forward/backward citation chasing on key studies [18] |
| Grey Literature Search Template | Systematic capture of non-peer-reviewed evidence | Structured approach to organizational website searching [2] |
| CEESAT Appraisal Tool | Quality assessment of evidence reviews | Evaluate reliability of included syntheses [14] |
| DMX-129 | DMX-129, MF:C19H17FN8, MW:376.4 g/mol | Chemical Reagent |
| Ganoderic acid Mk | Ganoderic acid Mk, MF:C34H50O7, MW:570.8 g/mol | Chemical Reagent |
The strategic implementation of search limitations requires careful consideration of both methodological integrity and practical constraints within environmental evidence synthesis. Date restrictions should be justified based on technological or policy relevance periods rather than arbitrary cutoffs. Language limitations must acknowledge and mitigate potential geographic biases, particularly for regionally specific environmental topics. Source type filtering should preserve access to critical grey literature while focusing database searching on most productive sources. By employing the protocols and decision framework outlined herein, researchers can implement defensible limitations while maintaining the comprehensive character essential to valid evidence synthesis. All limitation decisions must be transparently documented in protocols and final publications to enable critical appraisal and reproducibility.
In the realm of environmental evidence research, the completeness of a literature search directly determines the validity and reliability of its findings. Systematic searches require looking beyond a single database and a simple set of keywords to ensure all relevant evidence is captured [4]. This document provides detailed Application Notes and Protocols for expanding search strategies through the systematic identification and use of synonyms and related terms. The guidance is framed within the context of multiple database search strategies, a critical component of rigorous systematic reviews and other evidence synthesis methodologies in environmental science and drug development. Failure to adequately expand searches can lead to significant gaps in the evidence base; one study found that approximately 16% of relevant references in systematic reviews were uniquely found in only a single database [4]. This protocol outlines a structured methodology to mitigate this risk.
Searching multiple databases is not merely a recommendation but a necessity for robust evidence synthesis. A prospective exploratory study demonstrated that no single database retrieves all relevant references, and the combination of Embase, MEDLINE, Web of Science Core Collection, and Google Scholar was required to achieve 98.3% recall across a large sample of systematic reviews [4]. This is because databases have varying coverage, scope, and indexing practices. Furthermore, different databases employ different controlled vocabulariesâstructured, hierarchical lists of subject-specific termsâmeaning the same concept can be described with different terminology across platforms [5] [58].
A core challenge in information retrieval is the fact that the same concept can be described in multiple ways. Authors may use different words (synonyms), related terms, or broader/narrower terms to describe the same idea. A thesaurus, as defined by the USGS, is a "consistent collection of terms chosen for specific purposes with explicitly stated, logical constraints on their intended meanings and relationships" [58]. Leveraging these tools is fundamental to an effective search.
The relationships within a thesaurus provide the logical framework for expanding a search systematically:
This section provides a detailed, step-by-step methodology for developing a comprehensive search strategy that fully leverages synonyms and related terms across multiple databases.
Objective: To generate an exhaustive list of free-text keywords and identify relevant controlled vocabulary terms for each key concept in a research question.
Materials: Access to major relevant databases (e.g., PubMed/MEDLINE, Embase, Web of Science), a thesaurus or controlled vocabulary for at least one database, and a spreadsheet or word processor for documentation.
Workflow:
Table 1: Search Term Development Worksheet for the Concept "Microplastics"
| Key Concept | Free-Text Synonyms & Variants | Controlled Vocabulary (e.g., MeSH) | Broader Terms | Narrower Terms |
|---|---|---|---|---|
| Microplastics | "micro plastic", "micro-plastic", "plastic debris", "plastic particle", "synthetic polymer", "nurdle", "microfiber" | Microplastics (Scope Note: Synthetic polymers...)Non-Preferred Terms: "micro plastic", "plastic particulate" |
PlasticsWater Pollutants, Chemical |
Nanoplastics |
Diagram 1: Workflow for Systematic Term Identification
Objective: To adapt and run the developed search strategy efficiently and effectively across different database platforms, accounting for variations in syntax and vocabulary.
Materials: The completed Search Term Worksheet from Protocol 1, access to target databases, and documentation software.
Workflow:
OR.(microplastic* OR "plastic debris" OR microfiber*).* or $) to capture word variants (e.g., plastic* finds plastic, plastics) and quotation marks for phrase searching [5].[tiab] in PubMed vs .mp. in Ovid platforms [5].NEAR/3 in some platforms), which can be more precise than phrase searching [5].Table 2: Database Syntax Adaptation Guide
| Search Element | PubMed Syntax | Ovid (MEDLINE/Embase) Syntax | Web of Science Syntax | EBSCOhost Syntax |
|---|---|---|---|---|
| Title/Abstract | [tiab] |
.mp. or .ti,ab. |
TS= |
TI AB |
| Medical Subject Headings | [mh] |
/ (e.g., Microplastics/) |
N/A | MH |
| Truncation | * (e.g., plastic*) |
* or $ |
* |
* |
| Phrase Search | "plastic debris" |
"plastic debris" |
"plastic debris" |
"plastic debris" |
| Proximity Search | "plastic debris"[tiab]~5 |
(plastic ADJ3 debris).mp. |
plastic NEAR/3 debris |
plastic N3 debris |
Objective: To assess the comprehensiveness and efficiency of the executed searches and determine if expansion or refinement is necessary.
Materials: The list of pre-identified gold-standard articles, the combined results from all database searches, and a reference manager.
Workflow:
Diagram 2: Search Strategy Validation Workflow
Table 3: Essential Resources for Comprehensive Literature Searching
| Tool / Resource | Function / Application | Example / Notes |
|---|---|---|
| Bibliographic Databases | Host indexed scholarly literature; the primary source for systematic search results. | Embase: Strong coverage of pharmacology/environmental science.MEDLINE/PubMed: Life sciences and biomedicine.Web of Science Core Collection: Multidisciplinary science.Scopus: Large multidisciplinary abstract database. |
| Controlled Vocabularies | Provide standardized terminology to search concepts consistently, overcoming the challenge of synonyms. | MeSH (Medical Subject Headings): Used by NLM in MEDLINE [5].Emtree: Used in the Embase database [5].USGS Thesaurus: For geological and environmental science topics [58]. |
| Search Syntax Macros & Tools | Aid in translating complex search strategies between different database interfaces, saving time and reducing errors. | Custom macros (e.g., in Excel or text editors) or dedicated software can assist in converting field codes and operators [4]. |
| Validated Search Filters | Pre-tested search strings designed to identify specific study designs (e.g., randomized trials, observational studies). | Cochrane Collaboration's highly sensitive RCT filter for PubMed [5]. Use with caution for observational studies in environmental health. |
| Reference Management Software | Manages search results, removes duplicate records, and facilitates the screening process. | EndNote, Zotero, Rayyan. Critical for handling the large volume of records from multiple databases [4]. |
| PRESS Checklist | A standardized guideline for the peer review of electronic search strategies to improve quality and completeness. | Ensures search strategies are well-translated, free of errors, and use appropriate terms and logic before execution [5]. |
Within the rigorous domain of environmental evidence research, the development of a robust multiple-database search strategy is a foundational component of any systematic review or map. An iterative process for search development, characterized by repeated cycles of testing, evaluation, and refinement, is critical to minimizing bias and ensuring the comprehensive identification of relevant literature [59]. Failing to include relevant information can significantly affect and potentially skew the findings of a synthesis [15]. This protocol details the application of an iterative, tested, and peer-reviewed methodology for constructing search strategies that are both transparent and fit-for-purpose within the context of environmental evidence synthesis.
The development of a final search strategy is not a linear but a cyclical process. It requires conscious planning, execution, evaluation, and refinement to build and improve the strategy step-by-step [59]. The following workflow outlines the key stages, emphasizing that the process may loop back on itself until a satisfactory level of performance is achieved.
Diagram 1: The iterative search development workflow. This cycle continues until the search strategy demonstrates acceptable performance against the test list, after which it undergoes formal peer review before final execution and documentation.
An independently developed test list is crucial for objectively assessing search strategy performance [16].
3.1.1 Methodology:
3.1.2 Key Quantitative Benchmarks: Table 1: Performance metrics for search strategy evaluation.
| Metric | Calculation | Target Benchmark | Purpose |
|---|---|---|---|
| Sensitivity | (Number of test list articles retrieved / Total number of test list articles) x 100 | â¥90% | Measures the comprehensiveness of the search in retrieving known relevant studies [16]. |
| Precision | (Number of relevant studies retrieved / Total number of studies retrieved) x 100 | Varies by topic | Measures the efficiency of the search; a higher percentage reduces screening workload. |
A formal peer review process for search strategies is a critical step to identify errors and enhance quality [60].
3.2.1 Methodology:
3.2.2 PRESS Instrument Components: Table 2: Key elements assessed during the peer review of a search strategy.
| Component | Description | Example Review Questions |
|---|---|---|
| Boolean Operators | Checks the logical structure of the search string. | Are the AND/OR/NOT operators used correctly? Is the logic sound? [15] |
| Spelling & Syntax | Identifies typographical errors and database-specific command errors. | Are all terms spelled correctly? Are field codes (e.g., .ti,ab.) used appropriately? [16] |
| Term Selection | Evaluates the choice and comprehensiveness of keywords and subject headings. | Are key synonyms, related terms, and variant spellings included? [18] |
| Line-by-Line Review | Assesses each individual line of the search strategy for errors and omissions. | Does each segment of the search produce the expected results? |
| Translation | Ensures the strategy is correctly adapted for different databases. | Have subject headings been properly translated for each database? [18] |
Table 3: Key resources and tools for developing and executing a systematic search strategy.
| Tool / Resource | Category | Function & Application |
|---|---|---|
| Bibliographic Databases | Information Source | Primary sources of published literature. A minimum of 3-5 multidisciplinary and subject-specific databases should be searched. |
| Test List of Articles | Validation Tool | A set of known relevant articles used to objectively test the performance (sensitivity) of the draft search strategy [16]. |
| PRESS Instrument | Quality Assurance | A structured checklist used by peer reviewers to evaluate the completeness, syntax, and logic of a search strategy [60]. |
| Boolean Operators | Search Logic | The operators AND, OR, and NOT are used to combine search terms logically, defining the relationships between concepts. |
| Controlled Vocabulary | Terminology | Database-specific subject headings (e.g., MeSH in MEDLINE) that tag articles with standardized terms, improving search comprehensiveness [18]. |
| WebAIM Contrast Checker | Accessibility Tool | A free online tool to test color contrast ratios, ensuring visualizations and outputs meet accessibility guidelines. |
Adopting a structured, iterative approach to search strategy development is non-negotiable for producing high-quality, reliable environmental evidence syntheses. The integration of objective testing with an independent test list and formal peer review using the PRESS instrument provides a robust methodological framework that significantly reduces errors and biases [16] [60]. This protocol ensures that the resulting multiple-database search strategy is comprehensive, transparent, and reproducible, thereby solidifying the integrity of the entire systematic review or map.
In the rigorous world of evidence-based research, particularly in environmental evidence and drug development, the completeness of a literature search directly determines the validity and reliability of its conclusions. While bibliographic databases like MEDLINE and Embase form the cornerstone of systematic investigation, a growing body of metaresearch demonstrates that relying solely on these platforms risks missing substantial relevant evidence. Supplementary search methodsâdefined as non-database search techniquesâprovide a crucial mechanism for identifying studies and study reports that might be overlooked by bibliographic database searching alone [44]. These methods operate on different principles than keyword-based database queries, instead leveraging citation networks, expert knowledge, and specialized repositories to uncover the full spectrum of available evidence.
The imperative for comprehensive searching is particularly acute in environmental research and pharmacovigilance, where regulatory decisions and safety profiles depend on complete evidence synthesis. In Europe, for instance, pharmacovigilance regulations mandate risk-management plans and postauthorization safety studies for medicines, necessitating robust methodologies for evidence identification [61]. This application note establishes why supplementary searches are indispensable in contemporary evidence synthesis and provides detailed protocols for their implementation within a broader thesis on multiple database search strategies for environmental evidence research.
Empirical research substantiates the critical value of moving beyond single-database searching. A prospective exploratory study examining 58 published systematic reviews found that searching multiple sources is essential for adequate literature coverage [4]. The research revealed that 16% of included references across these reviews (291 articles) were found in only a single database, demonstrating the unique contributions of individual sources [4]. Embase produced the most unique references (n=132), followed by other database and supplementary sources.
Perhaps more strikingly, an analysis of database combinations found that even the most comprehensive database searching alone may be insufficient. The combination of Embase, MEDLINE, Web of Science Core Collection, and Google Scholar achieved an overall recall of 98.3%, reaching 100% recall in 72% of systematic reviews [4]. This indicates that despite exhaustive database searching, supplementary methods were still necessary to identify all relevant references in more than a quarter of all reviews. Researchers estimate that approximately 60% of published systematic reviews fail to retrieve 95% of all available relevant references because they do not search appropriate databases or employ supplementary techniques [4].
Table 1: Performance Metrics of Database Combinations in Systematic Reviews
| Database Combination | Overall Recall (%) | Reviews with 100% Recall (%) | Key Findings |
|---|---|---|---|
| Embase + MEDLINE + Web of Science + Google Scholar | 98.3% | 72% | Most effective combination tested |
| Single Database (best performing) | Varies | <40% | Inadequate for comprehensive recall |
| Current Practice in Published Reviews | <95% | Not reported | 60% of reviews miss >5% of relevant evidence |
The resource implications of supplementary searching are not trivial, but must be weighed against the cost of missing critical evidence. Time requirements vary significantly by method, from relatively efficient citation searching to more labor-intensive handsearching [44]. When designing search strategies for environmental evidence reviews, researchers should consider both the effectiveness and resource requirements of these supplementary methods to optimize their search workflow.
Supplementary search methods encompass a diverse set of approaches that operate on different mechanisms than traditional database searching. Where bibliographic databases rely on searching controlled indexing and free-text fields, supplementary methods locate studies through alternative pathways such as citation linkages, direct researcher communication, and specialized repositories [44].
Citation searching leverages reference lists and citation networks to identify related studies. This method is divided into:
Tools such as Citationchaser can assist with semi-automating this process, though manual review remains essential for accuracy [44]. The power of citation searching lies in its ability to overcome limitations of indexing, terminology, and database coverage by tracing intellectual connections between research works.
Direct communication with subject experts and study authors can uncover unpublished data, ongoing studies, additional reports, or clarifications about published work [44] [62]. For complex reviews, explaining your evidence needs to a topic expert may prove more efficient than formulating complex search strategies across multiple databases [44]. This approach is particularly valuable for identifying grey literature and study protocols that haven't been indexed in major databases.
Handsearching involves manually reviewing specific journals, conference proceedings, or other relevant sources page-by-page to identify studies that may not be properly indexed in databases [44] [62]. While time-consuming, this method can capture research presented in non-standard formats, early-stage publications, or content from specialized sources that lack comprehensive database indexing.
For environmental evidence and drug development research, regulatory agency sources and clinical trials registries provide access to crucial unpublished or incomplete trial data [44] [62]. These include:
These sources help counter publication bias by identifying studies regardless of their outcome or publication status.
Specialized web searching targets organizational websites, institutional repositories, and grey literature databases to identify:
The Centre for Reviews and Dissemination (CRD) Handbook distinguishes between general internet searching via search engines and targeted searching of specific relevant websites, recommending the latter as more practical for systematic reviews [44].
Table 2: Supplementary Search Methods and Their Applications
| Method | Primary Mechanism | Best For Identifying | Resource Intensity |
|---|---|---|---|
| Citation Searching | Following citation networks | Seminal works, related research clusters | Moderate |
| Contacting Authors | Direct expert communication | Unpublished data, ongoing studies, clarifications | Low-Moderate |
| Handsearching | Manual journal/conference review | Non-indexed studies, specialized sources | High |
| Regulatory Sources | Accessing agency repositories | Unpublished trial data, regulatory reports | Moderate |
| Web Searching | Targeted website searching | Grey literature, organizational reports | Variable |
Objective: To identify relevant studies through systematic exploration of citation networks.
Materials:
Methodology:
Quality Control: Dual screening of at least 20% of references with calculation of inter-rater reliability.
Objective: To identify unpublished or non-commercially published research relevant to the review question.
Materials:
Methodology:
Quality Control: Maintain search logs detailing dates, sources, strategies, and results.
Objective: To identify ongoing, completed, or unpublished clinical trials.
Materials:
Methodology:
Quality Control: Dual independent searching of key registries with comparison of results.
Implementing supplementary searches requires strategic integration with conventional database searching. The following workflow illustrates how these methods complement each other in a comprehensive search strategy:
Diagram 1: Supplementary Search Integration Workflow
This integration framework emphasizes the parallel nature of database and supplementary searching, with convergence at the reference management stage. Environmental evidence researchers should note that the specific supplementary methods emphasized may vary based on the research question, with ecological studies potentially prioritizing organizational grey literature while clinical questions may emphasize trial registries.
Table 3: Research Reagent Solutions for Supplementary Searching
| Tool Category | Specific Resources | Primary Function | Application Notes |
|---|---|---|---|
| Citation Tracking | Google Scholar, Web of Science, Scopus, Citationchaser | Identify citing and cited references | Google Scholar provides free access; subscription databases may offer more precise filtering |
| Grey Literature Sources | OpenGrey, GreyLit.org, OATD, WorldCat Theses | Locate unpublished reports, theses, conference papers | OATD specializes in open access theses; WorldCat provides comprehensive coverage |
| Trial Registries | ClinicalTrials.gov, WHO ICTRP, EU CTR | Identify registered trials regardless of publication | Essential for assessing publication bias in clinical research |
| Expert Networking | ResearchGate, institutional directories, professional associations | Facilitate contact with subject experts | Professional conferences also provide networking opportunities |
| Handsearching Aids | Journal tables of contents, conference programs | Identify relevant content in targeted sources | Most efficient when focused on high-yield sources |
| Reference Management | EndNote, Zotero, Mendeley | Organize, deduplicate, and track sources | Critical for managing results from multiple search methods |
Supplementary search methods represent not merely optional enhancements but essential components of rigorous evidence synthesis, particularly in environmental research and drug development where regulatory and policy decisions depend on complete evidence bases. The quantitative data clearly demonstrates that exclusive reliance on bibliographic databasesâeven multiple databasesârisks missing substantial relevant evidence. By systematically implementing the protocols and frameworks outlined in this application note, researchers can significantly enhance the comprehensiveness and reliability of their systematic reviews and evidence syntheses, ultimately leading to more robust conclusions and more informed decision-making in environmental management and pharmaceutical development.
The future of evidence synthesis lies in recognizing the complementary strengths of diverse search methods and strategically allocating resources across these approaches to maximize evidence identification while managing practical constraints. As distributed network models and common data models continue to evolve in multi-database studies [61], the integration of supplementary search methodologies will become increasingly sophisticated, further strengthening the foundation of evidence-based practice.
Citation chaining, also referred to as citation tracking or pearl growing, is a systematic search technique that exploits the connections between research articles to identify relevant literature for evidence synthesis [63]. This method is particularly valuable in comprehensive research methodologies such as systematic reviews, where minimizing procedural bias and ensuring literature saturation are paramount [64]. Within the context of environmental evidence research and drug development, systematic searching aims to build an unbiased and comprehensive evidence base by retrieving all possibly relevant studies from multiple sources [64]. Citation chasing serves as a crucial supplementary search method because it helps researchers identify potentially relevant studies that might not be retrieved by standard bibliographic database searches [64]. This is especially critical in fields like environmental science and pharmacology, where research terminology may be disconnected, inconsistent, or span multiple disciplinary boundaries [63].
The fundamental principle of citation chaining operates on the establishment of relationships between scholarly works. Researchers begin with a set of "seed references"âarticles known to be relevant to the research topic [63]. These seeds are then used to trace scholarly conversations both backward and forward through time. The terminology for these methods can vary, but they are generally sub-categorized into direct citation tracking (backward and forward) and indirect citation tracking (co-citation and co-citing) [63]. By creating this chain of related sources, researchers can efficiently expand their literature base beyond the limitations of keyword searches, which are often constrained by the specific terminology and indexing practices of individual databases [64]. This approach is indispensable for building on existing work and ensuring that syntheses of evidence, such as those required for environmental policy or drug safety evaluations, are as complete and unbiased as possible.
Citation tracking encompasses several distinct methods, each with a specific directional relationship to the seed article [63]. The most common forms include:
Empirical evidence demonstrates the substantial yield that citation chasing can provide in systematic search efforts. The following table summarizes potential results from a typical citation chasing exercise using modern digital tools:
Table 1: Representative Output from a Citation Chasing Exercise on 33 Seed Articles [67]
| Chaining Direction | Total References Retrieved | Unique References After Deduplication | Potential Focus Threshold |
|---|---|---|---|
| Backward | 1,374 | 1,144 | References cited by â¥5 seed articles |
| Forward | >9,582 | 9,582 | Citations from â¥5 seed articles |
The data illustrates the powerful expansion capability of forward citation chasing, which often yields a substantially larger volume of unique references than backward chasing [67]. The application of a "focus threshold"âfor instance, considering only those references or citations that are shared by multiple seed articlesâcan help manage the volume of results and prioritize highly influential or convergent works [67]. This quantitative benefit is complemented by a critical qualitative advantage: citation chasing is particularly effective at identifying studies that are semantically linked to the research topic but are "terminologically disconnected," meaning they would not be found using the specific keywords and Boolean operators of a standard database search [64] [63].
This section provides a detailed, step-by-step protocol for executing a comprehensive citation chase, suitable for systematic reviews in environmental evidence and drug development.
Objective: To identify a robust and comprehensive set of scholarly references related to a defined research question through backward and forward citation chasing from a validated set of seed articles.
Principle: This protocol uses the open-source tool Citationchaser (available as an R package and web-based Shiny app) due to its transparency, efficiency, and reliance on the extensive Lens.org academic index [64] [67]. The process can be iterative, with newly identified relevant references serving as new seed references for subsequent chasing rounds.
Materials and Reagents:
Citationchaser Shiny app (https://estech.shinyapps.io/citationchaser/) or the R package.Procedure:
Seed Article Preparation and Input
Citationchaser Shiny app. On the "Article input" tab, upload the RIS file. Alternatively, you can manually enter a list of DOIs separated by commas.Backward Citation Chasing
Forward Citation Chasing
Result Management and Screening
The following diagram illustrates the logical workflow and decision points for the citation chaining protocol described above.
Table 2: Essential Tools and Platforms for Effective Citation Chaining
| Tool / Resource | Type | Primary Function in Citation Chaining | Key Consideration |
|---|---|---|---|
| Citationchaser [64] [67] | Software Tool | An open-source Shiny app & R package for bulk backward and forward citation chasing using the Lens.org API. | Promotes transparency and reproducibility. Free to use. Coverage depends on Lens.org. |
| Lens.org [67] | Bibliographic Database | A massive open academic index that serves as the primary data source for Citationchaser. | Aggregates data from multiple sources; may have a lag in updating with the very latest publications. |
| Scopus [63] [65] | Commercial Citation Database | A curated abstract and citation database used for manual forward and backward citation chasing. | High-quality data but requires an institutional subscription. |
| Web of Science [63] [65] [66] | Commercial Citation Index | Another major curated citation index for manual forward citation chasing and identifying highly cited papers. | Requires an institutional subscription. |
| RIS (Research Information System) Format [67] | Data Standard | A standardized tag format for exchanging bibliographic data between reference managers, databases, and tools like Citationchaser. | Critical for seamless transfer of seed articles and results. Supported by all major reference managers. |
| Zotero / EndNote / Mendeley [67] | Reference Manager | Software to manage seed articles, export RIS files, import results from citation chasing, and deduplicate references. | Essential for organizing the large volume of references generated. |
For rigorous systematic reviews in environmental evidence and related fields, current best practice recommends against using citation tracking in isolation. Instead, it should be deployed as one component of a multi-pronged search strategy [64] [63]. The Cochrane Handbook for Systematic Reviews of Interventions, a gold-standard source for methodology, explicitly requires backward citation chasing of included studies, while forward citation chasing is strongly suggested for reviews on complex and public health interventions [67]. The principal advantage of integrating citation chaining is its ability to mitigate "procedural bias" and identify relevant studies that are missed by database searches due to inconsistencies in terminology, indexing, or vocabulary overlaps with other fields [64] [63]. By combining systematic database searching with supplementary methods like citation chasing, handsearching, and grey literature searching, researchers can approach a more complete and unbiased evidence base, thereby strengthening the conclusions and reliability of their synthesis.
Within the rigorous framework of environmental evidence research, comprehensive literature retrieval is paramount to minimizing bias and ensuring robust synthesis. While multiple database search strategies form the backbone of this process, supplementary methods are often necessary to capture the full spectrum of relevant evidence. This protocol details the application of two critical supplementary search techniques: handsearching key journals and consulting with subject experts. These methods are designed to identify studies or data that may be missed by standard electronic database searches due to inadequate indexing, recent publication, or non-traditional dissemination channels [68] [44]. This document provides detailed Application Notes and Experimental Protocols for their implementation, contextualized within a broader thesis on advanced search methodologies.
Handsearching and expert consultation serve as vital supplements to bibliographic database searching. Their primary function is to identify eligible study reports that might otherwise be overlooked [44]. Handsearching involves a manual page-by-page examination of the entire contents of journal issues or conference proceedings to identify all relevant reports, irrespective of their presence in databases or the quality of their indexing [68] [69]. Consulting with subject experts utilizes dialogue and discussion with topic authorities to locate unpublished reports, linked publications, or to clarify details in existing study reports [44]. The mechanism of action for these methods differs fundamentally from database searching; they rely on human scrutiny and professional networks rather than query formulation against indexed fields [44].
The justification for employing these resource-intensive methods is well-supported. Key reasons include:
Evidence suggests that handsearching, in particular, can achieve superior recall. One case study focusing on conference proceedings found that handsearching identified 604 potentially eligible abstracts and demonstrated perfect recall (100%) when compared to other search methods, though it was noted for poor efficiency in exporting records for screening [69].
To manually identify all potentially eligible studies or study reports published in a defined set of key journals through a page-by-page examination of each journal issue, covering a specified time period.
Table 1: Research Reagent Solutions for Handsearching
| Item Name | Function/Application |
|---|---|
| Journal TOCs Service | A service that emails subscribers the tables of contents of selected journals to track future issues automatically [71]. |
| BrowZine Account | A platform that allows access to and browsing of online journals, enabling organization of favorite journals into a personal "bookshelf" for ongoing monitoring [68]. |
| Bibliographic Management Tool | Software used to store, manage, and export references. Note that some journal websites may not support bulk export, impacting efficiency [69]. |
| Unpaywall Extension | A browser extension to find free Open Access versions of journal articles, which is useful when a journal is not available via institutional subscription [68]. |
Identify Key Journals: Compile a list of key journals relevant to the research topic.
Define Scope and Timeframe: Determine the number of years back from the present date for which you will examine each journal. This must be applied consistently across all selected journals [71].
Access Journal Content: Navigate to the publisher's website for each journal. Institutional library subscriptions often provide access [68]. Platforms like BrowZine can facilitate this process.
Conduct Manual Examination: For each issue within the defined timeframe, perform a page-by-page review of the entire contents. This includes scanning the table of contents, but also examining full articles, advertisements, announcements, book reviews, and any other content for potentially relevant material [70].
Record and Export Findings: Document all potentially eligible reports. The efficiency of this step varies; some publisher sites allow for bulk export, while others may require each abstract or citation to be identified and downloaded individually, which is a known resource bottleneck [69].
Integrate Results: Combine the records identified through handsearching with those from other search methods, removing duplicate entries.
The following workflow diagram illustrates the handsearching protocol:
To identify unpublished reports, linked publications, or clarify details in study reports through direct communication with experts in the field of interest.
Table 2: Research Reagent Solutions for Expert Consultation
| Item Name | Function/Application |
|---|---|
| Professional Networking Platforms | Sites like LinkedIn or academic-focused networks (e.g., ResearchGate) to identify and initiate contact with domain experts. |
| Institutional Websites | University department pages and professional organization directories to locate experts and their contact information. |
| Email Client | Primary tool for formal, documented communication with study authors and subject matter experts. |
| Citation Tracking Tools | Resources like Scopus, Web of Science, or Google Scholar to identify leading authors based on publication and citation metrics. |
Identify Potential Experts: Create a list of potential contacts. This can include:
Formulate Contact Strategy: Develop a standardized message or script explaining the purpose of your systematic review or evidence synthesis, the type of studies or data you are searching for, and why their input is valuable [44].
Initiate Contact: Reach out via professional email. The communication should be concise, respectful of the expert's time, and clearly state what you are requesting (e.g., information on unpublished studies, linked publications, or confirmation of data).
Document Interactions: Maintain a record of all experts contacted, the date of contact, the nature of the inquiry, and any responses received. This is crucial for the transparency and reproducibility of the search process.
Manage Acquired Information: Process any studies, data, or references provided by experts through the same screening and data extraction pipeline as records identified from other sources.
Acknowledge Contributions: Where appropriate, acknowledge the assistance of experts in your final systematic review or publication.
The following workflow diagram illustrates the expert consultation protocol:
The effectiveness and resource burden of supplementary search methods are critical considerations in the planning stages of a systematic review. The following table synthesizes quantitative data from empirical studies, providing a comparison for researchers.
Table 3: Comparison of Supplementary Search Method Effectiveness and Resource Use
| Search Method | Key Effectiveness Findings | Efficiency & Resource Considerations |
|---|---|---|
| Handsearching | Identified 100% of known eligible conference abstracts (604 records) in a case study; superior recall [69]. | Resource-intensive; exporting 604 records required individual download of each abstract, adding significant time [69]. |
| Citation Chasing | Cited as effective for identifying studies missed by database searching due to mis-indexing or recent publication [44]. | Tools like Citation Chaser can improve efficiency; traditional manual methods are time-consuming [72]. |
| Consulting Study Authors | Effective for identifying unpublished reports, linked publications, and clarifying data [44]. | Requires time for identifying contacts, correspondence, and managing responses. Dialogue can be more efficient than complex database searches [44]. |
| Bibliographic Database Searching | Primary method for comprehensive searching, but may miss relevant studies [44]. | Highly structured and efficient for searching large volumes of indexed literature; performance depends on search strategy quality [44]. |
For a comprehensive search in environmental evidence research, handsearching and expert consultation should not be performed in isolation. The following diagram illustrates how these methods integrate within a broader multiple database search strategy.
Within environmental evidence research, the ability to conduct comprehensive, unbiased evidence syntheses is paramount. Systematic reviews and maps rely on methodological rigor to minimize bias and provide reliable conclusions that can inform policy and management decisions [16] [15]. This document provides detailed Application Notes and Protocols for one of the most critical phases of this process: systematically searching clinical trials registries and regulatory agency sources. These sources are vital for mitigating publication biasâthe tendency for statistically significant or "positive" results to be published more readily than null or negative findingsâwhich can lead to a significant overestimation of effects in synthesis outcomes [16] [15]. The protocols outlined herein are designed to be integrated into a broader thesis on multiple database search strategies, ensuring that researchers can locate and incorporate the full spectrum of relevant evidence, including ongoing, completed, and unpublished studies.
A clear understanding of the following terms is essential for implementing the subsequent protocols effectively [16] [15]:
Failing to include all relevant evidence in a synthesis can significantly affect or bias its findings [16] [15]. Relying solely on published, English-language literature from commercial bibliographic databases introduces several risks:
A metaresearch study confirmed that searching two or more databases significantly decreases the risk of missing relevant studies, underscoring the importance of a multi-source approach [45]. This principle extends directly to the use of registries and regulatory sources to ensure all relevant study data is captured.
Searching is not an isolated activity but a key component of a broader systematic workflow. The figure below outlines the key stages of the search process within evidence synthesis, from planning to reporting.
Figure 1: Systematic Search Workflow for Evidence Synthesis. The process is iterative, moving from planning and development through to execution and final reporting, ensuring transparency and reproducibility [16] [2] [5].
To identify and retrieve records of completed, ongoing, and terminated clinical trials relevant to the evidence synthesis question, thereby minimizing publication and time-lag biases.
Step 1: Define Registry Scope Identify which national and international registries are most likely to host trials relevant to the research topic. A non-comprehensive list of major registries is provided in Table 1. The World Health Organization's International Clinical Trials Registry Platform (ICTRP) is a recommended starting point as it acts as a voluntary coordinating body for many international registries [73].
Step 2: Develop a Standardized Search String
Step 3: Execute and Document the Search
Step 4: Screen and Manage Records
Table 1: Select Clinical Trial Registries for Evidence Synthesis
| Registry Name | Scope | Access | URL |
|---|---|---|---|
| ClinicalTrials.gov | United States (largest registry) | Public | clinicaltrials.gov |
| WHO ICTRP Search Portal | International (aggregates from multiple registries) | Public | who.int/ictrp |
| EU Clinical Trials Register | European Union | Public | clinicaltrialsregister.eu |
| ISRCTN Registry | International (all study types) | Public | isrctn.com |
| ANZCTR | Australia & New Zealand | Public | anzctr.org.au |
| ChiCTR | China | Public | chictr.org.cn |
To identify regulatory documents, including published rules, notices, and docket materials from government agencies, which may contain unique data, reports, and scientific assessments not found in the academic literature.
Step 1: Identify Relevant Agencies and Portals Determine which government agencies (e.g., Environmental Protection Agency, Food and Drug Administration, European Medicines Agency) have jurisdiction over the topic. Identify the primary portals for accessing their regulatory information.
Step 2: Develop a Targeted Search Strategy
Step 3: Execute and Document the Search
Step 4: Retrieve and Archive Documents
Table 2: Key U.S. Regulatory Information Sources
| Source Name | Description | Content | URL |
|---|---|---|---|
| Regulations.gov | Portal for finding and commenting on U.S. federal regulations | Proposed & final rules, supporting documents, public comments | regulations.gov |
| Federal Register (GovInfo) | Official daily publication for rules, proposed rules, and notices of federal agencies | Volumes from 1936 to present | govinfo.gov |
| eCFR | Continuously updated online version of the Code of Federal Regulations (CFR) | Current, codified regulations | ecfr.gov |
| HeinOnline CFR | Historical archive of the CFR | Superseded regulations (1938-present) | HeinOnline |
| ProQuest Regulatory Insight | Compiles regulatory histories | Links rules to their enabling statutes and public comments | ProQuest |
This toolkit details essential resources for executing the search protocols described above.
Table 3: Essential Tools for Systematic Searching of Registries and Regulations
| Tool / Resource | Function | Application in Protocol |
|---|---|---|
| Citation Management Software (e.g., EndNote, Zotero) | Manages, deduplicates, and stores bibliographic records. Essential for handling large numbers of citations from diverse sources. | Used in both protocols to manage and organize search results prior to screening [2]. |
| WHO ICTRP Search Portal | Provides a single point of access to search multiple international trial registries simultaneously. | Used in Protocol 1, Step 1 to efficiently identify relevant trials across global registries [73]. |
| Regulations.gov | Central portal for accessing U.S. federal regulatory materials, including dockets and public comments. | The primary tool for Protocol 2, enabling discovery of regulatory grey literature [75]. |
| PRESS Checklist | A peer-review checklist designed to evaluate the quality of electronic search strategies. | Used during search strategy development to minimize errors and biases before final execution [5]. |
| Test-List of Articles | A benchmark set of known relevant articles compiled independently of the main search. | Used to test and validate the performance of the search strategy during its development phase [16]. |
| Data Management Plan | A formal document outlining how data will be handled, stored, and preserved during and after the research project. | Critical for both protocols to ensure the integrity and traceability of retrieved data and documentation [2]. |
Integrating systematic searches of clinical trials registries and regulatory agency sources is a non-negotiable component of a rigorous evidence synthesis in environmental research. The protocols detailed herein provide a structured, reproducible methodology for accessing this critical body of grey literature. By doing so, researchers can directly address pervasive biases like publication bias, thereby producing more reliable, comprehensive, and unbiased syntheses of the available evidence. This approach solidifies the scientific foundation upon which environmental policy and management decisions are made.
Systematic web searching and grey literature retrieval are fundamental components of comprehensive evidence synthesis, particularly in environmental evidence research where publication bias can significantly skew research findings. Grey literatureâdefined as research published outside traditional commercial publishing channelsâincludes technical reports, dissertations, conference proceedings, and trial registries that are essential for mitigating publication bias [77]. This bias occurs when studies with "positive" or statistically significant results are three times more likely to be published than those showing null or negative findings, creating a "file-drawer" problem that distorts the evidence base [77]. Within environmental evidence research, where data comes from diverse sources and terminology varies widely, systematic retrieval methodologies are particularly crucial for ensuring evidence syntheses are both comprehensive and reliable [1].
Grey literature encompasses multiple document types that are critical for evidence synthesis:
The strategic importance of grey literature is best illustrated through case examples like the antidepressant Agomelatine, where published trials showed modest benefits over placebo, while five unpublished trials found no effectiveness compared to placebo [77]. This publication bias created a distorted perception of drug efficacy that was only apparent through grey literature retrieval.
Publication bias represents a significant threat to evidence synthesis validity. The tendency for researchers, reviewers, and editors to preferentially publish studies with positive results leads to systematic evidence distortions. This bias operates at multiple levels:
Table 1: Types of Publication Bias in Research Synthesis
| Bias Type | Mechanism | Impact on Evidence |
|---|---|---|
| Submission Bias | Researchers don't submit null results | Overestimation of effects |
| Acceptance Bias | Journals reject null findings | Inflated significance claims |
| Language Bias | English publication preference | Reduced global perspective |
| Time-lag Bias | Delayed null result publication | Early over-optimism |
Environmental evidence research requires searching multiple database types to ensure comprehensive coverage. Cochrane guidelines recommend a minimum of three core databases be searched, typically including CENTRAL, MEDLINE, and Embase [78]. For environmental topics, specialized databases like Global Health via Ovid provide unique coverage of public health research with extensive grey literature inclusion, including international journals, research reports, patents, standards, dissertations, and conference proceedings [77].
Additional specialized databases critical for environmental evidence research include:
Effective search strategies for environmental evidence must account for differential search term sensitivityâwhere compound search terms do not perform equally across all subdomains [1]. This is particularly challenging in environmental research where terminology is rarely standardized.
The standard search approach combines population, intervention, and outcome terms in the format: ãpopulation termsã AND ãintervention termsã AND ãoutcome termsã [1]. However, for complex environmental topics like nutrient recovery from wastewater, additional targeted searches for specific subdomains (e.g., "urine AND struvite precipitation," "feces AND vermicomposting") are necessary to ensure comprehensive coverage [1].
Table 2: Search Strategy Components for Environmental Evidence
| Component | Purpose | Environmental Examples |
|---|---|---|
| Population Terms | Define subject scope | wastewater, human excreta, sewage sludge |
| Intervention Terms | Specify actions studied | nutrient recovery, reuse, recycle |
| Outcome Terms | Identify measured results | agricultural application, crop yield |
| Subdomain Terms | Capture specialized areas | struvite precipitation, vermicomposting |
Grey literature retrieval requires targeting specific repository types beyond traditional bibliographic databases. Essential sources include:
Conference abstract inclusion requires careful consideration. While Cochrane and the United States National Academy of Sciences recommend always including conference abstracts to mitigate publication bias, they present challenges: they often lack methodological details, report preliminary results, and may not be peer-reviewed [78]. The decision to include abstracts should be based on the review's purpose, timeline, and resources for following up with authors for additional information.
Data repositories provide access to underlying research data and can be crucial sources for evidence synthesis. These include subject-specific repositories for environmental data and general repositories like the King's Research Data Management System [77]. When searching repositories, consider:
Environmental evidence synthesis frequently encounters overwhelmingly large bodies of research. When facing resource constraints, researchers must streamline processes while maintaining validity. Efficiency measures include:
The Egestabase project, which involved screening over 150,000 studies and coding over 15,000, demonstrates how strategic prioritization enables management of large evidence bodies in environmental research [1].
Optimal search strategies for environmental evidence must address terminology challenges through:
Comparative analysis of evidence maps on nutrient recovery from wastewater showed surprisingly low overlapâonly about 10% of studies appeared in multiple evidence basesâhighlighting how search strategy differences significantly impact outcomes [1].
A systematic workflow for web searching and grey literature retrieval ensures comprehensive coverage and reproducibility:
Quality assurance mechanisms are essential for reliable evidence synthesis. Recommended approaches include:
Reported consistency checking methods include parallel screening of 0.85-1.8% of records by multiple reviewers followed by discussion of disagreements, providing measurable quality control [1].
Table 3: Essential Research Reagent Solutions for Systematic Searching
| Tool Category | Specific Resources | Function and Application |
|---|---|---|
| Bibliographic Databases | MEDLINE, Embase, CENTRAL | Core published literature searching [78] |
| Grey Literature Databases | Global Health (Ovid), HMIC, PsycEXTRA | Specialized unpublished literature retrieval [77] |
| Trial Registries | ClinicalTrials.gov, WHO ICTRP | Ongoing and completed trial identification [78] |
| Dissertation Databases | EThOS, ProQuest Dissertations, OATD | Graduate research theses locating [77] |
| Data Repositories | Subject-specific repositories, King's RDM | Underlying research data access [77] |
| Reference Management | Mendeley, Zotero, EndNote | Result deduplication and organization [50] |
Systematic web searching and grey literature retrieval require methodical approaches, particularly in environmental evidence research where terminology variability and publication bias present significant challenges. By implementing structured protocols for database searching, targeted grey literature retrieval, and rigorous quality assurance, researchers can produce more comprehensive and reliable evidence syntheses. The techniques outlined provide a framework for addressing the unique challenges of environmental evidence research while maintaining methodological rigor in the face of large evidence bodies and resource constraints.
A rigorous, multi-database search strategy is the non-negotiable foundation of any trustworthy environmental evidence synthesis. It systematically minimizes biases and maximizes the likelihood of capturing all relevant evidence, thereby protecting the integrity of the review's conclusions. This involves a meticulous process from planning and scoping to execution and validation, incorporating both traditional bibliographic databases and supplementary methods. For biomedical and clinical research, these principles are directly transferable, ensuring that drug development and health policy are informed by the most complete and unbiased body of evidence available. Future directions will involve leveraging new technological tools for search automation and further refining methods to efficiently manage the ever-growing volume of scientific literature.