AI and Authorship: Systematic Review with ☸️SAIMSARA



saimsara.com

Review Stats
Identification of studies via EPMC (all fields) Identification Screening Included Records identified:n=46333Records excluded:n=0 Records assessed for eligibilityn=46333Records excluded:n=46004 Studies included in reviewn=329 PRISMA Diagram generated by ☸️ SAIMSARA
⛛OSMA Triangle Effect-of Predictor → Outcome ai and authorship  →  Outcome Beneficial for patients ΣN=105 (0%) Harmful for patients ΣN=2892 (0%) Neutral ΣN=1728343 (100%) 0 ⛛OSMA Triangle generated by ☸️SAIMSARA
Outcome-Sentiment Meta-Analysis (OSMA): (LLM-only)
Frame: Effect-of Predictor → Outcome • Source: Europe PMC
Outcome: Outcome Typical timepoints: 5-y, 7-y. Reported metrics: %, CI, p.
Common endpoints: Common endpoints: complications, mortality, functional.
Predictor: ai and authorship — exposure/predictor. Routes seen: oral, topical. Typical comparator: a human. this, alleged human authors across, human authorship, human-written manuscripts.….




1) Introduction
The rapid integration of artificial intelligence (AI), particularly large language models (LLMs), into various domains of knowledge production has fundamentally reshaped traditional concepts of authorship. From academic publishing and scientific research to creative writing and marketing communications, AI tools offer unprecedented capabilities for content generation and assistance. However, this technological advancement introduces complex challenges concerning research integrity, the authenticity of content, ethical attribution, and the very definition of a "creator." Understanding the multifaceted impact of AI on authorship is critical for establishing robust guidelines and fostering responsible innovation in an increasingly AI-augmented world.

2) Aim
This paper aims to systematically review the current landscape of AI and authorship, synthesizing findings from recent studies to identify key themes, challenges, and emerging best practices across diverse fields.

3) Methods
3.1 Eligibility criteria: Original Studies; exclude editorials, conference papers, and reviews.
3.2 Study selection: Apply the session keyword gate (strict/fuzzy/off) already used upstream.
3.3 Risk of bias: The included studies exhibit considerable heterogeneity in design. A substantial portion of the evidence is derived from studies with unspecified designs (N/A) or mixed methods, which inherently carry a higher risk of bias compared to more rigorous experimental or randomized controlled trial (RCT) designs. While some RCTs and experimental studies are present, their populations and interventions vary widely, limiting direct comparability. The reliance on qualitative assessments and diverse metrics across studies further complicates a uniform assessment of bias.
3.4 Synthesis: Autonomous multilayer AI research agent: keyword normalization, retrieval & structuring, and paper synthesis (see SAIMSARA About section for details).

4) Results
4.1 Study characteristics
Studies frequently employed mixed-methods designs or did not specify a design, alongside experimental studies and randomized controlled trials. Populations varied widely, encompassing medical professionals, academic authors, students, consumers, and general participants. Follow-up periods were predominantly not specified, though some prospective RCTs included follow-up durations of up to 16 months [124].

4.2 Main numerical result aligned to the query
The median (unweighted) accuracy for human identification of AI authorship was 57.7%, ranging from 19.6% [85] to 79.41% [161]. Many studies also indicated that humans could not reliably distinguish AI-generated content from human-authored content [6, 30, 38, 52, 55, 115, 120, 151, 190].

4.3 Topic synthesis


5) Discussion
5.1 Principal finding
The median (unweighted) accuracy for human identification of AI authorship was 57.7%, ranging from 19.6% [85] to 79.41% [161], indicating a significant challenge for humans to reliably distinguish AI-generated content from human-authored text.

5.2 Clinical implications


5.3 Research implications / key gaps


5.4 Limitations


5.5 Future directions


6) Conclusion
The median (unweighted) accuracy for human identification of AI authorship was 57.7%, ranging from 19.6% [85] to 79.41% [161], indicating a significant challenge for humans to reliably distinguish AI-generated content from human-authored text. This highlights a critical blurring of lines in authorship across various settings, from academic publishing to creative endeavors. The most significant limitation affecting certainty in this landscape is the rapid evolution and inconsistent application of AI tools and detection methods, making findings quickly outdated and comparisons difficult. To maintain scientific integrity and foster responsible innovation, clear, standardized, and adaptable guidelines for AI use and authorship disclosure are urgently needed.

References
SAIMSARA Session Index — session.json

Figure 1. Publication-year distribution of included originals
Figure 1. Publication-year distribution of included originals

Figure 2. Study-design distribution of included originals
Figure 2. Study-design distribution

Figure 3. Study-type (directionality) distribution of included originals
Figure 3. Directionality distribution

Figure 4. Main extracted research topics
Figure 4. Main extracted research topics (Results)

Figure 5. Limitations of current studies (topics)
Figure 5. Limitations of current studies (topics)

Figure 6. Future research directions (topics)
Figure 6. Future research directions (topics)