Bio x AI Hackathon
  • Welcome to the Bio x AI Hackathon
  • Getting Started
    • Quickstart
    • Important Links
  • Developers
    • BioAgents
    • CoreAgents
    • Eliza Agent Framework
    • Knowledge Graphs
    • .cursorrules
    • Starter-repos
    • Plugin Guide
  • Vision and Mission
    • Bio x AI Hackathon
    • The Problems in Science
    • TechBio
    • Guidance from the Judges
      • Important Datasets and Code Repositories
      • Reading List
      • Common Mistakes for Developers new to Academia
    • Hackathon Ideas
      • Full Projects
        • The Complexity Slider - Finding Hypotheses at the Limits of Human Knowledge
        • [Hard Mode] Metadata Generation on datasets with No Manuscript or Code Associated
        • Inverse Reproducibility - Given Manuscript and Data, Make the Code
        • Atlas of Research Methods Formatted for Agentic Reuse
        • Utilizing Knowledge Graphs for the Detection of Potential Null Results
        • Creating an Iterative Publication Stack by Linking Together Existing Tooling
        • Longevity Atlas: Building a Decentralized Knowledge Network with Agentic Research Hypothesis Engine
        • CoreAgent Track - Opportunities to work with BioDAOs
        • SpineDAO Chronos Project Spec
      • Individual Plugins
        • Plug-ins for every piece of research tooling known to humankind
        • Reproducibility Assistant - Code Cleaning, Dockerization, etc
        • Finding and Differentiating Cardinal vs Supporting Assertions
        • [Easier Mode] Metadata Generation on Datasets Given the Manuscript and Code Repository
        • Sentiment Analysis on Existing Citations, Dissenting vs Confirming
        • Agentic Metadata Template Creation for Standard Lab Equipment
  • Ops
    • Calendar
      • Key Dates
      • Office Hours
    • Judges and Mentors
      • Communicating to Judges and Mentors
      • BioAgent Judging Panel
      • CoreAgent Judging Panel
      • Mentors
    • Prize Tracks
    • Hackathon Rules
    • Kickoff Speakers
    • FAQ
Powered by GitBook
On this page
  1. Vision and Mission
  2. Hackathon Ideas
  3. Individual Plugins

Sentiment Analysis on Existing Citations, Dissenting vs Confirming

Problem Statement:

Citations in academic manuscripts are not merely references; they often convey a specific relationship between the citing and cited works. Determining whether a citation indicates agreement, disagreement, or a neutral stance is crucial for understanding the evolution of scientific ideas and identifying potential controversies.

Challenge:

Develop a system that can analyze the context of citations within a scientific manuscript and classify the relationship between the citing and cited works as either agreement, disagreement, or neutral.

Detailed Description:

  • Input:

    • A scientific manuscript (PDF or text).

    • A list of citations within the manuscript.

  • System Functionality:

    • Context Extraction: Extract the sentences or paragraphs surrounding each citation.

    • Sentiment Analysis: Analyze the extracted context to determine the relationship between the citing and cited works.

    • Classification: Classify each citation as:

      • Agreement: The citing work supports or builds upon the cited work.

      • Disagreement: The citing work contradicts or challenges the cited work.

      • Neutral: The citing work simply refers to the cited work without expressing a clear opinion.

    • Output:

      • A list of citations with their corresponding sentiment classifications.

      • A structured report (e.g., JSON, CSV) summarizing the citation sentiment analysis.

  • Example Scenarios:

    • "Smith et al. (2020) demonstrated a significant increase in... This finding is consistent with our results." (Agreement)

    • "Contrary to the findings of Jones (2018), we observed..." (Disagreement)

    • "This method was based on the approach described by Brown (2015)." (Neutral)

  • Potential Technologies:

    • Natural Language Processing (NLP) for context extraction and sentiment analysis.

    • Sentence embedding models for semantic similarity analysis.

    • Machine learning classifiers for citation sentiment classification.

    • Large language models.

  • Evaluation Metrics:

    • Precision, recall, and F1-score for each sentiment class.

    • Accuracy of classification.

    • Human evaluation of the system outputs.

Desired Outcomes:

  • A functional system that can accurately classify citation sentiment.

  • A dataset of citations with labeled sentiment classifications.

  • Insights into the linguistic cues that indicate agreement, disagreement, or neutrality.

Previous[Easier Mode] Metadata Generation on Datasets Given the Manuscript and Code RepositoryNextAgentic Metadata Template Creation for Standard Lab Equipment

Last updated 1 month ago