23rd Discussion-20 Nov 2025

From BioWiki
Revision as of 10:50, 20 November 2025 by Skarunan (talk | contribs) (Meeting summary and document links)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigationJump to search

A knowledge sharing session on the usage of LLMs in Bioinfo cores hosted by Sivarajan Karunanithi (Siva) and James Gilbert.

Below is an AI generated summary of the [Meeting Notes](https://docs.google.com/document/d/1zdZvvpbsnktxlg-FbvhLLQ3JfXrvbX4X4r4krYi528g/edit?usp=sharing)

Summary:

* LLMs can generate code and assist with tasks like QC or documentation, but they perform poorly on complex analyses (e.g., multiomics, scRNA-seq), often producing plausible but incorrect results.
* Agentic LLM systems (Cell Voyager, Biomni) are difficult to debug, inconsistent, and prone to choosing wrong tests or outdated methods, making their analyses unreliable.
* While LLMs help non-experts write substantial code, users often lack the background knowledge to detect errors, increasing risk and reducing incentive to learn fundamentals.
* Teaching must adapt by emphasizing AI literacy, safe use, critical thinking, and maintaining foundational coding skills despite reduced student interest in learning basics.
* Institutions are exploring secure/open-source LLM deployments, with AI useful for snippets and documentation but not yet a threat to bioinformatics facilities due to its unreliability in advanced analysis.
* LLMs can create a false sense of competence, enabling inexperienced users to perform advanced analyses without understanding assumptions, which can lead to confidently wrong scientific conclusions.
* There’s no consensus on how to acknowledge AI use, raising questions about responsibility, transparency, and how AI-assisted work should be reported in scientific outputs.