Potentials and Pitfalls of Using Large Language Models
This repository contains the slides, literature, and figures for the talk "Tokens of Love" about Large Language Models (LLMs). It was delivered on February 14, 2025 (Valentine's Day) by Matthias Zepper at SciLifeLab in Stockholm.
The presentation explores the potentials and pitfalls of using LLMs, covering topics such as AI integrations, generative models, prompt engineering, and the future of AI in scientific research. It is a gentle introduction to the topic for non-developers and geared towards wet lab biologists.
- The presentation slides rendered to PDF
- The source file for the presentation is
scilifelab-llms.tex
.
To compile the slides, ensure you have LaTeX with Beamer installed, you will need BibLaTeX/Biber for bibliography management. Compile using lualatex
or xelatex
followed by biber
and another LaTeX run.
Note
My custom SciLifeLab theme is included in the repository. It uses the colors of the official visual guide, but not the font faces. I took the liberty to replace Lato/Lora with Biolinum by Philipp H. Poll, which you may have to install separately.
./literature/
: Includes the bibliography filescilifelabllms.bib
with references cited in the presentation../figures/
: Contains all the images and figures used in the slides../training_data_example/
: Some public training data examples that were eventually not used for the presentation.
- Valentine's Day 2025: Posits that mankind is in love with AI.
- AI Integrations and Services: Highlights various AI applications in data analysis, pipeline development, and synthetic DNA generation.
- Encounters with AI-Generated Content: Examines the impact of AI-generated content on scientific language and search engines.
- Crystal-ball gazing: Contrasts extremely negative and positive scenarios of the future, and suggests that knowing more about AI can never harm.
- Primer on Language Models: Provides a detailed explanation of language models, their training, and their applications.
- Architectures and Key Innovations: Discusses the architecture of LLMs, including transformers and token embeddings.
- Generative Models: Explores the process of text generation and the importance of context.
- Interacting with LLM Models: Covers prompt engineering and reasoning models.
- Augmenting LLM Models: Introduces retrieval-augmented generation (RAG) and the Model Context Protocol.
- Outlook: Speculates on the future of AI and its impact on information access.
All original (self/AI-generated) content in these slides and this repository is released under a Creative Commons Attribution 4.0 International License. However, certain copyrighted images, including works of art and cartoon characters, are used under Fair Use for nonprofit educational purposes and are not covered by this license.