π Bibliography: LLMs and Tokenization
Large Language Models Fundamentals
| Resource | Type/Focus | Language | Access |
|---|---|---|---|
| Hugging Face NLP Course | Complete course on NLP and transformers | EN | Online |
| Wikipedia: Large language model | Introduction to LLMs, history and applications | EN | Online |
| OpenAI: GPT-3 Paper | Technical paper on GPT-3 | EN | arXiv |
| Vaswani et al. (2017). "Attention Is All You Need" | Seminal paper on Transformer architecture | EN | arXiv |
Tokenization and BPE
| Resource | Type/Focus | Language | Access |
|---|---|---|---|
| Hugging Face: Byte-Pair Encoding tokenization | Detailed BPE tutorial with implementation | EN | Online |
| Wikipedia: Byte pair encoding | History and operation of BPE algorithm | EN | Online |
| Sennrich, R., Haddow, B., & Birch, A. (2016). "Neural Machine Translation of Rare Words with Subword Units" | Original paper on BPE for NLP | EN | arXiv |
| Hugging Face Tokenizers | Tokenizers library documentation | EN | Online |
Transformers and Architectures
| Resource | Type/Focus | Language | Access |
|---|---|---|---|
| The Illustrated Transformer | Visual explanation of Transformer architecture | EN | Online |
| Attention Is All You Need | Original Transformers paper (Google) | EN | arXiv |
| Stanford CS224N: Natural Language Processing with Deep Learning | Stanford course on NLP | EN | Online |
| Devlin, J. et al. (2018). "BERT: Pre-training of Deep Bidirectional Transformers" | BERT paper | EN | arXiv |
Pre-trained Models and Fine-tuning
| Resource | Type/Focus | Language | Access |
|---|---|---|---|
| Hugging Face Model Hub | Repository of pre-trained models | EN | Online |
| OpenAI API Documentation | GPT-3.5/4 documentation | EN | Online |
| Radford, A. et al. (2019). "Language Models are Unsupervised Multitask Learners" | GPT-2 paper | EN | |
| Brown, T. et al. (2020). "Language Models are Few-Shot Learners" | GPT-3 paper | EN | arXiv |
Medical Applications of LLMs
| Resource | Type/Focus | Language | Access |
|---|---|---|---|
| Nature: Large language models in medicine | LLM applications in medicine | EN | Article |
| PubMed: ChatGPT and Healthcare | Review on ChatGPT in healthcare | EN | Article |
| Med-PaLM: Google's Medical LLM | Medical-specialized LLM | EN | arXiv |
| WHO: Ethics and governance of AI for health | WHO ethical perspective on AI in health | EN |
Note: All links have been verified and are accessible online. arXiv articles are freely available. For journal articles, check your university library or Google Scholar.