Large Language Models (LLMs)
Created: =dateformat(this.file.ctime,"dd MMM yyyy, hh:mm a") | Modified: =dateformat(this.file.mtime,"dd MMM yyyy, hh:mm a")
Tags: knowledge
Overview
Related fields
Introduction
Perplexity Metric (PPL)
- metric to determine the accuracy of next token prediction for language models. where lower perplexity, better the model
- the metric applies specifically to classical language models (sometimes called autoregressive or causal language models) and is not well defined for masked language models like BERT
- defined as the exponentiated average negative log-likelihood of a sequence
- Intuitively, it can be thought of as an evaluation of the model’s ability to predict uniformly among the set of specified tokens in a corpus.
- the tokenization procedure has a direct impact on a model’s perplexity which should always be taken into consideration when comparing different models.
- Evaluation Metrics for Language Modeling
Theoretical References
Papers
Articles
Courses
- GitHub - mlabonne/llm-course: Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
- quite afew resources for blogposts regarding following topics:

Code References
Methods
- GitHub - haotian-liu/LLaVA: [NeurIPS’23 Oral] Visual Instruction Tuning: LLaVA (Large Language-and-Vision Assistant) built towards GPT-4V level capabilities.
- GitHub - SkunkworksAI/BakLLaVA
- microsoft/phi-2 · Hugging Face
- mistralai/Mistral-7B-v0.1 · Hugging Face
- mistralai/Mixtral-8x7B-v0.1 · Hugging Face
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B · Hugging Face
- SmolVLM2: Bringing Video Understanding to Every Device
Tools, Frameworks
- OpenAccess-AI-Collective/axolotl
- User-friendly and powerful fine-tuning tool that is used in a lot of state-of-the-art open-source models.
- GitHub - unslothai/unsloth: Finetune Llama 3.3, DeepSeek-R1, Gemma 3 & Reasoning LLMs 2x faster with 70% less memory! 🦥
- fintuning
- exo
- Determine model size reqs