Hugging Face is an open-source AI platform and ecosystem that makes it easier to build, train, fine-tune, and share machine learning models, especially large language models (LLMs).
Think of Hugging Face as:
GitHub for AI models
A toolkit for training and experimenting with AI
A learning playground for understanding how models actually work
Hugging Face provides:
Thousands of pre-trained models (e.g., Llama 2, Falcon, Mistral)
The Transformers library for working with language, vision, and multimodal models
Datasets, evaluation tools, and training utilities
A community-driven approach to AI development
In an academic context, Hugging Face is especially valuable because it allows students to:
See how models are structured internally
Experiment beyond closed, “black-box” systems
Learn real-world AI workflows used in research and industry
In practice:
Hugging Face provides the base model (e.g., Llama 2)
PEFT provides the efficient fine-tuning method
Students train small adapter layers instead of the full model
The result is a customized model behavior with minimal compute
Importantly:
The base model remains unchanged
The adapter can be shared, versioned, and evaluated independently
This aligns well with ethical and reproducible AI practices
PEFT stands for Parameter-Efficient Fine-Tuning.
PEFT is a method that allows you to fine-tune large AI models without retraining the entire model, which would otherwise require massive computing power and cost.
Instead of updating billions of parameters, PEFT:
Trains only a small subset of parameters
Keeps the original model mostly frozen
Adds lightweight “adapters” that learn your task or style
This makes fine-tuning:
Faster
Cheaper
More accessible to students
Possible on limited hardware
Adds small trainable layers to the model
One of the most popular PEFT methods
Widely used for fine-tuning LLMs like Llama 2
A memory-optimized version of LoRA
Allows fine-tuning large models on consumer GPUs
Very common in academic and research settings
Students typically fine-tune:
Tone and style
Formatting behavior
Domain-specific responses
Task-specific outputs