Lightning AI Studios: Never set up a local environment again →
Adapter is a lightweight adaptation method that adds a small...
Alpaca is a model fine-tuned from the LLaMA 7B model...
An attention mechanism is a component used in large language...
Cosine similarity is a measure used to determine the similarity...
Embeddings are dense, lower-dimensional representations of words, sentences, or other...
Falcon is the latest open-source large language model released by Technology Innovation...
GPT-NeoX-20B is a 20 billion parameter autoregressive language model trained...
GPTQ is a one-shot weight quantization method based on approximate...
Inference refers to the process of using a trained model...
Instruction tuning, in the context of large language models, is...
LLaMA is a foundational large language model that has been...
LoRA achieves parameter reduction in large language models by learning...
Multimodal learning models the combination of different types of data...
Prefix-tuning is a lightweight alternative to fine-tuning for natural language...
Pretraining an LLM involves training the model on a large,...
Prompt engineering refers to the process of carefully designing and...
Prompt-tuning is a cost-effective and efficient method for adapting an...
QLoRA is a highly effective method for fine-tuning, which significantly...
Question answering is a natural language processing task where a...
Reinforcement learning from human feedback (RLHF), also known as reinforcement...
Self-instruct is a method that combines self-supervised learning and reinforcement...
Semantic search is a search technique that aims to understand...
StableLM is a series of language models from Stability AI...
Text summarization is the process of condensing a longer piece...
Tokenization is the process of splitting text into smaller units...
Topic modeling is a statistical technique used to uncover latent...
The Transformer is a neural network architecture introduced in the...