PinnedPublished inTowards Data ScienceTuning-Free Longer Context Lengths For LLMs — A Review of Self-Extend (LLM Maybe LongLM)A simple strategy to enable LLMs to consume longer context length inputs during inference without the need for finetuning.Jan 4, 20241Jan 4, 20241
PinnedPublished inTowards Data ScienceDemystifying GQA — Grouped Query Attention for Efficient LLM Pre-trainingThe variant of multi-head attention powering LLMs like LLaMA-2, Mistral7B, etc.Dec 27, 20233Dec 27, 20233
PinnedPublished inTowards Data ScienceUnderstanding LoRA — Low Rank Adaptation For Finetuning Large ModelsFine-tuning large pre-trained models is computationally challenging, often involving adjustment of millions of parameters. This…Dec 22, 20234Dec 22, 20234
PinnedPublished inTowards Data ScienceQuantum Computing ?/!Zeros and ones. This is how we imagined computing till now. This is what classical computing is. But a whole new concept is now changing…Jul 8, 20181Jul 8, 20181
ORPO — Preference Optimization without Reference ModelCombing Instruction Fine-tuning and Preference Alignment in a single stage.Aug 29, 2024Aug 29, 2024
Training HaarCascade Model on Microsoft Azure.In this hands-on tutorial, we will learn how to train your own haar cascade model on Microsoft Azure. To understand Haarcascade I…Dec 24, 20182Dec 24, 20182
Blockchain ?/ ! The Start of a new revolutionToday every big company you can think of is investing in Blockchain. From Tech giants like Microsoft and IBM to Financial giants like…Oct 18, 20171Oct 18, 20171