a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
-
Updated
Jun 25, 2024 - Jupyter Notebook
a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
This repository features a custom-built decoder-only language model (LLM) with a total of 37 million parameters 🔥. I train the model to be able to ask question from a given context
a dna sequence generation/classification using transformers
Implementation of the GPT-3 paper: Language Models are Few-Shot Learners
A miniGPT inspired from the original NanoGPT released by OpenAI. This is a notebook to walk through the decoder part of the transformer architecture with details outlined.
An LLM based tool for generation of cheese advirtisements
Experimental project for AI and NLP based on Transformer Architecture
Transformers Intuition
A Decoder Based Semantic Parser that can be tested on four benchmark datasets (ATIS, GeoQuery, Jobs640 and Django)
Decoder model for language modelling
Generate caption on images using CNN Encoder- LSTM Decoder structure
Add a description, image, and links to the decoder-model topic page so that developers can more easily learn about it.
To associate your repository with the decoder-model topic, visit your repo's landing page and select "manage topics."