|||

ML Notes

Quick search

Contents:

  • Mathematics
  • NLP
    • GPT
      • Model Design
      • Input Encoding
      • Neural Networks
      • Training and Predicting
      • Review
      • Reference
  • Computer Vision
  • Micro-nanoplastics
  • Reference

GPTΒΆ

This project is an project to store notes and codes during the study of the Generative Pre-trained Transformer (GPT) language model.

Contents:

  • Model Design
    • Parameters
    • Workflow Chart
  • Input Encoding
    • Unicode Encoding
    • Tokenization
    • Token Embedding (WTE)
    • Positional Encoding (WPE)
    • Encoding Combination
  • Neural Networks
    • Layer Normalization
    • Feed-Forward
    • Attention Layer
  • Training and Predicting
    • Pre-Training
    • Meta-Learning
    • Examples
  • Review
    • Paper Review
    • Critics
  • Reference
    • Papers
    • Implementation
    • APIs
    • Fine-Tuning
    • Other Sources

Back to NLP.

<NLP
Model Design>
© Copyright 2022, Juan Cervantes. Created using Sphinx 8.2.3.

Styled using the Piccolo Theme