Aucune description

Gen TANG 2a962e005e update eng version il y a 1 an
ch03_linear b5db7b2945 update comment il y a 1 an
ch04_logit 2156c1ac35 change color for autograd il y a 1 an
ch05_econometrics 2d95750b98 update comment for ch05 il y a 2 ans
ch06_optimizer 05d6077132 update comment for ch06 il y a 2 ans
ch07_autograd c9f95f42ee update font in graph il y a 1 an
ch08_mlp c9f95f42ee update font in graph il y a 1 an
ch09_cnn 35ba0ddcea update comment for ch09 il y a 2 ans
ch10_rnn c9f95f42ee update font in graph il y a 1 an
ch11_llm 636815ed79 finish annotation for ch11 il y a 2 ans
ch12_rl af872ca8ce typo il y a 2 ans
ch13_others af872ca8ce typo il y a 2 ans
.gitignore e18835b1cd update comment for ch07 il y a 2 ans
LICENSE 0a9008828e Initial commit il y a 2 ans
README.md 2a962e005e update eng version il y a 1 an

README.md

Deconstructing Large Language Models: From Linear Regression to General Artificial Intelligence

This book is currently in the editing process and will be available soon.

Description

For classic models in AI(artificial intelligence), the tools, such as PyTorch, have provided well-encapsulated implementations, making their usage relatively straightforward. However, due to engineering considerations, these implementations introduce excessive details into the code, complicating model understanding. This book aims to enhance reader comprehension by re-implementing core model parts with detailed annotations. While explaining complex algorithms in human language can be challenging, reading the code proves more intuitive.

The code relies on external libraries, with installation commands given at the beginning of the scripts. Rerunning might yield slight result variations due to random numbers, but the overall impact is minimal. For large language models, running the code on a GPU is crucial to avoid a significant increase in computation time.

Outline