Extend to matrices and high-order tensors#94
Conversation
224da0a to
e7f552f
Compare
|
Why do you keep making PRs? The repo is fine the way it is. If you re-organize it, it will not match the video demos that go along with it. |
|
This is an extension to matrices and higher-order tensors. The notebook demos still run. To see how it is like, check out https://github.com/brief-ds/micrograd @dnparadice |
|
@karpathy could you review? You can check it out at https://github.com/brief-ds/micrograd:
The from micrograd import Value, tensordot |
|
@karpathy now the |
|
@karpathy you can check how it looks on https://github.com/brief-ds/micrograd
|
|
@karpathy this version works with tensors. As the core is just one 500-line Python file This blog post TensorFlow, Apple's MLX and our micrograd explains in terms of install size and performance micrograd is in par with MLX, but micrograd is extra easy to learn, play with and profile. Would you consider merging? |
|
I made micrograd tensor-capable with in mind the idea to try simplifying the attention mechanism. Below is a proposal. Any thought? @karpathy https://www.brief-ds.com/2026/02/10/roadmap-att.html Only micrograd's characteristics makes the study possible (simplicity, opening up forward and backward methods). Feel free to merge this PR. :) |
First clean-up:
assets/demos/unittestpackage for unit tests, one dependency (onpytest) lessNext, I plan to extend the engine to matrix or higher-order tensors, and make the
backward()more efficient so that it won't have to recompute the operator topology for each piece of new input data.