Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings
Discussion options

As part of an effort to deepen my expertise in Rust programming and to refresh my knowledge of Large Language Models I have developed a Rust version of the Python version. The implementation utilizes the Burn library for tensor operations.

I have seen someone already posted the book's code using the Candle library but my project is a little bit less ambitious it only implements the training and generation in a tool that can be used to play with the model.

I believe that having multiple implementations of the model in different languages and frameworks opens up a fantastic opportunity for bench-marking and performance analysis. Comparing the speed, memory usage, and overall efficiency of these different approaches could provide valuable insights.

Thank you for your work

You must be logged in to vote

Replies: 0 comments

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
1 participant
Morty Proxy This is a proxified and sanitized view of the page, visit original site.