Skip to content

[Incomplete] Initial fast llama modeling#24

Open
CoffeeVampir3 wants to merge 3 commits intomainfrom
llama-modeling
Open

[Incomplete] Initial fast llama modeling#24
CoffeeVampir3 wants to merge 3 commits intomainfrom
llama-modeling

Conversation

@CoffeeVampir3
Copy link

Initial modeling files for llama3 -- not hooked into llm-foundary yet.

@HangenYuu
Copy link

HangenYuu commented Feb 11, 2025

@CoffeeVampir3 mind if I push the FlashAttention commit to this branch? To put everything Llama in a single PR.

Also would be great if you can include some quick start guide for Llama, to check of implementation is running correctly.

@CoffeeVampir3
Copy link
Author

CoffeeVampir3 commented Feb 12, 2025

@CoffeeVampir3 mind if I push the FlashAttention commit to this branch? To put everything Llama in a single PR.

Yeah! That sounds like a plan. I think anything related to modeling/optimization this is a fine place to put.

Also would be great if you can include some quick start guide for Llama, to check of implementation is running correctly.

I do have a testing repo here, I did not want to pollute the repo with test code that was unrelated to llm-foundary, but I did confirm the modeling is correct and trainable here: https://github.com/CoffeeVampir3/Llama-3-Clean-Minimized

For the llm-foundary specific, there's a variety of hooks needed to be added into the modeling. I don't think it'll be too difficult, but just a matter of doing it. Quick start for the foundary stuff is waiting on that.

@AbdalrahmanWael
Copy link
Contributor

For the llm-foundary specific, there's a variety of hooks needed to be added into the modeling. I don't think it'll be too difficult, but just a matter of doing it. Quick start for the foundary stuff is waiting on that.

@CoffeeVampir3 I can try having it work with llm-foundary in a manner similar to how they did for their mpt model, if you haven't already started on that

@CoffeeVampir3
Copy link
Author

For the llm-foundary specific, there's a variety of hooks needed to be added into the modeling. I don't think it'll be too difficult, but just a matter of doing it. Quick start for the foundary stuff is waiting on that.

@CoffeeVampir3 I can try having it work with llm-foundary in a manner similar to how they did for their mpt model, if you haven't already started on that

I haven't started, would be great if you're interested 👍

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants