Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Bump llama.cpp from 2347e45 to 254a7a7 #103

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from

Conversation

dependabot[bot]
Copy link
Contributor

@dependabot dependabot bot commented on behalf of github Jun 14, 2023

Bumps llama.cpp from 2347e45 to 254a7a7.

Commits

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


Dependabot commands and options

You can trigger Dependabot actions by commenting on this PR:

  • @dependabot rebase will rebase this PR
  • @dependabot recreate will recreate this PR, overwriting any edits that have been made to it
  • @dependabot merge will merge this PR after your CI passes on it
  • @dependabot squash and merge will squash and merge this PR after your CI passes on it
  • @dependabot cancel merge will cancel a previously requested merge and block automerging
  • @dependabot reopen will reopen this PR if it is closed
  • @dependabot close will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
  • @dependabot ignore this major version will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this minor version will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this dependency will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)

Bumps [llama.cpp](https://github.com/ggerganov/llama.cpp) from `2347e45` to `254a7a7`.
- [Release notes](https://github.com/ggerganov/llama.cpp/releases)
- [Commits](ggml-org/llama.cpp@2347e45...254a7a7)

---
updated-dependencies:
- dependency-name: llama.cpp
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
@dependabot dependabot bot added the dependencies Pull requests that update a dependency file label Jun 14, 2023
Copy link
Contributor

@deadprogram deadprogram left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is the commit from llama.cpp that brings in CUDA support: ggml-org/llama.cpp#1827

Let's merge this!

@mudler
Copy link
Member

mudler commented Jun 15, 2023

Yes! Definitely! Problem is, I gave a try it yesterday , but there seems to be issues still. Current master works ( no full offloading), but bringing this in seems to break it entirely.

I'm testing this with a box with a GPU given from a community member ( as I don't have one to try it myself ), and seems to fail into offloading to the GPU. @deadprogram did you tried it already on your GPU?

@mudler
Copy link
Member

mudler commented Jun 15, 2023

for reference, the error running the example:

ggml_init_cublas: found 2 CUDA devices:
  Device 0: Tesla T4                                                                                      
  Device 1: Tesla T4     
llama.cpp: loading model from /home/ubuntu/WizardLM-7B-uncensored.ggmlv3.q4_0.bin
llama_model_load_internal: format     = ggjt v3 (latest)
llama_model_load_internal: n_vocab    = 32001
llama_model_load_internal: n_ctx      = 128 
llama_model_load_internal: n_embd     = 4096                                                              
llama_model_load_internal: n_mult     = 256
llama_model_load_internal: n_head     = 32      
llama_model_load_internal: n_layer    = 32
llama_model_load_internal: n_rot      = 128                                                               
llama_model_load_internal: ftype      = 2 (mostly Q4_0)
llama_model_load_internal: n_ff       = 11008                                                             
llama_model_load_internal: n_parts    = 1
llama_model_load_internal: model size = 7B                                                                
Model loaded successfully. 
>>> w                                                                                                     
                                                     
Sending w

LLAMA_ASSERT: /home/ubuntu/go-llama.cpp/llama.cpp/llama.cpp:1372: !!kv_self.ctx
SIGABRT: abort
PC=0x7f2d1026ea7c m=0 sigcode=18446744073709551610
signal arrived during cgo execution

goroutine 1 [syscall]:

@dependabot @github
Copy link
Contributor Author

dependabot bot commented on behalf of github Jun 15, 2023

Superseded by #104.

@dependabot dependabot bot closed this Jun 15, 2023
@dependabot dependabot bot deleted the dependabot/submodules/llama.cpp-254a7a7 branch June 15, 2023 19:03
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
dependencies Pull requests that update a dependency file
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants
Morty Proxy This is a proxified and sanitized view of the page, visit original site.