Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings
Permalink

Comparing changes

Choose two branches to see what’s changed or to start a new pull request. If you need to, you can also or learn more about diff comparisons.

Open a pull request

Create a new pull request by comparing changes across two branches. If you need to, you can also . Learn more about diff comparisons here.
base repository: ModelCloud/GPTQModel
Failed to load repositories. Confirm that selected base ref is valid, then try again.
Loading
base: v0.9.4
Choose a base ref
Loading
...
head repository: ModelCloud/GPTQModel
Failed to load repositories. Confirm that selected head ref is valid, then try again.
Loading
compare: v0.9.5
Choose a head ref
Loading
  • 13 commits
  • 64 files changed
  • 5 contributors

Commits on Jul 4, 2024

  1. Update version.py (#163)

    Qubitium authored Jul 4, 2024
    Configuration menu
    Copy the full SHA
    fb388f3 View commit details
    Browse the repository at this point in the history
  2. [CI] FIX test perplexity fail (#160)

    * fix not defined error
    
    * fix test_perplexity fail
    
    * modify dataset filter text length
    
    * modify assert the difference of ppl
    
    * modify dataset filter with text length
    ZYC-ModelCloud authored Jul 4, 2024
    Configuration menu
    Copy the full SHA
    d5c1024 View commit details
    Browse the repository at this point in the history
  3. [REFRACTOR] Remove Backend.CUDA and Backend.CUDA_OLD (#165)

    * remove Backend.CUDA and Backend.CUDA_OLD
    
    * fix unit test
    
    * remove cuda_64/ and cuda_256/
    ZX-ModelCloud authored Jul 4, 2024
    Configuration menu
    Copy the full SHA
    6f1eb58 View commit details
    Browse the repository at this point in the history

Commits on Jul 5, 2024

  1. Configuration menu
    Copy the full SHA
    b250a76 View commit details
    Browse the repository at this point in the history
  2. [BACKEND] Add QBits support (#137)

    * Support QBits kernel for CPU device
    
    Signed-off-by: Cheng Penghui <penghui.cheng@intel.com>
    
    * fix merge
    
    * format
    
    * fix merge
    
    * rename to meet with latest main style
    
    * rename to meet with latest main style
    
    * fix doc
    
    * revert commented codes
    
    * add warning for fallback to cpu
    
    * remove unneeded var
    
    * fix merge
    
    * get gpu from curl
    
    * update url & use matrix
    
    * revert to main
    
    * update codes with pr comments
    
    * no 2 bit
    
    * set min to 1.4.2
    
    * fix name
    
    * add test
    
    * remove cpu check, model.device is CPU, so it cause wrong type check there
    
    * remove cpu check, model.device is CPU, so it cause wrong type check there
    
    * temp disable cuda check
    
    * add cpu check back
    
    * check module type like main
    
    * fix torch_dtype wrong which caused qbits not work
    
    * check bits support with BITS_DTYPE_MAPPING
    
    * add qbits test
    
    * add qbit test to ci
    
    * remove for now
    
    * delete test_qbits_kernel.py, it can't pass all 4 bit tests
    
    * remove cpu check again.. not sure what it is
    
    * add qbits in format tests
    
    * move test_qbits to test_cpu
    
    * no need container
    
    * setup python
    
    * update cuda check
    
    * set python to 3.10
    
    * fix check
    
    * update runner
    
    * update runner
    
    * disable download other run's artifact
    
    * set --durations=0
    
    * quant_type removed from main
    
    * quant_type removed
    
    * override device=cpu for qbits 
    
    qbits must be explicit and we do not auto switch to qbits when device=cpu. we do the reverse, and force device=cpu and backend  set to qbits
    
    * Update base.py
    
    * Update qlinear_qbits.py
    
    * qbits supports 2, 3, 4, 8 bits
    
    * Update qlinear_qbits.py
    
    * reverse/rename asym into sym
    
    * ruff
    
    * rename
    
    * rename
    
    * load qbits only as needed
    
    * cleanup
    
    * cleanup
    
    * fix device override for qbits
    
    * cleanup
    
    * cuda has been removed
    
    * format
    
    * fix check condition
    
    * fix qbits RuntimeError
    
    * fix qbits RuntimeError
    
    * remove todo
    
    * add protobuf in req & remove buggy download artifact with runid: actions/download-artifact#295
    
    * ruff
    
    ---------
    
    Signed-off-by: Cheng Penghui <penghui.cheng@intel.com>
    Co-authored-by: Cheng Penghui <penghui.cheng@intel.com>
    Co-authored-by: Qubitium-ModelCloud <qubitium@modelcloud.ai>
    3 people authored Jul 5, 2024
    Configuration menu
    Copy the full SHA
    b39fa13 View commit details
    Browse the repository at this point in the history
  3. [FIX] Delete 8 bits test (#169)

    * revert comment
    
    * remove 8 bits test
    CSY-ModelCloud authored Jul 5, 2024
    Configuration menu
    Copy the full SHA
    efb77a2 View commit details
    Browse the repository at this point in the history
  4. [MODEL] Add 2 & 3 bits support for QBits (#170)

    * add 2 & 3 bits
    
    * update SUPPORT_BITS
    CSY-ModelCloud authored Jul 5, 2024
    Configuration menu
    Copy the full SHA
    03bd744 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    87ef93f View commit details
    Browse the repository at this point in the history
  6. [CI] [FIX] used wrong tokenizer get dataset (#171)

    * fix not defined error
    
    * fix test_perplexity fail
    
    * modify dataset filter text length
    
    * modify assert the difference of ppl
    
    * modify dataset filter with text length
    
    * fix use wrong tokenizer get dataset
    
    * simplify code
    ZYC-ModelCloud authored Jul 5, 2024
    Configuration menu
    Copy the full SHA
    61191d5 View commit details
    Browse the repository at this point in the history
  7. [FEATURE] BaseQuantLinear add SUPPORTED_DEVICES (#174)

    * Check QuantLinear Device
    
    * cleanup
    
    * REFRACTOR check_cuda by introducing SUPPORTED_DEVICE into BaseQuantLinear
    
    * make device type cuda/cpu an enum
    
    * cleanup
    
    * cleanup
    ZX-ModelCloud authored Jul 5, 2024
    Configuration menu
    Copy the full SHA
    6c35fd8 View commit details
    Browse the repository at this point in the history
  8. [MODEL] Add quant support for Qbits (#173)

    * add quant support for qbits
    
    * test quant with qbits
    
    * set real sym back to quantize_config
    
    * Update qlinear_qbits.py
    
    ---------
    
    Co-authored-by: Qubitium-ModelCloud <qubitium@modelcloud.ai>
    CSY-ModelCloud and Qubitium authored Jul 5, 2024
    Configuration menu
    Copy the full SHA
    8b3c1d3 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    50aa90a View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    f0a1ee8 View commit details
    Browse the repository at this point in the history
Loading
Morty Proxy This is a proxified and sanitized view of the page, visit original site.