Skip to content
This repository has been archived by the owner on Oct 11, 2024. It is now read-only.

andy/bump main to v0.3.2 #49

Closed
wants to merge 113 commits into from
Closed

andy/bump main to v0.3.2 #49

wants to merge 113 commits into from

Commits on Jan 26, 2024

  1. Configuration menu
    Copy the full SHA
    6b7de1a View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    3a0e1fc View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    390b495 View commit details
    Browse the repository at this point in the history

Commits on Jan 27, 2024

  1. Configuration menu
    Copy the full SHA
    beb89f6 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    220a476 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    3801700 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    5f036d2 View commit details
    Browse the repository at this point in the history

Commits on Jan 28, 2024

  1. Configuration menu
    Copy the full SHA
    f8ecb84 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    89be30f View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    7d64841 View commit details
    Browse the repository at this point in the history

Commits on Jan 29, 2024

  1. Support FP8-E5M2 KV Cache (vllm-project#2279)

    Co-authored-by: zhaoyang <[email protected]>
    Co-authored-by: Zhuohan Li <[email protected]>
    3 people authored Jan 29, 2024
    Configuration menu
    Copy the full SHA
    9090bf0 View commit details
    Browse the repository at this point in the history
  2. Fix error when tp > 1 (vllm-project#2644)

    Co-authored-by: zhaoyang-star <[email protected]>
    zhaoyang-star and zhaoyang-star authored Jan 29, 2024
    Configuration menu
    Copy the full SHA
    b72af8f View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    1b20639 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    ea8489f View commit details
    Browse the repository at this point in the history

Commits on Jan 30, 2024

  1. Configuration menu
    Copy the full SHA
    5d60def View commit details
    Browse the repository at this point in the history
  2. Fused MOE for Mixtral (vllm-project#2542)

    Co-authored-by: chen shen <[email protected]>
    pcmoritz and scv119 authored Jan 30, 2024
    Configuration menu
    Copy the full SHA
    ab40644 View commit details
    Browse the repository at this point in the history
  3. Fix 'Actor methods cannot be called directly' when using `--engine-us…

    …e-ray` (vllm-project#2664)
    
    * fix: engine-useray complain
    
    * fix: typo
    HermitSun authored Jan 30, 2024
    Configuration menu
    Copy the full SHA
    d79ced3 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    4f65af0 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    bbe9bd9 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    105a40f View commit details
    Browse the repository at this point in the history

Commits on Jan 31, 2024

  1. Configuration menu
    Copy the full SHA
    3dad944 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    1af090b View commit details
    Browse the repository at this point in the history
  3. Fixes assertion failure in prefix caching: the lora index mapping sho…

    …uld respect prefix_len (vllm-project#2688)
    
    Signed-off-by: Tao He <[email protected]>
    sighingnow authored Jan 31, 2024
    Configuration menu
    Copy the full SHA
    d69ff0c View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    c664b0e View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    89efcf1 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    d0d93b9 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    93b38be View commit details
    Browse the repository at this point in the history

Commits on Feb 1, 2024

  1. Configuration menu
    Copy the full SHA
    cd9e60c View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    923797f View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    b9e96b1 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    bb8c697 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    c410f5d View commit details
    Browse the repository at this point in the history
  6. Remove hardcoded device="cuda" to support more devices (vllm-proje…

    …ct#2503)
    
    Co-authored-by: Jiang Li <[email protected]>
    Co-authored-by: Kunshang Ji <[email protected]>
    jikunshang and bigPYJ1151 authored Feb 1, 2024
    Configuration menu
    Copy the full SHA
    96b6f47 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    0e163fc View commit details
    Browse the repository at this point in the history

Commits on Feb 2, 2024

  1. Configuration menu
    Copy the full SHA
    4abf633 View commit details
    Browse the repository at this point in the history

Commits on Feb 4, 2024

  1. Configuration menu
    Copy the full SHA
    5ed704e View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    51cd22c View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    5a6c81b View commit details
    Browse the repository at this point in the history

Commits on Feb 5, 2024

  1. Require triton >= 2.1.0 (vllm-project#2746)

    Co-authored-by: yangrui1 <[email protected]>
    whyiug and whyiug20231206 authored Feb 5, 2024
    Configuration menu
    Copy the full SHA
    c9b45ad View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    72d3a30 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    56f738a View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    b92adec View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    2ccee3d View commit details
    Browse the repository at this point in the history

Commits on Feb 6, 2024

  1. Configuration menu
    Copy the full SHA
    f0d4e14 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    ed70c70 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    fe6d09a View commit details
    Browse the repository at this point in the history

Commits on Feb 7, 2024

  1. Configuration menu
    Copy the full SHA
    c81dddb View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    931746b View commit details
    Browse the repository at this point in the history

Commits on Feb 8, 2024

  1. Configuration menu
    Copy the full SHA
    65b89d1 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    3711811 View commit details
    Browse the repository at this point in the history

Commits on Feb 11, 2024

  1. Configuration menu
    Copy the full SHA
    0580aab View commit details
    Browse the repository at this point in the history

Commits on Feb 12, 2024

  1. Configuration menu
    Copy the full SHA
    4ca2c35 View commit details
    Browse the repository at this point in the history
  2. Refactor 2 awq gemm kernels into m16nXk32 (vllm-project#2723)

    Co-authored-by: Chunan Zeng <[email protected]>
    zcnrex and Chunan Zeng authored Feb 12, 2024
    Configuration menu
    Copy the full SHA
    5638364 View commit details
    Browse the repository at this point in the history

Commits on Feb 13, 2024

  1. Configuration menu
    Copy the full SHA
    a4211a4 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    f964493 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    5c976a7 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    ea35600 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    a463c33 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    317b29d View commit details
    Browse the repository at this point in the history
  7. Add LoRA support for Mixtral (vllm-project#2831)

    * add mixtral lora support
    
    * formatting
    
    * fix incorrectly ported logic
    
    * polish tests
    
    * minor fixes and refactoring
    
    * minor fixes
    
    * formatting
    
    * rename and remove redundant logic
    
    * refactoring
    
    * refactoring
    
    * minor fix
    
    * minor refactoring
    
    * fix code smell
    tterrysun authored Feb 13, 2024
    Configuration menu
    Copy the full SHA
    2a543d6 View commit details
    Browse the repository at this point in the history

Commits on Feb 14, 2024

  1. Configuration menu
    Copy the full SHA
    7eacffd View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    0c48b37 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    7e45107 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    87069cc View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    4efbac6 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    25e86b6 View commit details
    Browse the repository at this point in the history

Commits on Feb 15, 2024

  1. Align LoRA code between Mistral and Mixtral (fixes vllm-project#2875) (

    …vllm-project#2880)
    
    * Fix AttributeError: MixtralModel object has no attribute org_vocab_size.
    
    * Make LoRA logic for Mistral and Mixtral the same
    
    ---------
    
    Co-authored-by: Pernekhan Utemuratov <[email protected]>
    pcmoritz and Pernekhan authored Feb 15, 2024
    Configuration menu
    Copy the full SHA
    31348df View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    d7afab6 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    4f2ad11 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    5255d99 View commit details
    Browse the repository at this point in the history

Commits on Feb 16, 2024

  1. Configuration menu
    Copy the full SHA
    64da65b View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    5f08050 View commit details
    Browse the repository at this point in the history

Commits on Feb 17, 2024

  1. Defensively copy sampling_params (vllm-project#2881)

    If the SamplingParams object passed to LLMEngine.add_request() is mutated after it returns, it could affect the async sampling process for that request.
    
    Suggested by @Yard1 vllm-project#2514 (comment)
    njhill authored Feb 17, 2024
    Configuration menu
    Copy the full SHA
    185b2c2 View commit details
    Browse the repository at this point in the history
  2. multi-LoRA as extra models in OpenAI server (vllm-project#2775)

    how to serve the loras (mimicking the [multilora inference example](https://github.com/vllm-project/vllm/blob/main/examples/multilora_inference.py)):
    ```terminal
    $ export LORA_PATH=~/.cache/huggingface/hub/models--yard1--llama-2-7b-sql-lora-test/
    $ python -m vllm.entrypoints.api_server \
     --model meta-llama/Llama-2-7b-hf \
     --enable-lora \
     --lora-modules sql-lora=$LORA_PATH sql-lora2=$LORA_PATH
    ```
    the above server will list 3 separate values if the user queries `/models`: one for the base served model, and one each for the specified lora modules. in this case sql-lora and sql-lora2 point to the same underlying lora, but this need not be the case. lora config values take the same values they do in EngineArgs
    
    no work has been done here to scope client permissions to specific models
    jvmncs authored Feb 17, 2024
    Configuration menu
    Copy the full SHA
    8f36444 View commit details
    Browse the repository at this point in the history

Commits on Feb 18, 2024

  1. Configuration menu
    Copy the full SHA
    786b7f1 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    537c975 View commit details
    Browse the repository at this point in the history

Commits on Feb 19, 2024

  1. Configuration menu
    Copy the full SHA
    a61f052 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    ab3a5a8 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    86fd8bb View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    e433c11 View commit details
    Browse the repository at this point in the history

Commits on Feb 20, 2024

  1. Configuration menu
    Copy the full SHA
    264017a View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    63e2a64 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    181b27d View commit details
    Browse the repository at this point in the history

Commits on Feb 21, 2024

  1. Configuration menu
    Copy the full SHA
    017d9f1 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    5253eda View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    c20ecb6 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    a9c8212 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    dc903e7 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    7d2dcce View commit details
    Browse the repository at this point in the history
  7. Bump up version to v0.3.2 (vllm-project#2968)

    This version is for more model support. Add support for Gemma models (vllm-project#2964) and OLMo models (vllm-project#2832).
    zhuohan123 authored Feb 21, 2024
    Configuration menu
    Copy the full SHA
    8fbd84b View commit details
    Browse the repository at this point in the history

Commits on Feb 22, 2024

  1. Configuration menu
    Copy the full SHA
    7c4304b View commit details
    Browse the repository at this point in the history
  2. Update README.md

    mgoin authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    5344a01 View commit details
    Browse the repository at this point in the history
  3. Semi-structured 2:4 sparsity via SparseSemiStructuredTensor #4

    magic_wand semi_structured_sparse_tensor_linear branch integrates 2:4 semi-structured sparsity into SparseTensor. This PR adds a new sparsity config for 2:4 sparsity to neuralmagic-vllm, using the SparseTensor 2:4 support.
    
    This PR also refactors the sparse linear method into a separate file, vllm/model_executor/layers/sparsity/sparse_w16a16_linear_method.py, which supports all sparsity formats.
    afeldman-nm authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    81dba47 View commit details
    Browse the repository at this point in the history
  4. Sparse fused gemm integration (#12)

    Summary:
    
    Initial integration for the sparse-fused gemm. To achieve this, we need
    to ensure that we compress the weight matrix only once and never
    decompress it, as decompression is currently unsupported.
    
    Before this change, using `SparseParameter(SparseTensor)` meant that in
    `MergedColumnParallelLinear` and `QKVParallelLinear` every time a new
    shard was loaded by the `weight_loader` (e.g., the "q" portion of
    `QKVParallelLinear`), we would decompress the tensor in-order to use
    narrow to update the appropriate section of the weight tensor. With this
    change, `SparseParameter(SparseTensor)` is replaced with
    `LazyCompressedParameter`, which allows us to operate on
    `uncompressed_data` until we explicitly compress it. At that point, the
    `uncompressed_data` is compressed into `compressed_data` and freed.
    Currently, the detection of when to call compress is somewhat hacky. For
    `QKVParallelLinear`, we compress only after inserting "q", "k", and "v"
    shard ids, and for `MergedColumnParallelLinear`, we compress once we've
    inserted the same number of shards as outputs (determined by
    `len(output_sizes)`), which implicitly assumes one shard per output.
    
    Moving away from `SparseParameter(SparseTensor)` means that
    `SparseTensor` no longer handles dispatching to the custom ops; instead,
    this is handled by `SparseW16A16LinearMethod`. I believe this is a
    positive change overall. `SparseTensor` was an unnecessary extra layer
    of abstraction/indirection originally designed for the SLoRA work, not
    vLLM.
    
    This did result in the 2:4 sparse implementation breaking. However, it
    turns out it was already broken (i.e., it was decompressing and running
    dense within `SparseTensor`), so we "disable" it for now ("disable"
    meaning decompress and run dense instead).
    
    We should revisit all of this infrastructure post-MVP.
    
    ---------
    
    Co-authored-by: Andrew Feldman <[email protected]>
    2 people authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    cf8eed7 View commit details
    Browse the repository at this point in the history
  5. Abf149/fix semi structured sparse (#16)

    SUMMARY:
    - Fix bug whereby 2:4 is not being invoked
    - Eschew SparseTensor based implementation
    
    TESTING:
    - examples/offline_inference_semi_structured_sparse.py
    
    ---------
    
    Co-authored-by: Lucas Wilkinson <[email protected]>
    2 people authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    7527b9c View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    3c11f56 View commit details
    Browse the repository at this point in the history
  7. seed workflow (#19)

    SUMMARY
    * add callable seed workflow for initial boundary testing
    
    Co-authored-by: marcella-found <[email protected]>
    2 people authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    8147811 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    e802bc2 View commit details
    Browse the repository at this point in the history
  9. Use naive decompress for SM<8.0 (#32)

    A warning will be printed out if this case is triggered:
    ```
    WARNING 02-20 22:21:27 sparse_w16a16.py:32] Unstructured sparse kernels are not optimized for NVIDIA SM < 8.0. Naive decompress kernels will be used and can be slower than dense models
    ```
    
    Works on a T4 with:
    ```python
    from vllm import LLM, SamplingParams
    
    model = LLM(
        "nm-testing/opt-125m-pruned2.4", 
        sparsity="sparse_w16a16",
        enforce_eager=True,
        dtype="float16",
    )
    
    sampling_params = SamplingParams(max_tokens=100, temperature=0)
    outputs = model.generate("Hello my name is", sampling_params=sampling_params)
    outputs[0].outputs[0].text
    ```
    
    Test within colab:
    https://colab.research.google.com/drive/15xRvWX5gNaTb00BcaXhxwMm6yxavIKGN?usp=sharing
    mgoin authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    b976653 View commit details
    Browse the repository at this point in the history
  10. Varun/benchmark workflow (#28)

    Add initial bechmark workflow
    
    ---------
    
    Co-authored-by: Varun Sundar Rabindranath <[email protected]>
    2 people authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    78ba5c1 View commit details
    Browse the repository at this point in the history
  11. initial GHA workflows for "build test" and "remote push" (#27)

    SUMMARY:
    * initial set of "actions with a little a" that are the building blocks
    for eventual CI system
    * "build test" workflow
    * "remote push" workflow on `a10g`
    * update some requirement files to have packages listed in alphabetical
    order
    
    NOTE: this PR is still somewhat nebulas as i'm still working through
    building and testing "neuralmagic-vllm" in our automation environment.
    
    TEST:
    currently, i'm working through various workflow components, i.e.
    "actions with a little a". the bits making up the actions in this PR
    have been constructed from my notes along the way.
    
    we can do a "complete" run that includes: linting, building, installing,
    and running tests.
    
    GHA link ...
    https://github.com/neuralmagic/neuralmagic-vllm/actions/runs/7975058564
    `testmo` ... https://neuralmagic.testmo.net/automation/runs/view/8097
    
    Latest GHA link ...
    https://github.com/neuralmagic/neuralmagic-vllm/actions/runs/7992489982
    
    ---------
    
    Co-authored-by: andy-neuma <[email protected]>
    2 people authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    fbfd764 View commit details
    Browse the repository at this point in the history
  12. Only import magic_wand if sparsity is enabled (#37)

    Tested by making sure magic_wand was uninstalled and this code for a
    dense model runs fine:
    ```python
    from vllm import LLM, SamplingParams
    model = LLM("nm-testing/opt-125m-pruned2.4", enforce_eager=True)
    ```
    
    Then testing with a sparse model run:
    ```python
    from vllm import LLM, SamplingParams
    model = LLM("nm-testing/opt-125m-pruned2.4", sparsity="sparse_w16a16", enforce_eager=True)
    ```
    output:
    ```
    ...
      File "/home/michael/code/neuralmagic-vllm/vllm/model_executor/weight_utils.py", line 93, in get_sparse_config
        from vllm.model_executor.layers.sparsity import get_sparsity_config
      File "/home/michael/code/neuralmagic-vllm/vllm/model_executor/layers/sparsity/__init__.py", line 6, in <module>
        raise ValueError(
    ValueError: magic_wand is not available and required for sparsity support. Please install it with `pip install magic_wand`
    ```
    mgoin authored and robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    37883e0 View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    acf16bf View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    dbf3cab View commit details
    Browse the repository at this point in the history
  15. Configuration menu
    Copy the full SHA
    0feedf9 View commit details
    Browse the repository at this point in the history
  16. Configuration menu
    Copy the full SHA
    ce8164d View commit details
    Browse the repository at this point in the history
  17. format

    robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    166c13b View commit details
    Browse the repository at this point in the history
  18. Configuration menu
    Copy the full SHA
    1b395b4 View commit details
    Browse the repository at this point in the history
  19. format

    robertgshaw2-neuralmagic committed Feb 22, 2024
    Configuration menu
    Copy the full SHA
    8d935be View commit details
    Browse the repository at this point in the history
  20. Configuration menu
    Copy the full SHA
    acb8615 View commit details
    Browse the repository at this point in the history
  21. Rs/marlin downstream v0.3.2 (#43)

    Co-authored-by: Andrew Feldman <[email protected]>
    Co-authored-by: Robert Shaw <[email protected]>
    Co-authored-by: alexm <[email protected]>
    4 people authored Feb 22, 2024
    Configuration menu
    Copy the full SHA
    4b44479 View commit details
    Browse the repository at this point in the history

Commits on Feb 23, 2024

  1. additional updates to "bump-to-v0.3.2" (#39)

    SUMMARY
    * update `TORCH_CUDA_ARCH_LIST` to match `magic_wand`
    * update "test vllm" action to run tests serially
    * add helper script to find *.py tests, run them serially, and output
    JUnit formatted xml
    
    TEST
    working through changes manually on debug instance
    
    ---------
    
    Co-authored-by: andy-neuma <[email protected]>
    andy-neuma and andy-neuma authored Feb 23, 2024
    Configuration menu
    Copy the full SHA
    9209f15 View commit details
    Browse the repository at this point in the history
  2. move to 4 x gpu

    andy-neuma committed Feb 23, 2024
    Configuration menu
    Copy the full SHA
    b1e14c2 View commit details
    Browse the repository at this point in the history