-
Notifications
You must be signed in to change notification settings - Fork 10
Commits on Jan 26, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 6b7de1a - Browse repository at this point
Copy the full SHA 6b7de1aView commit details -
Support for Stable LM 2 (vllm-project#2598)
Co-authored-by: Zhuohan Li <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 3a0e1fc - Browse repository at this point
Copy the full SHA 3a0e1fcView commit details -
Configuration menu - View commit details
-
Copy full SHA for 390b495 - Browse repository at this point
Copy the full SHA 390b495View commit details
Commits on Jan 27, 2024
-
Configuration menu - View commit details
-
Copy full SHA for beb89f6 - Browse repository at this point
Copy the full SHA beb89f6View commit details -
Configuration menu - View commit details
-
Copy full SHA for 220a476 - Browse repository at this point
Copy the full SHA 220a476View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3801700 - Browse repository at this point
Copy the full SHA 3801700View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5f036d2 - Browse repository at this point
Copy the full SHA 5f036d2View commit details
Commits on Jan 28, 2024
-
Configuration menu - View commit details
-
Copy full SHA for f8ecb84 - Browse repository at this point
Copy the full SHA f8ecb84View commit details -
Configuration menu - View commit details
-
Copy full SHA for 89be30f - Browse repository at this point
Copy the full SHA 89be30fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7d64841 - Browse repository at this point
Copy the full SHA 7d64841View commit details
Commits on Jan 29, 2024
-
Support FP8-E5M2 KV Cache (vllm-project#2279)
Co-authored-by: zhaoyang <[email protected]> Co-authored-by: Zhuohan Li <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9090bf0 - Browse repository at this point
Copy the full SHA 9090bf0View commit details -
Fix error when tp > 1 (vllm-project#2644)
Co-authored-by: zhaoyang-star <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for b72af8f - Browse repository at this point
Copy the full SHA b72af8fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 1b20639 - Browse repository at this point
Copy the full SHA 1b20639View commit details -
Configuration menu - View commit details
-
Copy full SHA for ea8489f - Browse repository at this point
Copy the full SHA ea8489fView commit details
Commits on Jan 30, 2024
-
DeepseekMoE support with Fused MoE kernel (vllm-project#2453)
Co-authored-by: roy <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 5d60def - Browse repository at this point
Copy the full SHA 5d60defView commit details -
Fused MOE for Mixtral (vllm-project#2542)
Co-authored-by: chen shen <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for ab40644 - Browse repository at this point
Copy the full SHA ab40644View commit details -
Fix 'Actor methods cannot be called directly' when using `--engine-us…
…e-ray` (vllm-project#2664) * fix: engine-useray complain * fix: typo
Configuration menu - View commit details
-
Copy full SHA for d79ced3 - Browse repository at this point
Copy the full SHA d79ced3View commit details -
Configuration menu - View commit details
-
Copy full SHA for 4f65af0 - Browse repository at this point
Copy the full SHA 4f65af0View commit details -
Configuration menu - View commit details
-
Copy full SHA for bbe9bd9 - Browse repository at this point
Copy the full SHA bbe9bd9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 105a40f - Browse repository at this point
Copy the full SHA 105a40fView commit details
Commits on Jan 31, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 3dad944 - Browse repository at this point
Copy the full SHA 3dad944View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1af090b - Browse repository at this point
Copy the full SHA 1af090bView commit details -
Fixes assertion failure in prefix caching: the lora index mapping sho…
…uld respect prefix_len (vllm-project#2688) Signed-off-by: Tao He <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for d69ff0c - Browse repository at this point
Copy the full SHA d69ff0cView commit details -
Configuration menu - View commit details
-
Copy full SHA for c664b0e - Browse repository at this point
Copy the full SHA c664b0eView commit details -
Configuration menu - View commit details
-
Copy full SHA for 89efcf1 - Browse repository at this point
Copy the full SHA 89efcf1View commit details -
Configuration menu - View commit details
-
Copy full SHA for d0d93b9 - Browse repository at this point
Copy the full SHA d0d93b9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 93b38be - Browse repository at this point
Copy the full SHA 93b38beView commit details
Commits on Feb 1, 2024
-
Configuration menu - View commit details
-
Copy full SHA for cd9e60c - Browse repository at this point
Copy the full SHA cd9e60cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 923797f - Browse repository at this point
Copy the full SHA 923797fView commit details -
Configuration menu - View commit details
-
Copy full SHA for b9e96b1 - Browse repository at this point
Copy the full SHA b9e96b1View commit details -
Configuration menu - View commit details
-
Copy full SHA for bb8c697 - Browse repository at this point
Copy the full SHA bb8c697View commit details -
Use revision when downloading the quantization config file (vllm-proj…
…ect#2697) Co-authored-by: Pernekhan Utemuratov <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for c410f5d - Browse repository at this point
Copy the full SHA c410f5dView commit details -
Remove hardcoded
device="cuda"
to support more devices (vllm-proje……ct#2503) Co-authored-by: Jiang Li <[email protected]> Co-authored-by: Kunshang Ji <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 96b6f47 - Browse repository at this point
Copy the full SHA 96b6f47View commit details -
Configuration menu - View commit details
-
Copy full SHA for 0e163fc - Browse repository at this point
Copy the full SHA 0e163fcView commit details
Commits on Feb 2, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 4abf633 - Browse repository at this point
Copy the full SHA 4abf633View commit details
Commits on Feb 4, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 5ed704e - Browse repository at this point
Copy the full SHA 5ed704eView commit details -
set&get llm internal tokenizer instead of the TokenizerGroup (vllm-pr…
…oject#2741) Co-authored-by: shujunhua1 <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 51cd22c - Browse repository at this point
Copy the full SHA 51cd22cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5a6c81b - Browse repository at this point
Copy the full SHA 5a6c81bView commit details
Commits on Feb 5, 2024
-
Require triton >= 2.1.0 (vllm-project#2746)
Co-authored-by: yangrui1 <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for c9b45ad - Browse repository at this point
Copy the full SHA c9b45adView commit details -
Configuration menu - View commit details
-
Copy full SHA for 72d3a30 - Browse repository at this point
Copy the full SHA 72d3a30View commit details -
Configuration menu - View commit details
-
Copy full SHA for 56f738a - Browse repository at this point
Copy the full SHA 56f738aView commit details -
Configuration menu - View commit details
-
Copy full SHA for b92adec - Browse repository at this point
Copy the full SHA b92adecView commit details -
Configuration menu - View commit details
-
Copy full SHA for 2ccee3d - Browse repository at this point
Copy the full SHA 2ccee3dView commit details
Commits on Feb 6, 2024
-
Configuration menu - View commit details
-
Copy full SHA for f0d4e14 - Browse repository at this point
Copy the full SHA f0d4e14View commit details -
modelscope: fix issue when model parameter is not a model id but path…
… of the model. (vllm-project#2489)
Configuration menu - View commit details
-
Copy full SHA for ed70c70 - Browse repository at this point
Copy the full SHA ed70c70View commit details -
Configuration menu - View commit details
-
Copy full SHA for fe6d09a - Browse repository at this point
Copy the full SHA fe6d09aView commit details
Commits on Feb 7, 2024
-
[ROCm] Fix build problem resulted from previous commit related to FP8…
… kv-cache support (vllm-project#2790)
Configuration menu - View commit details
-
Copy full SHA for c81dddb - Browse repository at this point
Copy the full SHA c81dddbView commit details -
Configuration menu - View commit details
-
Copy full SHA for 931746b - Browse repository at this point
Copy the full SHA 931746bView commit details
Commits on Feb 8, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 65b89d1 - Browse repository at this point
Copy the full SHA 65b89d1View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3711811 - Browse repository at this point
Copy the full SHA 3711811View commit details
Commits on Feb 11, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 0580aab - Browse repository at this point
Copy the full SHA 0580aabView commit details
Commits on Feb 12, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 4ca2c35 - Browse repository at this point
Copy the full SHA 4ca2c35View commit details -
Refactor 2 awq gemm kernels into m16nXk32 (vllm-project#2723)
Co-authored-by: Chunan Zeng <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 5638364 - Browse repository at this point
Copy the full SHA 5638364View commit details
Commits on Feb 13, 2024
-
Configuration menu - View commit details
-
Copy full SHA for a4211a4 - Browse repository at this point
Copy the full SHA a4211a4View commit details -
Configuration menu - View commit details
-
Copy full SHA for f964493 - Browse repository at this point
Copy the full SHA f964493View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5c976a7 - Browse repository at this point
Copy the full SHA 5c976a7View commit details -
Configuration menu - View commit details
-
Copy full SHA for ea35600 - Browse repository at this point
Copy the full SHA ea35600View commit details -
Configuration menu - View commit details
-
Copy full SHA for a463c33 - Browse repository at this point
Copy the full SHA a463c33View commit details -
Remove Yi model definition, please use
LlamaForCausalLM
instead (vl……lm-project#2854) Co-authored-by: Roy <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 317b29d - Browse repository at this point
Copy the full SHA 317b29dView commit details -
Add LoRA support for Mixtral (vllm-project#2831)
* add mixtral lora support * formatting * fix incorrectly ported logic * polish tests * minor fixes and refactoring * minor fixes * formatting * rename and remove redundant logic * refactoring * refactoring * minor fix * minor refactoring * fix code smell
Configuration menu - View commit details
-
Copy full SHA for 2a543d6 - Browse repository at this point
Copy the full SHA 2a543d6View commit details
Commits on Feb 14, 2024
-
Migrate InternLMForCausalLM to LlamaForCausalLM (vllm-project#2860)
Co-authored-by: Roy <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 7eacffd - Browse repository at this point
Copy the full SHA 7eacffdView commit details -
Configuration menu - View commit details
-
Copy full SHA for 0c48b37 - Browse repository at this point
Copy the full SHA 0c48b37View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7e45107 - Browse repository at this point
Copy the full SHA 7e45107View commit details -
Configuration menu - View commit details
-
Copy full SHA for 87069cc - Browse repository at this point
Copy the full SHA 87069ccView commit details -
Configuration menu - View commit details
-
Copy full SHA for 4efbac6 - Browse repository at this point
Copy the full SHA 4efbac6View commit details -
Configuration menu - View commit details
-
Copy full SHA for 25e86b6 - Browse repository at this point
Copy the full SHA 25e86b6View commit details
Commits on Feb 15, 2024
-
Align LoRA code between Mistral and Mixtral (fixes vllm-project#2875) (…
…vllm-project#2880) * Fix AttributeError: MixtralModel object has no attribute org_vocab_size. * Make LoRA logic for Mistral and Mixtral the same --------- Co-authored-by: Pernekhan Utemuratov <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 31348df - Browse repository at this point
Copy the full SHA 31348dfView commit details -
Configuration menu - View commit details
-
Copy full SHA for d7afab6 - Browse repository at this point
Copy the full SHA d7afab6View commit details -
Configuration menu - View commit details
-
Copy full SHA for 4f2ad11 - Browse repository at this point
Copy the full SHA 4f2ad11View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5255d99 - Browse repository at this point
Copy the full SHA 5255d99View commit details
Commits on Feb 16, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 64da65b - Browse repository at this point
Copy the full SHA 64da65bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5f08050 - Browse repository at this point
Copy the full SHA 5f08050View commit details
Commits on Feb 17, 2024
-
Defensively copy
sampling_params
(vllm-project#2881)If the SamplingParams object passed to LLMEngine.add_request() is mutated after it returns, it could affect the async sampling process for that request. Suggested by @Yard1 vllm-project#2514 (comment)
Configuration menu - View commit details
-
Copy full SHA for 185b2c2 - Browse repository at this point
Copy the full SHA 185b2c2View commit details -
multi-LoRA as extra models in OpenAI server (vllm-project#2775)
how to serve the loras (mimicking the [multilora inference example](https://github.com/vllm-project/vllm/blob/main/examples/multilora_inference.py)): ```terminal $ export LORA_PATH=~/.cache/huggingface/hub/models--yard1--llama-2-7b-sql-lora-test/ $ python -m vllm.entrypoints.api_server \ --model meta-llama/Llama-2-7b-hf \ --enable-lora \ --lora-modules sql-lora=$LORA_PATH sql-lora2=$LORA_PATH ``` the above server will list 3 separate values if the user queries `/models`: one for the base served model, and one each for the specified lora modules. in this case sql-lora and sql-lora2 point to the same underlying lora, but this need not be the case. lora config values take the same values they do in EngineArgs no work has been done here to scope client permissions to specific models
Configuration menu - View commit details
-
Copy full SHA for 8f36444 - Browse repository at this point
Copy the full SHA 8f36444View commit details
Commits on Feb 18, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 786b7f1 - Browse repository at this point
Copy the full SHA 786b7f1View commit details -
Configuration menu - View commit details
-
Copy full SHA for 537c975 - Browse repository at this point
Copy the full SHA 537c975View commit details
Commits on Feb 19, 2024
-
Configuration menu - View commit details
-
Copy full SHA for a61f052 - Browse repository at this point
Copy the full SHA a61f052View commit details -
Configuration menu - View commit details
-
Copy full SHA for ab3a5a8 - Browse repository at this point
Copy the full SHA ab3a5a8View commit details -
Configuration menu - View commit details
-
Copy full SHA for 86fd8bb - Browse repository at this point
Copy the full SHA 86fd8bbView commit details -
Configuration menu - View commit details
-
Copy full SHA for e433c11 - Browse repository at this point
Copy the full SHA e433c11View commit details
Commits on Feb 20, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 264017a - Browse repository at this point
Copy the full SHA 264017aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 63e2a64 - Browse repository at this point
Copy the full SHA 63e2a64View commit details -
Configuration menu - View commit details
-
Copy full SHA for 181b27d - Browse repository at this point
Copy the full SHA 181b27dView commit details
Commits on Feb 21, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 017d9f1 - Browse repository at this point
Copy the full SHA 017d9f1View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5253eda - Browse repository at this point
Copy the full SHA 5253edaView commit details -
Configuration menu - View commit details
-
Copy full SHA for c20ecb6 - Browse repository at this point
Copy the full SHA c20ecb6View commit details -
Configuration menu - View commit details
-
Copy full SHA for a9c8212 - Browse repository at this point
Copy the full SHA a9c8212View commit details -
Configuration menu - View commit details
-
Copy full SHA for dc903e7 - Browse repository at this point
Copy the full SHA dc903e7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7d2dcce - Browse repository at this point
Copy the full SHA 7d2dcceView commit details -
Bump up version to v0.3.2 (vllm-project#2968)
This version is for more model support. Add support for Gemma models (vllm-project#2964) and OLMo models (vllm-project#2832).
Configuration menu - View commit details
-
Copy full SHA for 8fbd84b - Browse repository at this point
Copy the full SHA 8fbd84bView commit details
Commits on Feb 22, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 7c4304b - Browse repository at this point
Copy the full SHA 7c4304bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5344a01 - Browse repository at this point
Copy the full SHA 5344a01View commit details -
Semi-structured 2:4 sparsity via SparseSemiStructuredTensor #4
magic_wand semi_structured_sparse_tensor_linear branch integrates 2:4 semi-structured sparsity into SparseTensor. This PR adds a new sparsity config for 2:4 sparsity to neuralmagic-vllm, using the SparseTensor 2:4 support. This PR also refactors the sparse linear method into a separate file, vllm/model_executor/layers/sparsity/sparse_w16a16_linear_method.py, which supports all sparsity formats.
Configuration menu - View commit details
-
Copy full SHA for 81dba47 - Browse repository at this point
Copy the full SHA 81dba47View commit details -
Sparse fused gemm integration (#12)
Summary: Initial integration for the sparse-fused gemm. To achieve this, we need to ensure that we compress the weight matrix only once and never decompress it, as decompression is currently unsupported. Before this change, using `SparseParameter(SparseTensor)` meant that in `MergedColumnParallelLinear` and `QKVParallelLinear` every time a new shard was loaded by the `weight_loader` (e.g., the "q" portion of `QKVParallelLinear`), we would decompress the tensor in-order to use narrow to update the appropriate section of the weight tensor. With this change, `SparseParameter(SparseTensor)` is replaced with `LazyCompressedParameter`, which allows us to operate on `uncompressed_data` until we explicitly compress it. At that point, the `uncompressed_data` is compressed into `compressed_data` and freed. Currently, the detection of when to call compress is somewhat hacky. For `QKVParallelLinear`, we compress only after inserting "q", "k", and "v" shard ids, and for `MergedColumnParallelLinear`, we compress once we've inserted the same number of shards as outputs (determined by `len(output_sizes)`), which implicitly assumes one shard per output. Moving away from `SparseParameter(SparseTensor)` means that `SparseTensor` no longer handles dispatching to the custom ops; instead, this is handled by `SparseW16A16LinearMethod`. I believe this is a positive change overall. `SparseTensor` was an unnecessary extra layer of abstraction/indirection originally designed for the SLoRA work, not vLLM. This did result in the 2:4 sparse implementation breaking. However, it turns out it was already broken (i.e., it was decompressing and running dense within `SparseTensor`), so we "disable" it for now ("disable" meaning decompress and run dense instead). We should revisit all of this infrastructure post-MVP. --------- Co-authored-by: Andrew Feldman <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for cf8eed7 - Browse repository at this point
Copy the full SHA cf8eed7View commit details -
Abf149/fix semi structured sparse (#16)
SUMMARY: - Fix bug whereby 2:4 is not being invoked - Eschew SparseTensor based implementation TESTING: - examples/offline_inference_semi_structured_sparse.py --------- Co-authored-by: Lucas Wilkinson <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 7527b9c - Browse repository at this point
Copy the full SHA 7527b9cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 3c11f56 - Browse repository at this point
Copy the full SHA 3c11f56View commit details -
SUMMARY * add callable seed workflow for initial boundary testing Co-authored-by: marcella-found <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 8147811 - Browse repository at this point
Copy the full SHA 8147811View commit details -
Configuration menu - View commit details
-
Copy full SHA for e802bc2 - Browse repository at this point
Copy the full SHA e802bc2View commit details -
Use naive decompress for SM<8.0 (#32)
A warning will be printed out if this case is triggered: ``` WARNING 02-20 22:21:27 sparse_w16a16.py:32] Unstructured sparse kernels are not optimized for NVIDIA SM < 8.0. Naive decompress kernels will be used and can be slower than dense models ``` Works on a T4 with: ```python from vllm import LLM, SamplingParams model = LLM( "nm-testing/opt-125m-pruned2.4", sparsity="sparse_w16a16", enforce_eager=True, dtype="float16", ) sampling_params = SamplingParams(max_tokens=100, temperature=0) outputs = model.generate("Hello my name is", sampling_params=sampling_params) outputs[0].outputs[0].text ``` Test within colab: https://colab.research.google.com/drive/15xRvWX5gNaTb00BcaXhxwMm6yxavIKGN?usp=sharing
Configuration menu - View commit details
-
Copy full SHA for b976653 - Browse repository at this point
Copy the full SHA b976653View commit details -
Varun/benchmark workflow (#28)
Add initial bechmark workflow --------- Co-authored-by: Varun Sundar Rabindranath <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 78ba5c1 - Browse repository at this point
Copy the full SHA 78ba5c1View commit details -
initial GHA workflows for "build test" and "remote push" (#27)
SUMMARY: * initial set of "actions with a little a" that are the building blocks for eventual CI system * "build test" workflow * "remote push" workflow on `a10g` * update some requirement files to have packages listed in alphabetical order NOTE: this PR is still somewhat nebulas as i'm still working through building and testing "neuralmagic-vllm" in our automation environment. TEST: currently, i'm working through various workflow components, i.e. "actions with a little a". the bits making up the actions in this PR have been constructed from my notes along the way. we can do a "complete" run that includes: linting, building, installing, and running tests. GHA link ... https://github.com/neuralmagic/neuralmagic-vllm/actions/runs/7975058564 `testmo` ... https://neuralmagic.testmo.net/automation/runs/view/8097 Latest GHA link ... https://github.com/neuralmagic/neuralmagic-vllm/actions/runs/7992489982 --------- Co-authored-by: andy-neuma <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for fbfd764 - Browse repository at this point
Copy the full SHA fbfd764View commit details -
Only import magic_wand if sparsity is enabled (#37)
Tested by making sure magic_wand was uninstalled and this code for a dense model runs fine: ```python from vllm import LLM, SamplingParams model = LLM("nm-testing/opt-125m-pruned2.4", enforce_eager=True) ``` Then testing with a sparse model run: ```python from vllm import LLM, SamplingParams model = LLM("nm-testing/opt-125m-pruned2.4", sparsity="sparse_w16a16", enforce_eager=True) ``` output: ``` ... File "/home/michael/code/neuralmagic-vllm/vllm/model_executor/weight_utils.py", line 93, in get_sparse_config from vllm.model_executor.layers.sparsity import get_sparsity_config File "/home/michael/code/neuralmagic-vllm/vllm/model_executor/layers/sparsity/__init__.py", line 6, in <module> raise ValueError( ValueError: magic_wand is not available and required for sparsity support. Please install it with `pip install magic_wand` ```
Configuration menu - View commit details
-
Copy full SHA for 37883e0 - Browse repository at this point
Copy the full SHA 37883e0View commit details -
Configuration menu - View commit details
-
Copy full SHA for acf16bf - Browse repository at this point
Copy the full SHA acf16bfView commit details -
Configuration menu - View commit details
-
Copy full SHA for dbf3cab - Browse repository at this point
Copy the full SHA dbf3cabView commit details -
Configuration menu - View commit details
-
Copy full SHA for 0feedf9 - Browse repository at this point
Copy the full SHA 0feedf9View commit details -
Configuration menu - View commit details
-
Copy full SHA for ce8164d - Browse repository at this point
Copy the full SHA ce8164dView commit details -
Configuration menu - View commit details
-
Copy full SHA for 166c13b - Browse repository at this point
Copy the full SHA 166c13bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 1b395b4 - Browse repository at this point
Copy the full SHA 1b395b4View commit details -
Configuration menu - View commit details
-
Copy full SHA for 8d935be - Browse repository at this point
Copy the full SHA 8d935beView commit details -
Configuration menu - View commit details
-
Copy full SHA for acb8615 - Browse repository at this point
Copy the full SHA acb8615View commit details -
Rs/marlin downstream v0.3.2 (#43)
Co-authored-by: Andrew Feldman <[email protected]> Co-authored-by: Robert Shaw <[email protected]> Co-authored-by: alexm <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 4b44479 - Browse repository at this point
Copy the full SHA 4b44479View commit details
Commits on Feb 23, 2024
-
additional updates to "bump-to-v0.3.2" (#39)
SUMMARY * update `TORCH_CUDA_ARCH_LIST` to match `magic_wand` * update "test vllm" action to run tests serially * add helper script to find *.py tests, run them serially, and output JUnit formatted xml TEST working through changes manually on debug instance --------- Co-authored-by: andy-neuma <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9209f15 - Browse repository at this point
Copy the full SHA 9209f15View commit details -
andy-neuma committed
Feb 23, 2024 Configuration menu - View commit details
-
Copy full SHA for b1e14c2 - Browse repository at this point
Copy the full SHA b1e14c2View commit details