-
Notifications
You must be signed in to change notification settings - Fork 17
Comparing changes
Open a pull request
base repository: AI-Hypercomputer/jetstream-pytorch
base: jetstream-v0.2.2
head repository: AI-Hypercomputer/jetstream-pytorch
compare: jetstream-v0.2.3
- 19 commits
- 34 files changed
- 7 contributors
Commits on Jun 4, 2024
-
Enable jax profiler server in run with ray (#112)
* add jax profiler server * update jetstream
Configuration menu - View commit details
-
Copy full SHA for fe328bb - Browse repository at this point
Copy the full SHA fe328bbView commit details -
Add for readme interleave multiple host with ray (#114)
* add interleave multiple host with ray readme * add interleave multiple host with ray readme
Configuration menu - View commit details
-
Copy full SHA for f4426c2 - Browse repository at this point
Copy the full SHA f4426c2View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7f6e45f - Browse repository at this point
Copy the full SHA 7f6e45fView commit details
Commits on Jun 6, 2024
-
Integrate disaggregated serving with JetStream (#117)
* add diaggregated server with ray support * add run_server wity ray * format
Configuration menu - View commit details
-
Copy full SHA for 52ec00f - Browse repository at this point
Copy the full SHA 52ec00fView commit details
Commits on Jun 7, 2024
-
Support HF LLaMA ckpt conversion (#118)
* support converting hf checkpoint
Configuration menu - View commit details
-
Copy full SHA for 94b576c - Browse repository at this point
Copy the full SHA 94b576cView commit details -
Configuration menu - View commit details
-
Copy full SHA for e07aee6 - Browse repository at this point
Copy the full SHA e07aee6View commit details
Commits on Jun 10, 2024
-
Add support for Llama3-70b (#101)
* Add support for Llama3-70b * Fix unit tests * assert model_name is one of llama-2 or llama-3 for weight sharding * Fix lint * Revert separate shardings for llama-2 and llama-3 * Fix lint
Configuration menu - View commit details
-
Copy full SHA for 4535bdf - Browse repository at this point
Copy the full SHA 4535bdfView commit details -
Configuration menu - View commit details
-
Copy full SHA for 87b8d92 - Browse repository at this point
Copy the full SHA 87b8d92View commit details
Commits on Jun 11, 2024
-
* Initial Mixtral enablement. * Adds the mistral tokenizer model. * Updates the convert checkpoint file to handle mistral model. * Renames the typo of the model name. * Fixing checkpoing loading. Still has some issue. Push to debug. * Running on CPU working, temporarily disable the generate jit to see it's moving. But the outputs doesn't make sense yet because weights are not loaded yet. * Fix checkpoint loading issue. Right now loading from the gpt-fast converter with qkv fusion. * Fix the ckpt conversion script for mistral model. Fix the freqs_cis for loading pth file. * Add quantized layer for moe quantization * Add the huggingface download script. Improved the convert checkpoints logging. * Clean up and fix lint errors. * Missing cleanups. * Add instructions for Mixtral. * Renames everything from mistral to mixtral. * Fix more lint errors. * Removes the unnecessary checkpoint name mapping from the original Mixtral checkpoints. * Fix the model calling arg sequence; Fix the checkpoint convert script. --------- Co-authored-by: Han Qi
Configuration menu - View commit details
-
Copy full SHA for d6bf068 - Browse repository at this point
Copy the full SHA d6bf068View commit details
Commits on Jun 12, 2024
-
Configuration menu - View commit details
-
Copy full SHA for e2ee7dd - Browse repository at this point
Copy the full SHA e2ee7ddView commit details -
Add activation quantization support to per-channel quantized linear l…
…ayers (#105) * add activation quant support * pyink * fix dtype * uncomment prompts * try fix test add debug print to debug remove print, add bias to asym quant tests lint * add comment
Configuration menu - View commit details
-
Copy full SHA for 8a125b6 - Browse repository at this point
Copy the full SHA 8a125b6View commit details
Commits on Jun 13, 2024
-
Configuration menu - View commit details
-
Copy full SHA for fe8dbde - Browse repository at this point
Copy the full SHA fe8dbdeView commit details
Commits on Jun 14, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 97aaeae - Browse repository at this point
Copy the full SHA 97aaeaeView commit details -
Add lock in prefill and generate to prevent starvation (#126)
add lock for prefill and generate to prevent starvation
Configuration menu - View commit details
-
Copy full SHA for dc90aea - Browse repository at this point
Copy the full SHA dc90aeaView commit details -
Configuration menu - View commit details
-
Copy full SHA for d8d2da4 - Browse repository at this point
Copy the full SHA d8d2da4View commit details
Commits on Jun 15, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 8bffb5d - Browse repository at this point
Copy the full SHA 8bffb5dView commit details
Commits on Jun 17, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 7526a90 - Browse repository at this point
Copy the full SHA 7526a90View commit details -
Configuration menu - View commit details
-
Copy full SHA for aa90b05 - Browse repository at this point
Copy the full SHA aa90b05View commit details
Commits on Jun 19, 2024
-
Configuration menu - View commit details
-
Copy full SHA for fa1f120 - Browse repository at this point
Copy the full SHA fa1f120View commit details
This comparison is taking too long to generate.
Unfortunately it looks like we can’t render this comparison for you right now. It might be too big, or there might be something weird with your repository.
You can try running this command locally to see the comparison on your machine:
git diff jetstream-v0.2.2...jetstream-v0.2.3