GitHub / AI-Hypercomputer / JetStream
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
JSON API: http://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/AI-Hypercomputer%2FJetStream
Stars: 325
Forks: 39
Open issues: 24
License: apache-2.0
Language: Python
Size: 6.39 MB
Dependencies parsed at: Pending
Created at: about 1 year ago
Updated at: 3 days ago
Pushed at: about 17 hours ago
Last synced at: about 12 hours ago
Commit Stats
Commits: 121
Authors: 21
Mean commits per author: 5.76
Development Distribution Score: 0.645
More commit stats: https://commits.ecosyste.ms/hosts/GitHub/repositories/AI-Hypercomputer/JetStream
Topics: gemma, gpt, gpu, inference, jax, large-language-models, llama, llama2, llm, llm-inference, llmops, mlops, model-serving, pytorch, tpu, transformer