Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Zhn/fish e2e merge #11426

Draft
wants to merge 71 commits into
base: main
Choose a base branch
from
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
71 commits
Select commit Hold shift + click to select a range
4cc189a
init
niuzheng168 Jul 7, 2024
b90de40
Merge branch 'main' of https://github.com/vllm-project/vllm
niuzheng168 Jul 7, 2024
0ec7803
foward logits
Jul 8, 2024
43e9214
foward logits
Jul 9, 2024
41e1655
update
Jul 10, 2024
9c4be7b
genrate multihead output token
Jul 10, 2024
d488634
update spk_emb
Jul 11, 2024
3cad09b
fix bug
Jul 11, 2024
1bf42a2
add output hiddens
Jul 12, 2024
2918e8b
add output hiddens
Jul 12, 2024
d1f961f
add async
Jul 13, 2024
9a86ee8
add spk emb
Jul 14, 2024
c9cc8f9
Merge branch 'vllm-project:main' into main
niuzheng168 Jul 15, 2024
693c2d9
add benchmark
Jul 15, 2024
0f79d54
Merge branch 'main' of https://github.com/niuzheng168/vllm
Jul 15, 2024
90802c5
add benchmark
Jul 15, 2024
a632185
remove weight norm
Jul 16, 2024
3e127db
VocabParallelEmbedding
Jul 16, 2024
b8c94af
use base64
Jul 29, 2024
2185376
merge main
Aug 20, 2024
bd0fc4f
merge main
Aug 20, 2024
a0acc44
fix merge
Aug 20, 2024
cea25d2
fix merge main
Aug 21, 2024
43252d9
fix merge
Aug 21, 2024
0666697
Merge pull request #2 from niuzheng168/vllm-project-main
niuzheng168 Aug 21, 2024
dbf5f67
fix broken llama
Aug 30, 2024
7e98539
fix broken llama
Aug 30, 2024
eff2cf0
fix break
niuzheng168 Aug 30, 2024
350a758
seperate emb
niuzheng168 Aug 31, 2024
4b26392
fix catpure run
niuzheng168 Aug 31, 2024
df4ba99
support output hidden
niuzheng168 Sep 3, 2024
ceea94d
Enable repetition_penalties maybe work??
niuzheng168 Sep 4, 2024
74a2edd
implement multihead sampler
niuzheng168 Sep 5, 2024
76195a5
for xptts
niuzheng168 Sep 6, 2024
91fea92
fix llama
niuzheng168 Sep 6, 2024
5c9d4a7
fp32 done
niuzheng168 Sep 9, 2024
1eee089
update tokenizer
Sep 9, 2024
7636836
fix batch greater than 1
niuzheng168 Sep 9, 2024
ee811bb
fix capture bug
niuzheng168 Sep 10, 2024
a63f012
update benchmark code
niuzheng168 Sep 10, 2024
120693a
update logits for first 2 tokens
Sep 10, 2024
b37a431
undo usless changes
Sep 10, 2024
0e6701f
undo usless changes
Sep 10, 2024
c6d051d
Merge remote-tracking branch 'upstream/main' into fish_merge
niuzheng168 Sep 10, 2024
a620e33
fix merge failures for fish
niuzheng168 Sep 10, 2024
52d49d4
NOT FINISHED: num_scheduler_steps for multi head sampler
niuzheng168 Sep 10, 2024
2b53f49
another fix
niuzheng168 Sep 12, 2024
e2dbf85
non fused mlp
niuzheng168 Sep 13, 2024
64d8f54
overall good, only kernal issue now
niuzheng168 Sep 13, 2024
b79d49b
fix bug
niuzheng168 Sep 13, 2024
44fe04f
udpate
niuzheng168 Sep 13, 2024
bac660b
fix bug
niuzheng168 Sep 13, 2024
d1210e4
clean up
niuzheng168 Sep 19, 2024
c34c64f
cleanup
niuzheng168 Sep 19, 2024
d91f9c7
Merge pull request #3 from niuzheng168/fish_merge
niuzheng168 Sep 19, 2024
9017387
clean code
niuzheng168 Sep 19, 2024
a6321bd
clean code
niuzheng168 Sep 19, 2024
069b591
Merge pull request #4 from niuzheng168/fish_merge
niuzheng168 Sep 19, 2024
985f54c
Merge branch 'main' into dev
niuzheng168 Sep 19, 2024
9dd86bd
replace torch layer to vllm layers
niuzheng168 Sep 20, 2024
26a0169
optimize chattts convert code
niuzheng168 Sep 24, 2024
00255ef
optimize chattts convert code
niuzheng168 Sep 26, 2024
4ae3509
e2e streaming
Sep 27, 2024
1eadf24
update
Sep 29, 2024
06244eb
try add trt
Sep 30, 2024
d8f9a59
trt work
Sep 30, 2024
0731554
Merge remote-tracking branch 'upstream/main' into zhn/fish-e2e-merge
Dec 19, 2024
81e66fb
fish_tts_changes
Dec 19, 2024
a2c67c3
make e2e work after merge
niuzheng168 Dec 23, 2024
4d53ec8
make e2e work after merge
niuzheng168 Dec 23, 2024
4dde441
make llama work
niuzheng168 Dec 24, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 2 additions & 1 deletion benchmarks/backend_request_func.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,8 @@ class RequestFuncInput:

@dataclass
class RequestFuncOutput:
generated_text: str = ""
generated_text: str = "",
output_tokens: Union[List[int], List[List[int]]] = [],
success: bool = False
latency: float = 0.0
ttft: float = 0.0 # Time to first token
Expand Down
Loading
Loading