-
Notifications
You must be signed in to change notification settings - Fork 4.2k
Issues: microsoft/DeepSpeed
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[BUG] inference ops unit tests are failing
bug
Something isn't working
inference
#6839
opened Dec 9, 2024 by
oelayan7
[BUG] using deepspeed slower inference time
bug
Something isn't working
inference
#6818
opened Dec 4, 2024 by
williamlin0518
[BUG] Enabling drop_tokens in MoE layer causes inference to hang
bug
Something isn't working
inference
#6809
opened Nov 29, 2024 by
Shamauk
[BUG] deepspeed inference for llama3.1 70b for 2 node, each node with 2 gpu
bug
Something isn't working
inference
#6805
opened Nov 28, 2024 by
rastinrastinii
[BUG] Non-Deterministic Model Responses when the Input Prompt Order Changes
bug
Something isn't working
inference
#6612
opened Oct 8, 2024 by
zcakzhuu
[BUG] Excessive CPU and GPU Memory Usage with Multi-GPU Inference Using DeepSpeed
bug
Something isn't working
inference
#5793
opened Jul 23, 2024 by
gawain000000
[BUG] inference ValueError
bug
Something isn't working
inference
#5685
opened Jun 19, 2024 by
zxrneu
[BUG] 'Invalidate trace cache' with Seq2SeqTrainer+predict_with_generate+Zero3
bug
Something isn't working
inference
#5662
opened Jun 14, 2024 by
Osterlohe
[BUG] fp6 can‘t load qwen1.5-34b-chat
bug
Something isn't working
inference
#5579
opened May 29, 2024 by
pointerhacker
[BUG] The specified pointer resides on host memory and is not registered with any CUDA device.
bug
Something isn't working
inference
#5561
opened May 22, 2024 by
La1c
[Question]how to run the mixtral inference in multi-node?
bug
Something isn't working
inference
#5544
opened May 17, 2024 by
leachee99
[BUG] Mis-typed free_blocks
bug
Something isn't working
inference
#5407
opened Apr 12, 2024 by
lshamis
[BUG] inference generate raises ValueError
bug
Something isn't working
inference
#5388
opened Apr 9, 2024 by
lshamis
[BUG] MegatronMoELayerPolicy not mapped
bug
Something isn't working
inference
#5285
opened Mar 15, 2024 by
nikit-srivastava
ValueError: Only able to place 0 replicas, but 2 replicas were requested.
bug
Something isn't working
inference
#5267
opened Mar 13, 2024 by
cooper12121
[BUG] Errors when running with Microsoft Phi models
bug
Something isn't working
inference
#5200
opened Feb 27, 2024 by
IlyasMoutawwakil
[BUG] Error: Sizes of tensors must match except in dimension 1. Expected size 64 but got size 512 for tensor number 2 in the list.
bug
Something isn't working
inference
#5091
opened Feb 6, 2024 by
mikob
[BUG] RuntimeError: Error building extension 'inference_core_ops'
bug
Something isn't working
inference
#5050
opened Feb 1, 2024 by
Evanlovea
[BUG] Inferencev2 Mistral problems
bug
Something isn't working
inference
#5024
opened Jan 29, 2024 by
tatiana-iazykova
[BUG] The inaccurate flop results after several rounds for model's inference
bug
Something isn't working
inference
#4976
opened Jan 19, 2024 by
BitCalSaul
[BUG] DeepSpeed Zero3 Inference behavior error when model.train() mode
bug
Something isn't working
inference
#4922
opened Jan 9, 2024 by
liu-zichen
[BUG] DeepSpeed Zero Inference (stage 3) Stuck When One Process Doesn't Execute Something isn't working
inference
model.generate()
bug
#4910
opened Jan 6, 2024 by
samuel21119
[BUG] Deepspeed MultiGpu inference not working with Something isn't working
inference
Llama-2-13b-hf
bug
#4874
opened Dec 26, 2023 by
Rishabhg71
[BUG] Mixtral inference OOM
bug
Something isn't working
inference
#4864
opened Dec 23, 2023 by
ShayDuane
llama2-13b-tp8 long sentence input error.[BUG]
bug
Something isn't working
inference
#4855
opened Dec 21, 2023 by
zhyajie
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.