-
Notifications
You must be signed in to change notification settings - Fork 4.5k
Issues: hiyouga/LLaMA-Factory
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Potential label leakage Issue: Label strings are in inputs for both training and inference.
invalid
This doesn't seem right
#6466
opened Dec 27, 2024 by
hohoCode
1 task done
LLaMA Factoryx昇腾910B
npu
This problem is related to NPU devices
pending
This problem is yet to be addressed
#6461
opened Dec 27, 2024 by
winni0
1 task done
Please upgrade transformers and deepspeed version in requirements.txt
pending
This problem is yet to be addressed
#6460
opened Dec 27, 2024 by
randydl
ollama无法加载本地模型
pending
This problem is yet to be addressed
#6459
opened Dec 27, 2024 by
lx687
1 task done
相同dpo数据,相同训练配置和训练参数,在safe-rlhf框架训练完成以后回复正常,在llama-factory训练以后模型重复输出
pending
This problem is yet to be addressed
#6458
opened Dec 27, 2024 by
Xuanwu-Gong
1 task done
[rank0]: RuntimeError: tensor does not have a device
pending
This problem is yet to be addressed
#6454
opened Dec 26, 2024 by
Juvenilecris
1 task done
华为昇腾NPU支持QLora训练吗?
npu
This problem is related to NPU devices
pending
This problem is yet to be addressed
#6452
opened Dec 26, 2024 by
sunxiaoyu12
1 task done
DeepSpeed支持yaml配置文件
pending
This problem is yet to be addressed
#6445
opened Dec 25, 2024 by
randydl
加载完模型刚刚开始训练时,显示 torch.utils.checkpoint.CheckpointError: torch.utils.checkpoint: Recomputed values for the following tensors have different metadata than during the forward pass.
pending
This problem is yet to be addressed
#6438
opened Dec 25, 2024 by
yedingda
1 task done
lora微调Mamba-Codestral-7B-v0.1出现了问题
pending
This problem is yet to be addressed
#6434
opened Dec 24, 2024 by
tongzeliang
1 task done
寒武纪:咱们是否能支持寒武纪?
pending
This problem is yet to be addressed
#6429
opened Dec 24, 2024 by
y149604146
1 task done
Ascend NPU 910B3采用deepspeed引擎训练,Q1:未调用NPU,Q2:NPU健康状态是否影响训练。
npu
This problem is related to NPU devices
pending
This problem is yet to be addressed
#6428
opened Dec 24, 2024 by
Lexlum
1 task done
奖励模型能否不是一个model,而是一个自己定义的函数
pending
This problem is yet to be addressed
#6423
opened Dec 23, 2024 by
cdhx
1 task done
Tokenizer does not derive the newer config
pending
This problem is yet to be addressed
#6415
opened Dec 21, 2024 by
xiaosu-zhu
1 task done
Questions about resuming training form ckpt
pending
This problem is yet to be addressed
#6414
opened Dec 21, 2024 by
Jiawei-Guo
1 task done
Why Speed per iteration slower when dataset is large
pending
This problem is yet to be addressed
#6410
opened Dec 20, 2024 by
coding2debug
1 task done
sft have bug while lora run successfully
pending
This problem is yet to be addressed
#6405
opened Dec 20, 2024 by
TimeFlysLeo
1 task done
How to reproduce the paper results?
pending
This problem is yet to be addressed
#6387
opened Dec 19, 2024 by
StiphyJay
1 task done
LLaMA-Factory对话预期之外存在问题
pending
This problem is yet to be addressed
#6386
opened Dec 19, 2024 by
3237522375
1 task done
如何把我训练的奖励模型放到ppo的工作管线里
pending
This problem is yet to be addressed
#6385
opened Dec 19, 2024 by
chcoo
1 task done
LLava Series (7B, 14B) freeze_vision_tower=false bug
pending
This problem is yet to be addressed
#6376
opened Dec 18, 2024 by
xirui-li
1 task done
训练PPO出现问题:ValueError: Target module ModuleDict( (default): Identity() (reward): Identity() ) is not supported. Currently, only the following modules are supported: This problem is yet to be addressed
torch.nn.Linear
, torch.nn.Embedding
, torch.nn.Conv2d
, transformers.pytorch_utils.Conv1D
.
pending
#6373
opened Dec 18, 2024 by
sunzjz
1 task done
多节点使用zero3速度很慢
pending
This problem is yet to be addressed
#6372
opened Dec 18, 2024 by
HelloWorld506
1 task done
Previous Next
ProTip!
Adding no:label will show everything without a label.