PKU-Alignment/safe-rlhf
Safe-RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
Language: Python
#ai_safety #alpaca #datasets #deepspeed #large_language_models #llama #llm #llms #reinforcement_learning #reinforcement_learning_from_human_feedback #rlhf #safe_reinforcement_learning #safe_reinforcement_learning_from_human_feedback #safe_rlhf #safety #transformers #vicuna
Stars: 279 Issues: 0 Forks: 14
https://github.com/PKU-Alignment/safe-rlhf
Safe-RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
Language: Python
#ai_safety #alpaca #datasets #deepspeed #large_language_models #llama #llm #llms #reinforcement_learning #reinforcement_learning_from_human_feedback #rlhf #safe_reinforcement_learning #safe_reinforcement_learning_from_human_feedback #safe_rlhf #safety #transformers #vicuna
Stars: 279 Issues: 0 Forks: 14
https://github.com/PKU-Alignment/safe-rlhf
GitHub
GitHub - PKU-Alignment/safe-rlhf: Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback - PKU-Alignment/safe-rlhf
baichuan-inc/baichuan-7B
A large-scale 7B pretraining language model developed by Baichuan
Language: Python
#artificial_intelligence #ceval #chatgpt #gpt_4 #huggingface #large_language_models #llama #mmlu #natural_language_processing
Stars: 754 Issues: 10 Forks: 39
https://github.com/baichuan-inc/baichuan-7B
A large-scale 7B pretraining language model developed by Baichuan
Language: Python
#artificial_intelligence #ceval #chatgpt #gpt_4 #huggingface #large_language_models #llama #mmlu #natural_language_processing
Stars: 754 Issues: 10 Forks: 39
https://github.com/baichuan-inc/baichuan-7B
GitHub
GitHub - baichuan-inc/Baichuan-7B: A large-scale 7B pretraining language model developed by BaiChuan-Inc.
A large-scale 7B pretraining language model developed by BaiChuan-Inc. - baichuan-inc/Baichuan-7B