Skip to content
View none0663's full-sized avatar

Block or report none0663

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
none0663/README.md
  • 👋 Hi, I’m @none0663
  • 🧠 Reinforcement Learning Specialist with 6 years of hands-on experience
  • ⚡ Interests: RL algorithms design & RLHF (Reinforcement Learning from Human Feedback)
  • 🤝 Seeking collaboration: Open-source RL toolkits, human-AI alignment projects, and novel RLHF applications
  • 📫 Let's connect: [email protected]
  • 💡 Fun fact: Trained an RL agent and RLHF
  • 🌱 Always learning: Latest papers on reinforcement learning and ethical AI alignment

Popular repositories Loading

  1. none0663 none0663 Public

    Config files for my GitHub profile.

  2. PARL PARL Public

    Forked from PaddlePaddle/PARL

    A high-performance distributed training framework for Reinforcement Learning

    Python

  3. verl verl Public

    Forked from volcengine/verl

    veRL: Volcano Engine Reinforcement Learning for LLM

    Python 1

  4. OpenRLHF OpenRLHF Public

    Forked from OpenRLHF/OpenRLHF

    An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)

    Python

  5. Megatron-LM Megatron-LM Public

    Forked from NVIDIA/Megatron-LM

    Ongoing research training transformer models at scale

    Python