{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"OpenRLHF","owner":"OpenLLMAI","isFork":false,"description":"An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)","allTopics":["reinforcement-learning","raylib","transformers","deepspeed","large-language-models","reinforcement-learning-from-human-feedback","vllm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":48,"starsCount":1534,"forksCount":132,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,5,8,51,56,15,16,33,4,13,16,5,16,21,31,15,0,19,5,6,6,15,9,15,22,34,13,17,10,3,12,9,7,17,7,0,0,1,16,2,6,15,11,3,8,2,22],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-10T06:01:53.342Z"}},{"type":"Public","name":"OpenLLMDE","owner":"OpenLLMAI","isFork":false,"description":"OpenLLMDE: An open source data engineering framework for LLMs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":9,"starsCount":16,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-09T10:45:34.411Z"}}],"repositoryCount":2,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}