rui-ye / openfedllm Goto Github PK
View Code? Open in Web Editor NEWLicense: Apache License 2.0
License: Apache License 2.0
Hello, directly run your code, the configuration has not changed, the test steps are also according to your evaluation/open_ended process, but gen_judge_mtbench.py results are all ERROR. What's the problem? Looking forward to your reply.
Run the following command:
Here are the results of running on the A40:
model_answer.zip
model_judgment.zip
Thanks for your brilliant work. I would like to do SFT with multiple GPUs. Does your framework support this feature by design or I need to make some modifications?
Thx for ur good work. Could u pls tell me how can we run the local dataset? If the prompt in my dataset is same with the vicgalle/alpaca-gpt4, but it's not the hugging face format, it's the jason format. How can we run the code?
Hello and thank you for your great work. In addition to MT-Bench, Vicuna benchmark, and Advbench, do you have test details or codes for the other datasets?
I want to perform the local training without any federated learning operations in the same settings. So how can I modify the parameters in the off-the-shelf code framework?
I have generated several checkpoint model folders after 200 rounds of training, can I continue training with this model next time? If so, how to use the trained model parameters.
Congratulations on your excellent work!
Currently, the evaluation scripts only include open-ended scripts. Could you also release the closed-ended and other test dataset evaluation scripts? Due to the different settings (such as few-shot numbers and templates) used for different datasets, the results can vary significantly. To facilitate better comparisons and support the community's ongoing work, it would be beneficial to make this part of the code open-source.
Thanks to the authors for contributing OpenFedLLM to the open source community. Currently, I finished training the model using the default run_sft.sh script, and save to model to the output file, checkpoint-XX. When I try to use the evaluation script <gen_model_answer_mt.py> and loading the checkpoint path, I'm not able to run due to some config error
(Can't load the configuration of <file_path>, If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure <file_path> is the correct path to a directory containing a config.json file).
Any help or advice?
The summation of L2 losses over named_parameters()
does not equal the L2 loss of the entire model. Please check fed_local_dpo.py and fed_local_sft.py
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.