leehanchung / lora-instruct Goto Github PK
View Code? Open in Web Editor NEWFinetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA
License: Apache License 2.0
Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA
License: Apache License 2.0
Hi, I truncated "alpaca_data.json" to reduce training time and saved it in another file for training. Unfortunately, using this file gives me the error "pyarrow.lib.ArrowInvalid: JSON parse error: Column() changed from object to array in row 0".
Can this codebase be applicable for finetuning larger models, e.g., falcon-40b?
i 'm trying to finetune MPT-7b instruct, but getting this error.
File "/home/sulabh/new_gen_data/lora-instruct/finetune.py", line 383, in train
trainer.train(resume_from_checkpoint=resume_from_checkpoint)
File "/opt/conda/envs/lora-instruct/lib/python3.10/site-packages/transformers/trainer.py", line 1664, in train
return inner_training_loop(
File "/opt/conda/envs/lora-instruct/lib/python3.10/site-packages/transformers/trainer.py", line 1940, in _inner_training_loop
tr_loss_step = self.training_step(model, inputs)
File "/opt/conda/envs/lora-instruct/lib/python3.10/site-packages/transformers/trainer.py", line 2735, in training_step
loss = self.compute_loss(model, inputs)
File "/opt/conda/envs/lora-instruct/lib/python3.10/site-packages/transformers/trainer.py", line 2767, in compute_loss
outputs = model(**inputs)
File "/opt/conda/envs/lora-instruct/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
return forward_call(*args, **kwargs)
File "/opt/conda/envs/lora-instruct/lib/python3.10/site-packages/peft/peft_model.py", line 678, in forward
return self.base_model(
File "/opt/conda/envs/lora-instruct/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
return forward_call(*args, **kwargs)
File "/opt/conda/envs/lora-instruct/lib/python3.10/site-packages/accelerate/hooks.py", line 165, in new_forward
output = old_forward(*args, **kwargs)
TypeError: MPTForCausalLM.forward() got an unexpected keyword argument 'inputs_embeds'
Hey @leehanchung
Does lora-instruct support fine tuning mpt-30b family of models out of the box?
Hi, I'd like to know if training on Colab is possible
Hi I got a message when I try to use Lora to train MPT-B, do you have any ideas to solve it?
ValueError: Unable to create tensor, you should probably activate truncation and/or padding with 'padding=True' 'truncation=True' to have batched tensors with the same length. Perhaps your features (instruction
in this case) have excessive nesting (inputs type list
where type int
is expected).
Thanks for the wonderful code. While training Falcon-7B on Alpaca dataset, the training loss is not reducing. It used to work fine. Were there any recent changes?
Hello. Is there support for quantized lora finetuning of those llms?
Thanks
Thanks for the wonderful training code. Please release the inference code as well.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.