Giter VIP home page Giter VIP logo

Comments (5)

Beomi avatar Beomi commented on June 9, 2024 1

Hello!

I noticed that you're working with the KoAlpaca-65B-LoRA repository at huggingface, which contains only the 'LoRA-finetuned' additional weights.
To load the original llama weights, you can use other codes, such as the alpaca-lora found here: https://github.com/tloen/alpaca-lora.

If you're looking to load the model and test it on your own device, please note that you'll need an A100 80G GPU or H100 GPU to load it in a single device, even when using 8-bit quantization.
I won't be discussing pipeline parallel or tensor parallel in this repository, as it isn't the right place for that.

Assuming you have the necessary GPU, you can follow these steps to load the model and try it out: https://github.com/deep-diver/Alpaca-LoRA-Serve

To install alpaca-lora-serve, just follow the instructions in the repository.
Once that's done, you can run the following command:

export BASE_URL=decapoda-research/llama-65b-hf
export FINETUNED_CKPT_URL=beomi/KoAlpaca-65B-LoRA

python app.py --base_url $BASE_URL --ft_ckpt_url $FINETUNED_CKPT_URL --port 6006

After that, you can access the chatbot-like web UI from your browser at http://localhost:6006. Enjoy and happy coding!

from koalpaca.

SoroorMa avatar SoroorMa commented on June 9, 2024

from koalpaca.

SoroorMa avatar SoroorMa commented on June 9, 2024

I just did whatever you mentioned, but I got this error:
unsupported model type. only llamastack, alpaca, flan, and baize are supported

did I miss something?
(Alpaca-LoRA as a Chatbot Service works well)

from koalpaca.

Beomi avatar Beomi commented on June 9, 2024

I utilized the alpaca-lora framework for training the LoRA model; however, I did not employ it for loading or using the model.
My experience with the LoRA checkpoint has been limited to its integration within the Chatbot service.

What are you trying to do exactly?

from koalpaca.

SoroorMa avatar SoroorMa commented on June 9, 2024

based on your guidance, I just install alpaca-lora-serve
then run the exact same command that you shared to check it on my machine
but I got that error
for now I just want to test it on my system, and maybe later try to use the model for the specific task :)

from koalpaca.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.