Comments (3)
There is no --use_chat_template
parameter and if using chat template, it's automatically default behavior if the model has chat template and you don't pass --prompt_type
so I don't recommend passing that.
I fixed the particular issue, thanks.
python generate.py --use_safetensors=True
--max_seq_len=8192
--base_model=lightblue/suzume-llama-3-8B-multilingual
--use_auth_token=$HUGGING_FACE_HUB_TOKEN
--add_disk_models_to_ui=False
works now
from h2ogpt.
It already has a chat template, which means it would work OOTB.
from h2ogpt.
Documents Parsing doesnt work:
Step 1: Create DB
#!/bin/sh
export HUGGING_FACE_HUB_TOKEN=XXX
export CUDA_VISIBLE_DEVICES="0,1"
docker run \
--gpus all \
--runtime=nvidia \
--shm-size=64g \
-p 7860:7860 \
--rm --init \
-v /etc/passwd:/etc/passwd:ro \
-v /etc/group:/etc/group:ro \
-u 1000:1000 \
-v /opt/h2ogpt_data/.cache:/workspace/.cache \
-v /opt/h2ogpt_data/save:/workspace/save \
-v /opt/h2ogpt_data/db_dir_MyData:/workspace/db_dir_MyData \
-v /opt/h2ogpt_data/db_dir_UserData:/workspace/db_dir_UserData \
-v /opt/h2ogpt_data/tmp:/tmp \
-v /opt/h2ogpt_data/DATA:/workspace/user_path/DATA \
-e HUGGING_FACE_HUB_TOKEN=$HUGGING_FACE_HUB_TOKEN \
-e TOKENIZERS_PARALLELISM=false \
-e ADMIN_PASS=XXX \
-e CONCURRENCY_COUNT=1 \
gcr.io/vorvan/h2oai/h2ogpt-runtime:latest /workspace/src/make_db.py --db_type=chroma
# EOF
Step 2: Run
#!/bin/sh
export HUGGING_FACE_HUB_TOKEN=XXX
export CUDA_VISIBLE_DEVICES="0,1"
docker run \
--gpus all \
--runtime=nvidia \
--shm-size=64g \
-p 7860:7860 \
--rm --init \
-v /etc/passwd:/etc/passwd:ro \
-v /etc/group:/etc/group:ro \
-u 1000:1000 \
-v /opt/h2ogpt_data/.cache:/workspace/.cache \
-v /opt/h2ogpt_data/save:/workspace/save \
-v /opt/h2ogpt_data/db_dir_UserData:/workspace/db_dir_UserData \
-v /opt/h2ogpt_data/tmp:/tmp \
-v /opt/h2ogpt_data/DATA:/workspace/user_path/DATA \
-e HUGGING_FACE_HUB_TOKEN=$HUGGING_FACE_HUB_TOKEN \
-e TOKENIZERS_PARALLELISM=false \
-e ADMIN_PASS=XXX \
-e CONCURRENCY_COUNT=1 \
-e API_OPEN=1 \
-e ALLOW_API=1 \
gcr.io/vorvan/h2oai/h2ogpt-runtime:latest /workspace/generate.py \
--use_safetensors=True \
--prompt_type=unknown \
--max_seq_len=8192 \
--use_chat_template=True \
--base_model=lightblue/suzume-llama-3-8B-multilingual \
--save_dir='/workspace/save/' \
--use_auth_token=$HUGGING_FACE_HUB_TOKEN \
--use_gpu_id=False \
--allow_upload_to_user_data=False \
--allow_upload_to_my_data=True \
--enable_ocr='off' \
--enable_pdf_ocr='off' \
--langchain_mode='UserData' \
--langchain_modes="['LLM', 'UserData', 'MyData']" \
--user_path=/workspace/user_path \
--db_type=chroma \
--visible_h2ogpt_header=False \
--visible_doc_selection_tab=False \
--visible_doc_view_tab=False \
--visible_chat_history_tab=False \
--visible_expert_tab=False \
--visible_models_tab=False \
--visible_system_tab=False \
--visible_tos_tab=False \
--visible_hosts_tab=False
# EOF
Error:
Traceback (most recent call last):
File "/h2ogpt_conda/envs/h2ogpt/lib/python3.10/site-packages/gradio/queueing.py", line 566, in process_events
response = await route_utils.call_process_api(
... ... ...
File "/workspace/src/gpt_langchain.py", line 7985, in get_chain
assert hasattr(llm, 'chat_conversation')
AssertionError
from h2ogpt.
Related Issues (20)
- No way to save prompt/response pairs in a database?
- error intalling from linux_install_full.sh HOT 5
- Failed to import transformers.pipelines HOT 6
- Intel ARC GPU support
- Document Storage HOT 2
- How should I upload my personal data to the h2o website I deployed and make it persistent? HOT 1
- Collection Selection showen multiple times HOT 1
- ValueError: load_in_8bit must be a boolean HOT 5
- Question: correct prompts template for llama3-instruct HOT 9
- httpx.ConnectError with --openai_server=True --ssl-verify=False HOT 12
- h2ogpt on ubuntu server HOT 3
- branding capitalization HOT 1
- OCR issue HOT 1
- shared / personal collections HOT 1
- Failed to initial linux full script intallation HOT 2
- random assertion errors due to evaluate_nochat HOT 13
- Run docker image on any machine which haven't internet connection HOT 19
- h2ogpt vllm-check init-container stuck when istio injection
- GPU offloading mistralai_mistral-7b-instruct-v0.2 HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from h2ogpt.