Code Monkey home page Code Monkey logo

huatuo-llama-med-chinese's People

Contributors

dyr1 avatar flowolfzzz avatar s65b40 avatar scir-hi avatar sharpbai avatar thinksoso avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

huatuo-llama-med-chinese's Issues

代码的问题

在多轮对话infei时会出现大量的#号,这是为什么

: 小张最近感觉身体不适,出现心悸、气促等症状。体检发现心脏扩大、搏动减弱。 : 我建议您及家人及您的医生进行进一步的检查,以确定小张的病因。同时,我建让您及家人做好的饮食和生活习惯,以避免其他的健康问题。###########################################################################################################################: 小王被确诊为肝炎双重感染,最可能的并发症是什么? : 根据相关文献,肝炎双重感染的危险因素主要包括肝炎、营养不良、免疫抑制剂、病毒敏感免疫等。因此,我建让您及家人及您的医生进行进一步的检查,以确定小张的病因。

询问一下,运行infer.py,报错TypeError: 'NoneType' object is not subscriptable

def main(
load_8bit: bool = True,
base_model: str = "decapoda-research/llama-7b-hf",
# the infer data, if not exists, infer the default instructions in code
instruct_dir: str = "./data/infer.json",
use_lora: bool = True,
lora_weights: str = "tloen/alpaca-lora-7b",
# The prompt template to use, will default to med_template.
prompt_template: str = "med_template",
):我的配置参数是这样的,其他的都和初始的一样,但我运行之后报错
│ │
│ /usr/local/lib/python3.9/dist-packages/bitsandbytes/autograd/_functions.py:3 │
│ 80 in forward │
│ │
│ 377 │ │ │ if state.CxB is not None: │
│ 378 │ │ │ │ outliers = F.extract_outliers(state.CxB, state.SB, sta │
│ 379 │ │ │ else: │
│ ❱ 380 │ │ │ │ outliers = state.CB[:, state.idx.long()].clone() │
│ 381 │ │ │ │
│ 382 │ │ │ state.subB = (outliers * state.SCB.view(-1, 1) / 127.0).t( │
│ 383 │ │ │ CA[:, state.idx.long()] = 0 │
╰──────────────────────────────────────────────────────────────────────────────╯
TypeError: 'NoneType' object is not subscriptable

Process finished with exit code 1
我不知道为什么会这样,有人可以给我指导一下吗

lora alpaca med

使用lora-alpaca-med进行推理的时候,prompt_template应该选择什么呢, med-template.json还是alpaca_short.json?

如何复现基于Chinese-alpaca-7b的医学知识和医学文献对话模型?

在本地测了Chinese-alpaca-7b + lora-alpaca-med-alldata,在多轮对话上,效果很好。

请问:

  1. 如何利用llama.json以及liver_cancer.json,如何合并数据集(一个多轮对话,一个单轮对话),直接合并放在一个文件么?
  2. 复现您们的训练效果,该使用哪个模板?('med_template' or 'literature_template.json')
  3. 目前我有 8 X A100(40G), micro_batch_size和batch_size该如何设置(都是64么?为您们设置的一半?)
  4. 关于LoRA的rank和alpha值?听说在垂直领域微调时,加大LoRA的rank值效果会好些,想问下您们关于LoRA的rank值以及alpha值的选取的相关经验

感谢您们百忙之中抽空解答,祝项目越来越好,感谢!

运行报错:json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)

/home/halo_op/anaconda3/envs/py39/lib/python3.9/site-packages/bitsandbytes/cuda_setup/main.py:145: UserWarning: Found duplicate ['libcudart.so', 'libcudart.so.11.0', 'libcudart.so.12.0'] files: {PosixPath('/usr/local/cuda/lib64/libcudart.so.11.0'), PosixPath('/usr/local/cuda/lib64/libcudart.so')}.. We'll flip a coin and try one of these, in order to fail forward.
Either way, this might cause trouble in the future:
If you get CUDA error: invalid device function errors, the above might be the cause and the solution is to make sure only one ['libcudart.so', 'libcudart.so.11.0', 'libcudart.so.12.0'] in the paths that we search based on your env.
warn(msg)
CUDA SETUP: CUDA runtime path found: /usr/local/cuda/lib64/libcudart.so.11.0
CUDA SETUP: Highest compute capability among GPUs detected: 7.5
CUDA SETUP: Detected CUDA version 114
CUDA SETUP: Loading binary /home/halo_op/anaconda3/envs/py39/lib/python3.9/site-packages/bitsandbytes/libbitsandbytes_cuda114.so...
The tokenizer class you load from this checkpoint is not the same type as the class this function is called from. It may result in unexpected tokenization.
The tokenizer class you load from this checkpoint is 'LLaMATokenizer'.
The class this function is called from is 'LlamaTokenizer'.
Loading checkpoint shards: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 33/33 [00:14<00:00, 2.29it/s]
using lora ./lora-llama-med
Traceback (most recent call last):
File "/mnt/datadisk0/code/Huatuo-Llama-Med-Chinese/infer.py", line 125, in
fire.Fire(main)
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/site-packages/fire/core.py", line 141, in Fire
component_trace = _Fire(component, args, parsed_flag_args, context, name)
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/site-packages/fire/core.py", line 475, in _Fire
component, remaining_args = _CallAndUpdateTrace(
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/site-packages/fire/core.py", line 691, in _CallAndUpdateTrace
component = fn(*varargs, **kwargs)
File "/mnt/datadisk0/code/Huatuo-Llama-Med-Chinese/infer.py", line 47, in main
model = PeftModel.from_pretrained(
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/site-packages/peft/peft_model.py", line 156, in from_pretrained
PeftConfig.from_pretrained(model_id, subfolder=kwargs.get("subfolder", None)).peft_type
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/site-packages/peft/utils/config.py", line 109, in from_pretrained
loaded_attributes = cls.from_json_file(config_file)
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/site-packages/peft/utils/config.py", line 129, in from_json_file
json_object = json.load(file)
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/json/init.py", line 293, in load
return loads(fp.read(),
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/json/init.py", line 346, in loads
return _default_decoder.decode(s)
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/home/halo_op/anaconda3/envs/py39/lib/python3.9/json/decoder.py", line 355, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
Uploading 捕获.PNG…

请问是否以经典中医医书为主?

包括但不限于易经,针灸大成,神农本草经,黄帝内经,黄帝外经,伤寒论,金匮要略?是否以经方为主,是否训练了名医书籍和医案?

4 张 3090 上运行 finetune,运行结束报错 UnboundLocalError: local variable 'load_result' referenced before assignment

  1. 期望复现 llama lora 使用文中提到的语料库训练;
  2. 只修改 finetune.sh 中对应的 base model 路径,其他都未做修改;
  3. 运行完之后,命令行报错:
he intermediate checkpoints of PEFT may not be saved correctly, using `TrainerCallback` to save adapter_model.bin in corresponding folders, here are some examples https://github.com/huggingface/peft/issues/96
Traceback (most recent call last):
  File "/home/m1l03053/llama/Huatuo-Llama-Med-Chinese/finetune.py", line 280, in <module>
    fire.Fire(train)
  File "/home/m1l03053/.conda/envs/3.9/lib/python3.9/site-packages/fire/core.py", line 141, in Fire
    component_trace = _Fire(component, args, parsed_flag_args, context, name)
  File "/home/m1l03053/.conda/envs/3.9/lib/python3.9/site-packages/fire/core.py", line 475, in _Fire
    component, remaining_args = _CallAndUpdateTrace(
  File "/home/m1l03053/.conda/envs/3.9/lib/python3.9/site-packages/fire/core.py", line 691, in _CallAndUpdateTrace
    component = fn(*varargs, **kwargs)
  File "/home/m1l03053/llama/Huatuo-Llama-Med-Chinese/finetune.py", line 270, in train
    trainer.train(resume_from_checkpoint=resume_from_checkpoint)
  File "/home/m1l03053/.conda/envs/3.9/lib/python3.9/site-packages/transformers/trainer.py", line 1696, in train
    return inner_training_loop(
  File "/home/m1l03053/.conda/envs/3.9/lib/python3.9/site-packages/transformers/trainer.py", line 2095, in _inner_training_loop
    self._load_best_model()
  File "/home/m1l03053/.conda/envs/3.9/lib/python3.9/site-packages/transformers/trainer.py", line 2292, in _load_best_model
    self._issue_warnings_after_load(load_result)
UnboundLocalError: local variable 'load_result' referenced before assignment
wandb: Waiting for W&B process to finish... (failed 1). Press Control-C to abort syncing.
wandb: 
wandb: Run history:
wandb:               eval/loss █▅▄▃▃▂▂▂▂▂▁▁▁▁▁▁▁▁▁▁
wandb:            eval/runtime ▁▃▄▆▇▅▂▆▅██▃▅█▆█▆█▅▃
wandb: eval/samples_per_second █▆▅▃▂▄▇▃▄▁▁▆▃▁▃▁▃▁▄▆
wandb:   eval/steps_per_second █▆▅▃▂▄▇▃▄▁▁▆▄▁▂▁▃▁▄▆
wandb:             train/epoch ▁▁▁▂▂▂▂▂▂▃▃▃▃▃▃▄▄▄▄▄▅▅▅▅▅▆▆▆▆▆▆▇▇▇▇▇▇███
wandb:       train/global_step 

eval_loss为nan是什么原因

image
如上图
训练过程中eval_loss为nan,参数我只修改了 batch,别的都保持一致
请问这是什么原因?

训练后的数据不理想

这是我改过的infer-literature-single.sh脚本内容
image

这是我调用bash ./scripts/finetune.sh执行的finetune脚本
image

这是结果
image

CUDA Setup failed despite GPU being available.

===================================BUG REPORT===================================
Welcome to bitsandbytes. For bug reports, please run

python -m bitsandbytes

and submit this information together with your error trace to: https://github.com/TimDettmers/bitsandbytes/issues

bin C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\libbitsandbytes_cpu.so
CUDA_SETUP: WARNING! libcudart.so not found in any environmental path. Searching in backup paths...
C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: WARNING: The following directories listed in your path were found to be non-existent: {WindowsPath('/usr/local/cuda/lib64')}
warn(msg)
CUDA SETUP: WARNING! libcuda.so not found! Do you have a CUDA driver installed? If you are on a cluster, make sure you are on a CUDA machine!
C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: WARNING: No libcudart.so found! Install CUDA or the cudatoolkit package (anaconda)!
warn(msg)
C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: WARNING: No GPU detected! Check your CUDA paths. Proceeding to load CPU-only library...
warn(msg)
CUDA SETUP: Loading binary C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\libbitsandbytes_cpu.so...
argument of type 'WindowsPath' is not iterable
CUDA SETUP: Problem: The main issue seems to be that the main CUDA library was not detected.
CUDA SETUP: Solution 1): Your paths are probably not up-to-date. You can update them via: sudo ldconfig.
CUDA SETUP: Solution 2): If you do not have sudo rights, you can do the following:
CUDA SETUP: Solution 2a): Find the cuda library via: find / -name libcuda.so 2>/dev/null
CUDA SETUP: Solution 2b): Once the library is found add it to the LD_LIBRARY_PATH: export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:FOUND_PATH_FROM_2a
CUDA SETUP: Solution 2c): For a permanent solution add the export from 2b into your .bashrc file, located at ~/.bashrc
Traceback (most recent call last):
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\infer.py", line 8, in
from peft import PeftModel
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\peft_init_.py", line 22, in
from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING, PEFT_TYPE_TO_CONFIG_MAPPING, get_peft_config, get_peft_model
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\peft\mapping.py", line 16, in
from .peft_model import (
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\peft\peft_model.py", line 31, in
from .tuners import AdaLoraModel, LoraModel, PrefixEncoder, PromptEmbedding, PromptEncoder
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\peft\tuners_init_.py", line 20, in
from .lora import LoraConfig, LoraModel
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\peft\tuners\lora.py", line 40, in
import bitsandbytes as bnb
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes_init_.py", line 6, in
from . import cuda_setup, utils, research
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\research_init_.py", line 1, in
from . import nn
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\research\nn_init_.py", line 1, in
from .modules import LinearFP8Mixed, LinearFP8Global
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\research\nn\modules.py", line 8, in
from bitsandbytes.optim import GlobalOptimManager
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\optim_init_.py", line 6, in
from bitsandbytes.cextension import COMPILED_WITH_CUDA
File "C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main\venv\lib\site-packages\bitsandbytes\cextension.py", line 20, in
raise RuntimeError('''
RuntimeError:
CUDA Setup failed despite GPU being available. Please run the following command to get more information:

    python -m bitsandbytes

    Inspect the output of the command and see if you can locate CUDA libraries. You might need to add them
    to your LD_LIBRARY_PATH. If you suspect a bug, please take the information from python -m bitsandbytes
    and open an issue at: https://github.com/TimDettmers/bitsandbytes/issues

(venv) PS C:\Users\86152\PycharmProjects\Huatuo-Llama-Med-Chinese-main>
出现这种bug,不知道是什么引起的,应该怎么解决

申请入群

请问如何加群,群显示超过200人需要邀请,下面的个人码显示被加太多,被腾讯风控了

运行infer.sh文件出现错误

运行infer.sh报错:没有找到adapter_config.json文件。
保存的权重文件如下:
image
请问是哪里出错导致没有adapter_config.json文件?

关于SUS的计算

你好,很有意思的工作,工作中提出的SUS评判标准具体是怎么计算得到的呢

申请入群,运行infer文件报错,想请教一下如何解决!

55d0a9a33e0c7ec040ab73f294027ea
当我运行infer.py文件时,报错CUDA Setup failed despite GPU being available.但是,torch.cuda.is_available=true,试了很多方法都没能解决这个问题。
When I ran the infer.py file, I reported an error CUDA Setup failed despite GPU being available. However, torch. cuda. is_ Available=true, I have tried many methods but have not been able to solve this problem.
如有小伙伴知晓如何解决,请加我wechat:15665877987,不胜感激!

About Huatuo

Dear contributors,

Thanks for your insightful work and open-source project. I am happy to share our project (https://github.com/FreedomIntelligence/LLMZoo) with you, which includes a tailored version of LLM in biomedince, and coincidently has the same name "Huatuo". The only difference is that we used different characters (驼or佗). Interestingly, we submitted a paper called "Huatuo" for ACL 2023 last December and meanwhile we have a website called https://www.huatuogpt.cn/ . Plus, we have registered related trademarks in Feb. 2023. See an earlier news in http://www.sribd.cn/article/722 .

We do not intend to offend you but share some similar work with you. Hopefully, some win-win discussions could happend. If you not mind, maybe both names (华驼or华佗) could co-exist -- it might need some further discussions.

Thanks for your project and we highly value the contributions in NLP from SCIR lab. Feel free to write to me via email or something else.

Best regards,
Benyou Wang
[email protected]
HuatuoGPT team.

[BUG]运行infer.py cuda设置失败报错

请问一下,测试用的环境是什么啊。我是用的python3.10.11,安装完环境,cuda(版本为11.7)后运行infer.py
python infer.py --base_model 'decapoda-research/llama-7b-hf' --lora_weights './lora-llama-med' --use_lora True --instruct_dir './data/infer.json' --prompt_template 'med_template'会报如下错误:
Welcome to bitsandbytes. For bug reports, please run

python -m bitsandbytes

and submit this information together with your error trace to: https://github.com/TimDettmers/bitsandbytes/issues

bin C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\libbitsandbytes_cpu.so
C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: WARNING: The following directories listed in your path were found to be non-existent: {WindowsPath('/Users/cqy/.conda/envs/huatuo_cuda117/lib'), WindowsPath('C')}
warn(msg)
C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: C:\Users\cqy.conda\envs\huatuo_cuda117 did not contain ['libcudart.so', 'libcudart.so.11.0', 'libcudart.so.12.0'] as expected! Searching further paths...
warn(msg)
C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: WARNING: The following directories listed in your path were found to be non-existent: {WindowsPath('/127.0.0.1'), WindowsPath('7890'), WindowsPath('http')}
warn(msg)
CUDA_SETUP: WARNING! libcudart.so not found in any environmental path. Searching in backup paths...
C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: WARNING: The following directories listed in your path were found to be non-existent: {WindowsPath('/usr/local/cuda/lib64')}
warn(msg)
CUDA SETUP: WARNING! libcuda.so not found! Do you have a CUDA driver installed? If you are on a cluster, make sure you are on a CUDA machine!
C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: WARNING: No libcudart.so found! Install CUDA or the cudatoolkit package (anaconda)!
warn(msg)
C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\cuda_setup\main.py:145: UserWarning: WARNING: No GPU detected! Check your CUDA paths. Proceeding to load CPU-only library...
warn(msg)
CUDA SETUP: Loading binary C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\libbitsandbytes_cpu.so...
argument of type 'WindowsPath' is not iterable
CUDA SETUP: Problem: The main issue seems to be that the main CUDA library was not detected.
CUDA SETUP: Solution 1): Your paths are probably not up-to-date. You can update them via: sudo ldconfig.
CUDA SETUP: Solution 2): If you do not have sudo rights, you can do the following:
CUDA SETUP: Solution 2a): Find the cuda library via: find / -name libcuda.so 2>/dev/null
CUDA SETUP: Solution 2b): Once the library is found add it to the LD_LIBRARY_PATH: export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:FOUND_PATH_FROM_2a
CUDA SETUP: Solution 2c): For a permanent solution add the export from 2b into your .bashrc file, located at ~/.bashrc
Traceback (most recent call last):
File "E:\cqy-gpt\gpt-fine-tune\Huatuo-Llama-Med-Chinese\infer.py", line 8, in
from peft import PeftModel
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\peft_init_.py", line 22, in
from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING, PEFT_TYPE_TO_CONFIG_MAPPING, get_peft_config, get_peft_model
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\peft\mapping.py", line 16, in
from .peft_model import (
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\peft\peft_model.py", line 31, in
from .tuners import (
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\peft\tuners_init_.py", line 21, in
from .lora import LoraConfig, LoraModel
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\peft\tuners\lora.py", line 40, in
import bitsandbytes as bnb
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes_init_.py", line 6, in
from . import cuda_setup, utils, research
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\research_init_.py", line 1, in
from . import nn
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\research\nn_init_.py", line 1, in
from .modules import LinearFP8Mixed, LinearFP8Global
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\research\nn\modules.py", line 8, in
from bitsandbytes.optim import GlobalOptimManager
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\optim_init_.py", line 6, in
from bitsandbytes.cextension import COMPILED_WITH_CUDA
File "C:\Users\cqy.conda\envs\huatuo_cuda117\lib\site-packages\bitsandbytes\cextension.py", line 20, in
raise RuntimeError('''
RuntimeError:
CUDA Setup failed despite GPU being available. Please run the following command to get more information:

    python -m bitsandbytes

    Inspect the output of the command and see if you can locate CUDA libraries. You might need to add them
    to your LD_LIBRARY_PATH. If you suspect a bug, please take the information from python -m bitsandbytes
    and open an issue at: https://github.com/TimDettmers/bitsandbytes/issues

请问这个问题要如何解决?感谢

chinese-alpaca使用的模板与华驼不一致

chinese-alpaca使用原版stanford-alpaca不带input的模板,是英文的,华驼使用中文的模板在chinese-alpaca的基础上继续指令精调,这样的话,他们模板不一致,会不会产生很大的gap呢?

运行bash ./scripts/infer.sh 报错

异常信息:
ValueError: The current device_map had weights offloaded to the disk. Please provide an offload_folder for them. Alternatively, make sure you have safetensors installed if the model you are using offers the weights in this format.

关于训练数据的生成

您好,想请教以下问题:
1、知识库中的单条数据用于生成几条训练数据?训练数据生成后采用了什么筛选策略呢?如何判断知识库中的信息被正确利用了?
2、med_template.json里的prompt仅用于推理阶段吗?
不胜感激!

训练数据集的构建

请问你们是如何利用GPT3.5的api加本地已有的数据,来构建的训练数据集呢?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.