{ "cells": [ { "cell_type": "code", "execution_count": 1, "id": "d2ff171c-f5f8-4590-9ce0-21c87e3d5b39", "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "INFO 2023-06-01 20:26:48,576-1d: \n", "loading model config\n", "llm device: cuda\n", "embedding device: cuda\n", "dir: /media/gpt4-pdf-chatbot-langchain/dev-langchain-ChatGLM\n", "flagging username: 7daba79785044bceb6896b9e6f8f9894\n", "\n" ] } ], "source": [ "import sys\n", "sys.path.append('/media/gpt4-pdf-chatbot-langchain/dev-langchain-ChatGLM/')\n", "from langchain.llms.base import LLM\n", "import torch\n", "import transformers \n", "import models.shared as shared \n", "from abc import ABC\n", "\n", "from langchain.llms.base import LLM\n", "import random\n", "from transformers.generation.logits_process import LogitsProcessor\n", "from transformers.generation.utils import LogitsProcessorList, StoppingCriteriaList\n", "from typing import Optional, List, Dict, Any\n", "from models.loader import LoaderCheckPoint \n", "from models.base import (BaseAnswer,\n", " AnswerResult)\n", "\n" ] }, { "cell_type": "code", "execution_count": 3, "id": "68978c38-c0e9-4ae9-ba90-9c02aca335be", "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "Loading vicuna-7b-hf...\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Overriding torch_dtype=None with `torch_dtype=torch.float16` due to requirements of `bitsandbytes` to enable model loading in mixed int8. Either pass torch_dtype=torch.float16 or don't pass this argument at all to remove this warning.\n", "/media/gpt4-pdf-chatbot-langchain/pyenv-langchain/lib/python3.10/site-packages/bitsandbytes/cuda_setup/main.py:149: UserWarning: /media/gpt4-pdf-chatbot-langchain/pyenv-langchain did not contain ['libcudart.so', 'libcudart.so.11.0', 'libcudart.so.12.0'] as expected! Searching further paths...\n", " warn(msg)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\n", "===================================BUG REPORT===================================\n", "Welcome to bitsandbytes. For bug reports, please run\n", "\n", "python -m bitsandbytes\n", "\n", " and submit this information together with your error trace to: https://github.com/TimDettmers/bitsandbytes/issues\n", "================================================================================\n", "bin /media/gpt4-pdf-chatbot-langchain/pyenv-langchain/lib/python3.10/site-packages/bitsandbytes/libbitsandbytes_cuda118.so\n", "CUDA SETUP: CUDA runtime path found: /opt/cuda/lib64/libcudart.so.11.0\n", "CUDA SETUP: Highest compute capability among GPUs detected: 8.6\n", "CUDA SETUP: Detected CUDA version 118\n", "CUDA SETUP: Loading binary /media/gpt4-pdf-chatbot-langchain/pyenv-langchain/lib/python3.10/site-packages/bitsandbytes/libbitsandbytes_cuda118.so...\n" ] }, { "data": { "application/vnd.jupyter.widget-view+json": { "model_id": "9b61d05e18044b009c72b862c84ab5cb", "version_major": 2, "version_minor": 0 }, "text/plain": [ "Loading checkpoint shards: 0%| | 0/2 [00:00