From 1a939319aa2ebca7ced5486877cbc210887ac84e Mon Sep 17 00:00:00 2001 From: camenduru <54370274+camenduru@users.noreply.github.com> Date: Wed, 16 Aug 2023 19:37:57 +0300 Subject: [PATCH] test --- README.md | 2 +- ...r-gpt-7b-text-generation-webui-colab.ipynb | 59 +++++++++++++++++++ 2 files changed, 60 insertions(+), 1 deletion(-) create mode 100644 doctor-gpt-7b-text-generation-webui-colab.ipynb diff --git a/README.md b/README.md index 61c1a04..17c8e4c 100644 --- a/README.md +++ b/README.md @@ -7,6 +7,7 @@ | Colab | Info | --- | --- | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/camenduru/DoctorGPT-colab/blob/main/DoctorGPT_colab.ipynb) | DoctorGPT_colab (Gradio) +[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/camenduru/DoctorGPT-colab/blob/main/doctor-gpt-7b-text-generation-webui-colab.ipynb) | doctor-gpt-7b-text-generation-webui-colab https://github.com/oobabooga/text-generation-webui (Thanks to @oobabooga ❤) ## Main Repo https://github.com/llSourcell/DoctorGPT @@ -14,4 +15,3 @@ https://github.com/llSourcell/DoctorGPT ## Model https://huggingface.co/llSourcell/medllama2_7b/tree/main -## Output diff --git a/doctor-gpt-7b-text-generation-webui-colab.ipynb b/doctor-gpt-7b-text-generation-webui-colab.ipynb new file mode 100644 index 0000000..5eabefc --- /dev/null +++ b/doctor-gpt-7b-text-generation-webui-colab.ipynb @@ -0,0 +1,59 @@ +{ + "cells": [ + { + "cell_type": "markdown", + "metadata": { + "id": "view-in-github" + }, + "source": [ + "[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/camenduru/DoctorGPT-colab/blob/main/doctor-gpt-7b-text-generation-webui-colab.ipynb)" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "id": "VCFOzsQSHbjM" + }, + "outputs": [], + "source": [ + "%cd /content\n", + "!apt-get -y install -qq aria2\n", + "\n", + "!git clone -b v2.1 https://github.com/camenduru/text-generation-webui\n", + "%cd /content/text-generation-webui\n", + "!pip install -r requirements.txt\n", + "\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/resolve/main/model-00001-of-00002.safetensors -d /content/text-generation-webui/models/medllama2_7b -o model-00001-of-00002.safetensors\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/resolve/main/model-00002-of-00002.safetensors -d /content/text-generation-webui/models/medllama2_7b -o model-00002-of-00002.safetensors\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/raw/main/model.safetensors.index.json -d /content/text-generation-webui/models/medllama2_7b -o model.safetensors.index.json\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/raw/main/special_tokens_map.json -d /content/text-generation-webui/models/medllama2_7b -o special_tokens_map.json\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/raw/main/tokenizer.json -d /content/text-generation-webui/models/medllama2_7b -o tokenizer.json\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/raw/main/tokenizer_config.json -d /content/text-generation-webui/models/medllama2_7b -o tokenizer_config.json\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/raw/main/config.json -d /content/text-generation-webui/models/medllama2_7b -o config.json\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/raw/main/generation_config.json -d /content/text-generation-webui/models/medllama2_7b -o generation_config.json\n", + "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/4bit/medllama2_7b_s/resolve/main/tokenizer.model -d /content/text-generation-webui/models/medllama2_7b -o tokenizer.model\n", + "\n", + "%cd /content/text-generation-webui\n", + "!python server.py --share --model /content/text-generation-webui/models/medllama2_7b" + ] + } + ], + "metadata": { + "accelerator": "GPU", + "colab": { + "provenance": [] + }, + "gpuClass": "standard", + "kernelspec": { + "display_name": "Python 3", + "name": "python3" + }, + "language_info": { + "name": "python" + } + }, + "nbformat": 4, + "nbformat_minor": 0 + } + \ No newline at end of file