chatgpt
This article delves into the integration of generative question-answering and reinforcement learning in language models. It highlights projects such as OpenAI, CarperAI, Rallio67’s dataset, KoboldAI, OPT, GPT-Neo, Sentence Transformers, ChatGPT deployment, and various marketing/customer service applications.
GPT4 is out.
besides from decent processors, RAM and optimized runtime, in order to load LLMs fast, one would store the model weights on SSDs.
now colossalai supports chatgpt training with a single gpu, using open-source code
check humata for paper QA and information extraction/language understanding from PDF files
the syntax of chatgpt’s response is obviously markdown.
in order to be unblocked by chatgpt just because we are using static ip of corp’s wifi, we can connect through our phone’s hotspot.
Microsoft’s EdgeGPT needs you to open in Edge browser and join the waitlist of new Bing, having 3rd party API here
Merlin is an extension based on ChatGPT which is avaliable for free and all countries, with 11 queries for free each day. Pro subscriptions incoming.
Rallio67 builds dataset for RLHF and has released multiple chatgpt-like models on huggingface. namely, joi, chip and rosey, all based on pythia or neox-20b. laion people tend to share loads to CPU in order to run these huge models properly.
KoboldAI considered OPT and GPT-Neo as generic LMs. special models like NSFW shits may serve some purposes better.
many alternatives, but many are specialized in marketing and content generation, some are chatgpt replica, like chatsonic (with google knowledge) and youchat (from you.com (awesome!))
open assistant now has a data collection website, in which you can only perform tasks given and earn points (working for free? nah?)
it is adviced to run this chatgpt program with libraries instead of manually, to prevent issues.
my account has been banned from trying chatgpt. though it is not going to be free forever, you need to moderate your input (multi-language support, not only english but chinese) using some api to prevent similar incidents. also some topics outside of blacklist are banned intentionally so you need to check if the model is really producing the answer. if not you should avoid or change the way of asking it.
moderation via official openai api, perspective api (free), or via some projects like content moderation deeplearning, bert text moderation, bert-base-uncased-hatexplain, toxic-bert, copilot-toxicity and multilingual-hate-speech-robacofi, train on datasets like hate_speech_offensive, toxicity (by surge-ai, a dataset labelling workforce) and multilingual-hate-speech
from my point of view, this is a service you cannot replicate at home, either requires smaller models with different architecture, or requires crowd-sourced computational power.
saying chatgpt is powered by ray, increasing parallelism.
bigscience petals colab and petals repo
discord chatroom for reproducing chatgpt
since many different models are derived from the original pretrained language model, opendelta can save disk space by freezing main parameters, only tuning few of them.
this gpt seems really good. currently only api access.
but it is provided by openai which is no longer so “open” in the sense of “open-source”.
stability.ai is providing alternative open-source implementations of SOTA AI algorithms, which includes carper.ai, eleuther.ai, dreamstudio, harmonai (audio), laion.ai (datasets and projects)
viable approaches to chatgpt
according to my point of view, chatgpt is just specialized on chat, or socialized in other words.
the elo rating system is the key to facebook social network, many zero-sum games. basically it is some revolution rating system. to do such rating system effectively one shall use along with classifiers and embeddings.
according to the training process of instructgpt and webgpt, we know that gpt has learned more by interacting with people (multiple QA), doing self-examination (learning a reward model) and performing actions (searching and quoting on web).
RLHF
chainer, prompt engineering
langchain extending llm by advanced prompts, llm wrappers actions, databases and memories
RL algorithms, tools for providing feedback
Awesome-RLHF paper and code about RLHF
Efficient few-shot learning with Sentence Transformers, used by FewShotRLGPT (no updates till now?)
RLHF models
non-language models
language models
chatrwkv pure rnn language model, with chinese support
blenderbot2 a bot which can search internet, blenderbot3 is US only. install ParlAI then clone ParlAI_SearchEngine. tutorial
promptCLUE based on T5, created by clueai, trained on pCLUE
openchatgpt-neox-125m trained on chatgpt prompts, can be tested here, trained from pythia
copycat chatgpt replicate
medicine-chatgpt shit sick of COVID-19
baby-rlhf both cartpole and languge model
textrl 100+stars
PaLM-RLHF claims RETRO will be integrated soon?
RL4LMs with multiple rl methods
webgpt-cli interface openai api to browse web and answer questions
lm-human-preferences by openai
rlhf-magic using trlx (supports GPT3-like models) which has PPO and ILQL (as trainable model)
trl only has PPO on GPT2
Tk-Instruct T5 trained on natural instruct dataset. is it trained on RLHF systems?
datasets
whisperhub collection of chatgpt prompts by plugin
dataset building tools
open-chatgpt-prompt-collective
crowd-kit purify noisy data
reward models
rankgen scores model generations given a prefix (or prompt)
electra-webgpt-rm and electra-large-reward-model is based on electra discriminator
GPT3-like models
galactica is opt trained on scientific data
bloomz and mt0 trained on xP3 (multilingual prompts and code)
T0PP T0 optimized for zero-shot prompts, despite much smaller than GPT-3
RETRO another model with GPT-3 capabilities with fewer parameters?
gpt3 is gpt2 with sparse attension, which enables it to generate long sequence
metaseq provides OPT, which is basically GPT3
GPT-JT altered in many ways, trained on natural instructions huggingface space
Bloom large language model by bigscience
autonomous learning
autonomous-learning-library doc and repo
Gu-X doing god-knows-what experiments
analysis about how to make such model
gpt3 is capable of imitation (cause it is unsupervised.)
but! if you want to get things done (when you really need it!), you better want some aligned AI.
two similar models by openai: webgpt and instructgpt
about instructgpt
it is first fine-tuned on supervised datasets, then train some reward model, then use the reward model to handle prompts and do reinforcement learning with PPO.
details on webgpt environment
guess: create states by performing actions, then generate templates to allow model filling blanks.
1 | Our text-based web-browsing environment is written mostly in Python with some JavaScript. For a |
projects related to chatgpt
voice assistants
voice assistant in cpp
ChatWaifu with anime voice, ChatWaifu with live2d
hacking
give longterm memory and external resources to gpt3
hackgpt exploit vulnerabilities
vulchatgpt ida plugin for reverse engineering
chatgpt-universe things related to chatgpt
记笔记
12.27更新了一个更精简的应用
强烈建议部署到服务器上
huggingface参考:https://huggingface.co/spaces/Mahiruoshi/Lovelive-Nijigasaku-Chat-iSTFT-GPT3
GitHub:https://github.com/Paraworks/vits_with_chatgpt-gpt3
地址:https://drive.google.com/drive/folders/1vtootVMQ7wTOQwd15nJe6akzJUYNOw4d?usp=share_link
你可以先尝试在服务器上部署,之后可以直接解压进文件夹后运行exe(mac、安卓端需要用renpy自行编译)
去https://beta.openai.com/account/api-keys获取api-key
参数照着敲就好了
人物id通常是从0开始的数字,我的模型最大到12
api部署方法:把inference_api.py放入你的vits目录下,进入文件修改config和checkpoint.pth的路径,比起应用程序来说十分简单,可以自行设计。码龄三个月写出的的雪山代码警告
——————————————————————————————————————————————————
Chatgpt部署方法已于12.26更新(视频后部分)
vits参考:https://github.com/CjangCjengh/vits
服务器端建议用ISTFT VITS:https://github.com/innnky/MB-iSTFT-VITS
model库:https://github.com/CjangCjengh/TTSModels
也可以用我的https://huggingface.co/spaces/Mahiruoshi/MIT-VITS-Nijigaku
CHATGPT参考:https://github.com/rawandahmad698/PyChatGPT
示例视频(纯服务器api,gpt3)https://www.bilibili.com/video/BV1hP4y1B7wH/?spm_id_from=333.999.0.0&vd_source=7e8cf9f5c840ec4789ccb5657b2f0512
穗乃果配音来自缪斯的模型@Freeze_Phoenix
gpt3加载参考@ぶらぶら散策中
chatgpt use cases curated list
DAILA use chatgpt
to identify function calls in decompiler
awesome transformer language models a huge collection on transformer based LMs, huge models by megacorps, with some introduction and analogy on chatgpt
huggingface blog on RLHF containing similar projects and source code
bilibili sends me lots of videos (and articles) on hacking and ai (including chatgpt) via its android app. recommend you to scrape this source and collect transcription and screenshots for searching and content generation.
b站有做免杀 绕过杀软的
chatgpt对接搜索引擎
下载链接:
github: https://github.com/josStorer/chat-gpt-search-engine-extension/releases/
百度网盘: https://pan.baidu.com/s/1MnFJTDIatyIIPr5kUMWsAw?pwd=1111
提取码:1111
原项目: https://github.com/wong2/chat-gpt-google-extension
我创建的fork, 添加了多个搜索引擎支持的版本: https://github.com/josStorer/chat-gpt-search-engine-extension
PR: https://github.com/wong2/chat-gpt-google-extension/pull/31
已修复先前百度需要手动刷新的问题
access via api
https://github.com/altryne/chatGPT-telegram-bot
https://github.com/taranjeet/chatgpt-api
https://github.com/acheong08/ChatGPT
https://github.com/vincelwt/chatgpt-mac
https://github.com/transitive-bullshit/chatgpt-api
https://github.com/rawandahmad698/PyChatGPT
models like chatgpt
lfqa retrival based generative QA
lm-human-preferences by openai
trl Train transformer language models with reinforcement learning based on gpt2
trlx A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF) by CarperAI
RL4LMs A modular RL library to fine-tune language models to human preferences
PaLM-rlhf-pytorch saying this is basically chatgpt with palm
gpt-gmlp saying this design integrates gpt with gmlps so will use less ram and can be trained on a single gpu
tk-instruct with all models by allenai can be multilingual, trained on natural instructions
there’s a ghosted repo named instructgpt-pytorch found in bing but no cache preserved, also an empty repo called InstructFNet wtf?
AidMe Code and experiment of the article AidMe User-in-the-loop Adaptative Intent Detecttion for Instructable Digital Assistant
cheese Used for adaptive human in the loop evaluation of language and embedding models.
Kelpie Explainable AI framework for interpreting Link Predictions on Knowledge Graphs
GrIPS Gradient-free, Edit-based Instruction Search for Prompting Large Language Models
queakily nlp datasets cleaner
gpt-j
super big bilingual model GLM-130B
multi-modal deeplearning paper collections
bloom a huge model like gpt-3
notice, gpt-2 is somehow inferior to gpt-3 since it has smaller model parameters
dialogue-generation Generating responses with pretrained XLNet and GPT-2 in PyTorch.
personaGPT Implementation of PersonaGPT Dialog Model
DialoGPT Large-scale pretraining for dialogue