Posted: . At: 9:50 AM. This was 7 months ago. Post ID: 18543
Page permalink. WordPress uses cookies, or tiny pieces of information stored on your computer, to verify who you are. There are cookies for logged in users and for commenters.
These cookies expire two weeks after they are set.


Useful resources on the Internet for working with Large Language Models.


There are many useful resources on the Internet for working with Large Language resources. Such as comparing Large Language Models side by side. Visit this website. This allows comparing two random Chatbots side by side and then ranking the performance of each one.

This is the site. Give the chatbots a request and see the performance of each one. Given a programming challenge, it is striking how each one handles it.

https://chat.lmsys.org/.

It is also possible to run a model locally on your PC. This requires a lot of processing power and GPU VRAM. So a 24 GiB card would be better than a card with 6GiB. But the process of training a model would be the fun part. Find a model here to download and train. Runs locally on YOUR CPU, on YOUR model you trained yourself (but there are some available for download), and it’s Open Source.

https://github.com/nomic-ai/gpt4all.

Other promising projects or those that have future potential: https://github.com/Stability-AI/StableLM/ (demo in https://stablediffusionweb.com/StableLM ).

https://github.com/tatsu-lab/stanford_alpaca https://github.com/tloen/alpaca-lora (from what I understand it is like an optimized version of the previous one to use on modest hardware).

https://github.com/mosaicml/llm-foundry/ (more specifically https://huggingface.co/mosaicml/mpt-7b-chat but other models are available at https://www.mosaicml.com/blog/mpt-7b ).

https://huggingface.co/BlinkDL/rwkv-4-raven and https://huggingface.co/BlinkDL/rwkv-4-world https://github.com/LAION-AI/Open-Assistant (there is a demo at https://open-assistant.io/ but it requires registration).

https://huggingface.co/lmsys/fastchat-t5-3b-v1.0 https://github.com/nlpxucan/WizardLM https://github.com/artidoro/qlora (demo in https://huggingface.co/spaces/uwnlp/guanaco-playground-tgi ).

https://github.com/young-geng/EasyLM/blob/main/docs/koala.md.

https://github.com/databrickslabs/dolly.

https://github.com/THUDM/ChatGLM-6B/blob/main/README_en.md.

In addition to the demos mentioned, you can also try all previous models at https://chat.lmsys.org except rwkv-4-world and StableLM (by default it loads vicuna-13b).

Stanford Alpaca Large Language Model.

https://github.com/tatsu-lab/stanford_alpaca.

A Large Language model in the form of a humanoid android looking through documents in a large library, 1761, high fidelity painting, late afternoon lighting.
A Large Language model in the form of a humanoid android looking through documents in a large library, 1761, high fidelity painting, late afternoon lighting.

These resources should be very useful for setting up your own GPT-4 chatbot. Until the Google LLM is released, this is supposed to be far more massive than GPT-4. PaLM 2 is going to be good competitor to any other Large Language Model. Try Google Bard in the meantime.

https://bard.google.com/chat/.

I asked Google Bard to make this request.

Write a simple program in C for Linux to read all files from a directory and print the output.

And I got this response.

dirread.c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
#include <dirent.h>
#include <stdio.h>
 
int main() {
  DIR *dir;
  struct dirent *entry;
 
  // Open the directory.
  dir = opendir(".");
  if (dir == NULL) {
    perror("opendir");
    return 1;
  }
 
  // Read all entries in the directory.
  while ((entry = readdir(dir)) != NULL) {
    // Check if the entry is a regular file.
    if (entry->d_type == DT_REG) {
      // Print the filename.
      printf("%s\n", entry->d_name);
    }
  }
 
  // Close the directory.
  closedir(dir);
 
  return 0;
}

This program listed the directory instead, but it did work right away with no fuss.


Leave a Comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.