1
mirror of https://github.com/cocktailpeanut/dalai synced 2024-11-23 04:16:53 +01:00
Go to file
cocktailpeanut a32b544014 windows fix
2023-03-16 19:57:02 -04:00
bin windows fix 2023-03-16 19:57:02 -04:00
demo Remove unused line of code with syntax error 2023-03-13 04:38:53 -04:00
docs dalai alpaca 2023-03-16 18:46:41 -04:00
.gitignore win32 support 2023-03-15 09:24:47 -07:00
.prettierignore Add .prettierignore file to avoid auto-formatting files (and complicate PR merges); but also add .prettierrc and yarn script to auto-format the entire repo to encourage the decision to enable this ability 2023-03-13 04:43:06 -04:00
.prettierrc.json Add .prettierignore file to avoid auto-formatting files (and complicate PR merges); but also add .prettierrc and yarn script to auto-format the entire repo to encourage the decision to enable this ability 2023-03-13 04:43:06 -04:00
alpaca.js windows fix 2023-03-16 19:57:02 -04:00
dalai Add ./dalai script which runs the cli script for test running, and fix yarn scripts 2023-03-13 04:40:16 -04:00
index.js windows fix 2023-03-16 19:57:02 -04:00
llama.js windows fix 2023-03-16 19:57:02 -04:00
package-lock.json update 2023-03-16 19:15:14 -04:00
package.json update ver 2023-03-16 18:50:13 -04:00
yarn.lock Use Visual Studio on Windows 2023-03-15 13:48:17 -04:00

Dalai

Run LLaMA and Alpaca on your computer.

Github Twitter Discord


JUST RUN THIS

or

TO GET

Both alpaca and llama working on your computer!

alpaca.gif


  1. Powered by llama.cpp, llama-dl CDN, and alpaca.cpp
  2. Hackable web app included
  3. Ships with JavaScript API
  4. Ships with Socket.io API

Intro

1. Cross platform

Dalai runs on all of the following operating systems:

  1. Linux
  2. Mac
  3. Windows

2. System Requirements

Runs on most modern computers. Unless your computer is very very old, it should work.

3. Disk Space Requirements

Alpaca

Currently only the 7B model is available via alpaca.cpp

7B

Alpaca comes fully quantized (compressed), and the only space you need for the 7B model is 4.21GB:

alpaca_spec.png

LLaMA

You need a lot of space for storing the models.

You do NOT have to install all models, you can install one by one. Let's take a look at how much space each model takes up:

NOTE

The following numbers assume that you DO NOT touch the original model files and keep BOTH the original model files AND the quantized versions.

You can optimize this if you delete the original models (which are much larger) after installation and keep only the quantized versions.

7B

  • Full: The model takes up 31.17GB
  • Quantized: 4.21GB

7b.png

13B

  • Full: The model takes up 60.21GB
  • Quantized: 4.07GB * 2 = 8.14GB

13b.png

30B

  • Full: The model takes up 150.48GB
  • Quantized: 5.09GB * 4 = 20.36GB

30b.png

65B

  • Full: The model takes up 432.64GB
  • Quantized: 5.11GB * 8 = 40.88GB

65b.png


Quickstart

Mac

Step 1. Install node.js

Install Node.js

Step 2. Install Dalai

First install dalai:

npm install -g dalai

Step 3. Install Engines

Currently supported engines are llama and alpaca.

Install LLaMA

To install llama, run:

dalai llama install

Install Alpaca

To install alpaca, run:

dalai alpaca install

Step 4. Get Models

Download LLaMA models

To download llama models, you can run:

dalai llama get 7B

or to download multiple models:

dalai llama get 7B 13B

Download Alpaca models

Currently alpaca only has the 7B model:

dalai alpaca get 7B

Step 3. Run Web UI

After everything has been installed, run the following command to launch the web UI server:

dalai serve

and open http://localhost:3000 in your browser. Have fun!


Windows

Step 1. Install Visual Studio

On windows, you need to install Visual Studio before installing Dalai.

Press the button below to visit the Visual Studio downloads page and download:

Download Microsoft Visual Studio

IMPORTANT!!!

When installing Visual Studio, make sure to check the 3 options as highlighted below:

  1. Python development
  2. Node.js development
  3. Desktop development with C++

vs.png


Step 2.1. Install Dalai

First install dalai:

npm install -g dalai

If this worked without any errors, go to step 3.

Ohterwise try the troubleshoot below:


Step 2.2. Troubleshoot (optional)

In case above steps fail to install, try installing node.js and python separately.

Install Python:

Download Python

Install Node.js:

Download Node.js

After both have been installed, open powershell and type python to see if the application exists. And also type node to see if the application exists as well.

Once you've checked that they both exist, try the npx dalai llama command again.


Step 3. Install Engines

Currently supported engines are llama and alpaca.

Install LLaMA

To install llama, run:

dalai llama install

Install Alpaca

To install alpaca, run:

dalai alpaca install

Step 4. Get Models

Download LLaMA models

To download llama models, you can run:

dalai llama get 7B

or to download multiple models:

dalai llama get 7B 13B

Download Alpaca models

Currently alpaca only has the 7B model:

dalai alpaca get 7B

Step 5. Run Web UI

After everything has been installed, run the following command to launch the web UI server:

dalai serve

and open http://localhost:3000 in your browser. Have fun!


Linux

Step 1. Install Dependencies

You need to make sure you have the correct version of Python and Node.js installed.

Step 1.1. Python <= 3.10

Download node.js

Make sure the version is 3.10 or lower (not 3.11) Python must be 3.10 or below (pytorch and other libraries are not supported yet on the latest)

Step 1.2. Node.js >= 18

Download node.js

Make sure the version is 18 or higher

Step 2. Install Dalai

First install dalai:

npm install -g dalai

Step 3. Install Engines

Currently supported engines are llama and alpaca.

Install LLaMA

To install llama, run:

dalai llama install

Install Alpaca

To install alpaca, run:

dalai alpaca install

Step 4. Get Models

Download LLaMA models

To download llama models, you can run:

dalai llama get 7B

or to download multiple models:

dalai llama get 7B 13B

Download Alpaca models

Currently alpaca only has the 7B model:

dalai alpaca get 7B

Step 3. Run Web UI

After everything has been installed, run the following command to launch the web UI server:

dalai serve

and open http://localhost:3000 in your browser. Have fun!


Commands

1. install

LLaMA

Install the core engine for the model

dalai llama install

Alpaca

Install the core engine for the model

dalai alpaca install

2. get

Download the full LLaMA model and convert and compress them

LLaMA

Download one model:

dalai llama get 7B

Download multiple models:

dalai llama get 7B 13B

Alpaca

Currently only 7B available:

dalai alpaca get 7B

3. serve

Start a dalai server and an API endpoint (powered by socket.io)

dalai serve

API

Dalai is also an NPM package:

  1. programmatically install
  2. locally make requests to the model
  3. run a dalai server (powered by socket.io)
  4. programmatically make requests to a remote dalai server (via socket.io)

Dalai is an NPM package. You can install it using:

npm install dalai

1. constructor()

Syntax

const dalai = new Dalai(home)
  • home: (optional) manually specify the llama.cpp folder

By default, Dalai automatically stores the entire llama.cpp repository under ~/llama.cpp.

However, often you may already have a llama.cpp repository somewhere else on your machine and want to just use that folder. In this case you can pass in the home attribute.

Examples

Basic

Creates a workspace at ~/llama.cpp

const dalai = new Dalai()

Custom path

Manually set the llama.cpp path:

const dalai = new Dalai("/Documents/llama.cpp")

2. request()

Syntax

dalai.request(req, callback)
  • req: a request object. made up of the following attributes:
    • prompt: (required) The prompt string
    • model: (required) The model type + model name to query. Takes the following form: <model_type>.<model_name>
      • Example: alpaca.7B, llama.13B, ...
    • url: only needed if connecting to a remote dalai server
      • if unspecified, it uses the node.js API to directly run dalai locally
      • if specified (for example ws://localhost:3000) it looks for a socket.io endpoint at the URL and connects to it.
    • threads: The number of threads to use (The default is 8 if unspecified)
    • n_predict: The number of tokens to return (The default is 128 if unspecified)
    • seed: The seed. The default is -1 (none)
    • top_k
    • top_p
    • repeat_last_n
    • repeat_penalty
    • temp: temperature
    • batch_size: batch size
    • skip_end: by default, every session ends with \n\n<end>, which can be used as a marker to know when the full response has returned. However sometimes you may not want this suffix. Set skip_end: true and the response will no longer end with \n\n<end>
  • callback: the streaming callback function that gets called every time the client gets any token response back from the model

Examples

1. Node.js

Using node.js, you just need to initialize a Dalai object with new Dalai() and then use it.

const Dalai = require('dalai')
new Dalai().request({
  model: "7B",
  prompt: "The following is a conversation between a boy and a girl:",
}, (token) => {
  process.stdout.write(token)
})

2. Non node.js (socket.io)

To make use of this in a browser or any other language, you can use thie socket.io API.

Step 1. start a server

First you need to run a Dalai socket server:

// server.js
const Dalai = require('dalai')
new Dalai().serve(3000)     // port 3000
Step 2. connect to the server

Then once the server is running, simply make requests to it by passing the ws://localhost:3000 socket url when initializing the Dalai object:

const Dalai = require("dalai")
new Dalai().request({
  url: "ws://localhost:3000",
  model: "7B",
  prompt: "The following is a conversation between a boy and a girl:",
}, (token) => {
  console.log("token", token)
})

3. serve()

Syntax

Starts a socket.io server at port

dalai.serve(port)

Examples

const Dalai = require("dalai")
new Dalai().serve(3000)

4. http()

Syntax

connect with an existing http instance (The http npm package)

dalai.http(http)
  • http: The http object

Examples

This is useful when you're trying to plug dalai into an existing node.js web app

const app = require('express')();
const http = require('http').Server(app);
dalai.http(http)
http.listen(3000, () => {
  console.log("server started")
})

5. get()

Syntax

await dalai.install(model_type, model_name1, model_name2, ...)
  • model_type: the name of the model. currently supports:
    • "alpaca"
    • "llama"
  • model1, model2, ...: the model names to install ("7B"`, "13B", "30B", "65B", etc)

Examples

Install the "7B" and "13B" models:

const Dalai = require("dalai");
const dalai = new Dalai()
await dalai.install("7B", "13B")

6. installed()

returns the array of installed models

Syntax

const models = await dalai.installed()

Examples

const Dalai = require("dalai");
const dalai = new Dalai()
const models = await dalai.installed()
console.log(models)     // prints ["7B", "13B"]

FAQ

Updating to the latest

As of dalai@0.3.0 the recommended way to use dalai is through npm install -g (not the npx method)

The simplest way to make sure you have the correct version is running:

npm install -g dalai@0.3.0    

Staying up to date

Have questions or feedback? Follow the project through the following outlets:

Github Twitter Discord