Skip to content
This repository has been archived by the owner on Jun 26, 2024. It is now read-only.

Releases: aniketmaurya/llm-inference

v0.0.6

16 Jul 15:19
Compare
Choose a tag to compare

What's Changed

Full Changelog: v0.0.5...v0.0.6

Chatbot with Lit-GPT x LangChain

06 Jul 14:38
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.0.4...v0.0.5

Chatbot support & bug fixes

17 Apr 10:36
Compare
Choose a tag to compare

What's Changed

How to use Chatbot

from chatbot import LLaMAChatBot

checkpoint_path = f"state_dict.pth"
tokenizer_path = f"tokenizer.model"

bot = LLaMAChatBot(
    checkpoint_path=checkpoint_path, tokenizer_path=tokenizer_path
)

print(bot.send("hi, what is the capital of France?"))

Full Changelog: v0.0.2...v0.0.3

v0.0.2

16 Apr 19:22
Compare
Choose a tag to compare

What's Changed

For inference

from llama_inference import LLaMAInference
import os

WEIGHTS_PATH = os.environ["WEIGHTS"]

checkpoint_path = f"{WEIGHTS_PATH}/lit-llama/7B/state_dict.pth"
tokenizer_path = f"{WEIGHTS_PATH}/lit-llama/tokenizer.model"

model = LLaMAInference(checkpoint_path=checkpoint_path, tokenizer_path=tokenizer_path, dtype="bfloat16")

print(model("New York is located in"))

For serving a REST API

# app.py
from llama_inference.serve import ServeLLaMA, Response

import lightning as L

component = ServeLLaMA(input_type=PromptRequest, output_type=Response)
app = L.LightningApp(component)

Full Changelog: v0.0.1...v0.0.2

v0.0.1

16 Apr 19:22
Compare
Choose a tag to compare

What's Changed

Full Changelog: https://github.com/aniketmaurya/LLaMA-Inference-API/commits/v0.0.1