Skip to content

Deploying Libraries and Tools

This document provides an overview of various libraries and tools that can be used for deploying AI models. It is divided into several sections, each focusing on a specific aspect of deployment. The sections include LLM Ops, Models, Finetuning, Serving, Programming Convenience, Memory Interaction, Executors and Interpreters, Data Creation, and General.

LLM Ops

LLM Ops refers to operations related to Large Language Models. Here are a couple of tools that can assist in managing these operations:

  • LLM Ops: This is a Microsoft tool for managing large language models.
  • Reliable GPT: This is a wrapper that prevents failures due to rate limiting requests.

Models

This section provides a selection of repositories that enable the creation of models:

  • Hugging Face Transformers: This is a popular library for creating transformer models.

  • Chatall: This tool allows interaction with multiple chatbots at the same time.

  • LocalAI: This is a drop-in replacement REST API thatā€™s compatible with OpenAI API specifications for local inferencing.
GitHub Repo stars 'This project (ToolLLM) Tool Bench 'This project (ToolLLM) aims to construct open-source, large-scale, high-quality instruction tuning SFT data to facilitate the construction of powerful LLMs with general tool-use capability.'

image

Chat with Memory

Anything LLM is a full-stack application that enables you to turn any document, resource, or piece of content into context that any LLM can use as references during chatting.

This application allows you to pick and choose which LLM or Vector Database you want to use as well as supporting multi-user management and permissions.

MemGPT allows you to build LLM agents with self-editing memory

Serving

GitHub Repo stars Open LLM to run inference with any open-source large-language models, deploy to the cloud or on-premises, and build powerful AI apps.

Distributed

GitHub Repo stars Petals Run large language models at home, BitTorrent-style.

Generate text with distributed LLaMA 2 (70B), Stable Beluga 2, Guanaco-65B or BLOOM-176B and fineā€‘tune them for your own tasks ā€” right from your desktop computer or Google Cola Launch your own swarm

Programming Convenience

GitHub Repo stars Magentic for decorators

A nice and simple plugin that allows a @prompt decorator to call functions as an llm, including function-choice calls. Their example](https://github.com/jackmpcollins/magentic)

from typing import Literal

from magentic import prompt, FunctionCall


def activate_oven(temperature: int, mode: Literal["broil", "bake", "roast"]) -> str:
    """Turn the oven on with the provided settings."""
    return f"Preheating to {temperature} F with mode {mode}"


@prompt(
    "Prepare the oven so I can make {food}",
    functions=[activate_oven],
)
def configure_oven(food: str) -> FunctionCall[str]:
    ...


output = configure_oven("cookies!")
# FunctionCall(<function activate_oven at 0x1105a6200>, temperature=350, mode='bake')
output()
# 'Preheating to 350 F with mode bake'