Power User
  • Communities
  • Create Post
  • Create Community
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
panchovix@alien.topB to LocalLLaMAEnglish · 2 years ago

TabbyAPI released! A pure LLM API for exllama v2.

github.com

external-link
message-square
6
link
fedilink
1
external-link

TabbyAPI released! A pure LLM API for exllama v2.

github.com

panchovix@alien.topB to LocalLLaMAEnglish · 2 years ago
message-square
6
link
fedilink
GitHub - theroyallab/tabbyAPI
github.com
external-link
Contribute to theroyallab/tabbyAPI development by creating an account on GitHub.
  • a_beautiful_rhind@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 years ago

    Nice. A lightweight loader. Will make us free of gradio.

    • oobabooga4@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 years ago

      Gradio is a 70MB requirement FYI. It has become common to see people calling text-generation-webui “bloated”, when most of the installation size is in fact due to Pytorch and the CUDA runtime libraries.

      https://preview.redd.it/pgfsdld7xw0c1.png?width=370&format=png&auto=webp&s=c50a14804350a1391d57d0feac8a32a5dcf36f68

      • tronathan@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 years ago

        Gradio is a 70MB requirement

        That doesn’t make it fast, just small. Inefficient code can be compact.

      • kpodkanowicz@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 years ago

        I think there is room for everyone - Text Gen is a piece of art - it’s the only thing in the whole space that always works and is reliable. However, if im building an agent and getting a docker build, I can not afford to change text gen etc.

LocalLLaMA

localllama

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: !localllama@poweruser.forum

Community to discuss about Llama, the family of large language models created by Meta AI.

Visibility: Public
globe

This community can be federated to other instances and be posted/commented in by their users.

  • 4 users / day
  • 4 users / week
  • 4 users / month
  • 4 users / 6 months
  • 1 local subscriber
  • 5 subscribers
  • 1.02K Posts
  • 5.82K Comments
  • Modlog
  • mods:
  • communick
  • BE: 0.19.11
  • Modlog
  • Instances
  • Docs
  • Code
  • join-lemmy.org