Power User
  • Communities
  • Create Post
  • Create Community
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
migtissera@alien.topB to LocalLLaMAEnglish · 2 years ago

Introducing Tess: Tess-M with 200K Context Length

message-square
message-square
29
link
fedilink
1
message-square

Introducing Tess: Tess-M with 200K Context Length

migtissera@alien.topB to LocalLLaMAEnglish · 2 years ago
message-square
29
link
fedilink

I’m blown away. See for yourself.

https://migel.substack.com/p/a-conversation-with-tess

Tess, welcome to the world!

Model is Open Source with 200K context length.

Available at: https://huggingface.co/migtissera/Tess-M-v1.0

​

https://preview.redd.it/qehogf32mx0c1.jpg?width=800&format=pjpg&auto=webp&s=333356e7c95e1a1d8fc32dca1d7829720ad2cc84

  • YearZero@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 years ago

    Testing it now, but it’s worse than 7b models on logic questions for me. Huge disappointment compared to Dolphin and Nous-Capybara, both Yi finetunes and are the best models I’ve tested so far. It just goes to show you how much difference finetuning a base model can make.

    • drifter_VR@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 years ago

      Nice, did you manage to make a difference between Dolphin and Nous-Capybara ? Bothe are pretty close to me

      • YearZero@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 years ago

        Nope they’re both really good and very close to each other in my tests: https://docs.google.com/spreadsheets/d/1NgHDxbVWJFolq8bLvLkuPWKC7i_R6I6W/edit?usp=sharing&ouid=102314596465921370523&rtpof=true&sd=true

        • drifter_VR@alien.topB
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 years ago

          Thanks, I remember your tests, it’s great you are still on it.So according to your tests, 34b models compete with GPT3.5. I am not too surprised. And Mistral-7b is not so far behind, what a beast !
          Will you benchmark 70b models too ?

          • YearZero@alien.topB
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 years ago

            Unfortunately I don’t have enough ram/gpu, and too broke right now to afford paying for extra! But in the future I hope I will

LocalLLaMA

localllama

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: !localllama@poweruser.forum

Community to discuss about Llama, the family of large language models created by Meta AI.

Visibility: Public
globe

This community can be federated to other instances and be posted/commented in by their users.

  • 4 users / day
  • 4 users / week
  • 4 users / month
  • 4 users / 6 months
  • 1 local subscriber
  • 11 subscribers
  • 1.02K Posts
  • 5.82K Comments
  • Modlog
  • mods:
  • communick
  • BE: 0.19.11
  • Modlog
  • Instances
  • Docs
  • Code
  • join-lemmy.org