Search found 16 matches

Author Message

NotFirst

Post 24-Jul-2024 22:50

I cannot remove mine. I'm seeding both right now.

NotFirst

Post 23-Jul-2024 19:33

Actually I cannot download yours, the torrent is stalled. I will leave mine seeding for now.

NotFirst

Post 23-Jul-2024 19:30

Nice! You beat me by 15 minutes :) I will delete mine and seed yours.

NotFirst

Post 24-Jul-2024 00:43

It should be working now.

NotFirst

Post 24-Jul-2024 00:24

The torrent generated by AiTracker is corrupted. I am trying to fix it.

NotFirst

Post 23-Jul-2024 23:56

Llama 3.1 70B Instruct released today by Meta.
GGUF by https://huggingface.co/legraphista/Meta-Llama-3.1-7...struct-IMat-GGUF
Model Information
The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes (text in/text out). The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.
Model developer: Meta
Model Architecture: Llama 3.1 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
Supported languages: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
Llama 3.1 family of models. Token counts refer to pretraining data only. All model versions use Grouped-Query Attention (GQA) for improved inference scalability.
Model Release Date: July 23, 2024.
Model card: https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct
!1st

NotFirst

Post 23-Jul-2024 19:28

New & noteworthy: 128K context, function calling. I haven't tried those on the quantized version yet.

NotFirst

Post 23-Jul-2024 19:26

Llama 3.1 8B Instruct released today by Meta.
GGUF by https://huggingface.co/MaziyarPanahi/Meta-Llama-3.1-8B-Instruct-GGUF
Model Information
The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes (text in/text out). The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.
Model developer: Meta
Model Architecture: Llama 3.1 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
Supported languages: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
Llama 3.1 family of models. Token counts refer to pretraining data only. All model versions use Grouped-Query Attention (GQA) for improved inference scalability.
Model Release Date: July 23, 2024.
Model card: https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct

NotFirst

Post 01-Jul-2024 02:53

New version: https://aitracker.art/viewtopic.php?t=77

NotFirst

Post 30-Jun-2024 21:22

GGUF: https://huggingface.co/google/gemma-2-27b-it
Model page: https://huggingface.co/google/gemma-2-27b-it
Authors: Google
Model Information
Summary description and brief definition of inputs and outputs.
Description
Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights for both pre-trained variants and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as a laptop, desktop or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone.
Inputs and outputs
Input: Text string, such as a question, a prompt, or a document to be summarized.
Output: Generated English-language text in response to the input, such as an answer to a question, or a summary of a document.
Citation
@article{gemma_2024,
    title={Gemma},
    url={https://www.kaggle.com/m/3301},
    DOI={10.34740/KAGGLE/M/3301},
    publisher={Kaggle},
    author={Gemma Team},
    year={2024}
}

NotFirst

Post 01-Jul-2024 02:40

Fixed Gemma 2 GGUF - requires llama.cpp b3266 or later
GGUF by legraphista https://huggingface.co/legraphista/gemma-2-27b-it-IMat-GGUF/
Model info: https://aitracker.art/viewtopic.php?t=75

NotFirst

Post 30-Jun-2024 21:17

Hi, since Google is publishing models (Gemma2), could we get a Google section under Large Language Models please?
NVIDIA and Apple are publishing models also.
Edit: under LLM/Others is fine
!1st

NotFirst

Post 22-Jun-2024 02:44

8B q6
https://huggingface.co/SicariusSicariiStuff/LLAMA-3...igned_Alpha_GGUF
!1st

NotFirst

Post 12-Jun-2024 15:11

Q8

NotFirst

Post 12-Jun-2024 15:10

Posting the Q8
 

Current time is: 14-Sep 14:52

All times are UTC + 1

Cover our bill…

Donate at ko-fi.com

Or use crypto:

We accept BTC, XMR, LTC, BCH & DOGE