Author |
Message |
8B
»
Meta-Llama-3.1-8B-Instruct-Q8_0.gguf
|
NotFirst
|
I cannot remove mine. I'm seeding both right now.
|
NotFirst
|
Actually I cannot download yours, the torrent is stalled. I will leave mine seeding for now.
|
NotFirst
|
Nice! You beat me by 15 minutes :) I will delete mine and seed yours.
|
70B
»
Meta-Llama-3.1-70B-Instruct-GGUF [Q8_0]
|
NotFirst
|
It should be working now.
|
NotFirst
|
The torrent generated by AiTracker is corrupted. I am trying to fix it.
|
NotFirst
|
Llama 3.1 70B Instruct released today by Meta.
GGUF by https://huggingface.co/legraphista/Meta-Llama-3.1-7...struct-IMat-GGUF
Model Information The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes (text in/text out). The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks. Model developer: Meta Model Architecture: Llama 3.1 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety. Supported languages: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. Llama 3.1 family of models. Token counts refer to pretraining data only. All model versions use Grouped-Query Attention (GQA) for improved inference scalability. Model Release Date: July 23, 2024.
Model card: https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct
!1st
|
8B
»
Meta-Llama-3.1-8B-Instruct-GGUF [Q8_0]
|
NotFirst
|
New & noteworthy: 128K context, function calling. I haven't tried those on the quantized version yet.
|
NotFirst
|
Llama 3.1 8B Instruct released today by Meta.
GGUF by https://huggingface.co/MaziyarPanahi/Meta-Llama-3.1-8B-Instruct-GGUF
Model Information The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes (text in/text out). The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks. Model developer: Meta Model Architecture: Llama 3.1 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety. Supported languages: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. Llama 3.1 family of models. Token counts refer to pretraining data only. All model versions use Grouped-Query Attention (GQA) for improved inference scalability. Model Release Date: July 23, 2024.
Model card: https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
|
Misc
»
bartowski/gemma-2-27b-it-GGUF Gemma 2 IT 27B Q8_0_L
|
NotFirst
|
|
NotFirst
|
GGUF: https://huggingface.co/google/gemma-2-27b-it
Model page: https://huggingface.co/google/gemma-2-27b-it
Authors: Google Model Information
Summary description and brief definition of inputs and outputs. Description
Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights for both pre-trained variants and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as a laptop, desktop or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone. Inputs and outputs
Input: Text string, such as a question, a prompt, or a document to be summarized.
Output: Generated English-language text in response to the input, such as an answer to a question, or a summary of a document. Citation
@article{gemma_2024,
title={Gemma},
url={https://www.kaggle.com/m/3301},
DOI={10.34740/KAGGLE/M/3301},
publisher={Kaggle},
author={Gemma Team},
year={2024}
}
|
Misc
»
legraphista/gemma-2-27b-it-IMat-GGUF Gemma 2 27B IMat Q8_0 GGUF
|
NotFirst
|
|
Site suggestions & feedback
»
Subcategory request: LLM/Others/Google
|
NotFirst
|
Hi, since Google is publishing models (Gemma2), could we get a Google section under Large Language Models please? NVIDIA and Apple are publishing models also. Edit: under LLM/Others is fine !1st
|
8B
»
SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha_GGUF (q6)
|
NotFirst
|
|
Misc
»
bartowski/Codestral-22B-v0.1-GGUF (q8)
|
NotFirst
|
|
Misc
»
bartowski/Codestral-22B-v0.1-GGUF (q4_k_m)
|
NotFirst
|
|
|