[Chinese Version] Mixtral-8x7B model | 中文Mixtral-8x7B模型
#73 opened 5 months ago
by
wangrongsheng
Update the deprecated Flash Attention call parameter in from_pretrained() method
#72 opened 5 months ago
by
DeathReaper0965
can't load the model
2
#71 opened 5 months ago
by
JayZhang1
What is the best way for the inference process in LORA in PEFT approach
8
#70 opened 5 months ago
by
Pradeep1995
How to use system prompt?
1
#69 opened 5 months ago
by
mznw
Is there any simple way to solve the problem of redundant output
3
#68 opened 5 months ago
by
jjplane
Which is the actual way to store the adapters after PEFT finetuning
4
#67 opened 5 months ago
by
Pradeep1995
Failed to import transformers.models.mixtral.modeling_mixtral because of the following error (look up to see its traceback): libcudart.so.12: cannot open shared object file: No such file or directory
1
#66 opened 5 months ago
by
MukeshSharma
Model not loading, even with 4-bit quantization
1
#65 opened 5 months ago
by
soumodeep-semut
did Mixtral start from Mistral or from-scratch?
1
#64 opened 5 months ago
by
DaehanKim
How many GPUs do we need to run this out of box?
3
#63 opened 5 months ago
by
kz919
Is this model can choose expert for every token? Or just choose two expert for a input
#62 opened 5 months ago
by
PandaMaster
AutoTokenizer.from_pretrained show OSError
1
#61 opened 5 months ago
by
sean29
does file with .safetensors necessary for continue sft training?
#60 opened 5 months ago
by
hegang126
Incomplete Answers
7
#59 opened 5 months ago
by
samparksoftwares
How can we enable continuous learning with the LLM model ?
#58 opened 5 months ago
by
Tapendra
Inference generation extremely slow
6
#57 opened 6 months ago
by
aledane
Optimizing Mixtral-8x7B-Instruct-v0.1 for Hugging Face Chat
1
#54 opened 6 months ago
by
Husain
SageMaker Deployment Error
11
#53 opened 6 months ago
by
seabasshn
killed on Loading checkpoint shards
1
#52 opened 6 months ago
by
asmatveev
Playground?
1
#51 opened 6 months ago
by
pbourmeau
vectorstore
3
#50 opened 6 months ago
by
philgrey
Enable inference API
2
#49 opened 6 months ago
by
mrfakename
How to use consolidated.xx.pt?
1
#47 opened 6 months ago
by
Wan62
Model not loading and not printing any error message
2
#45 opened 6 months ago
by
robotrage
open weights???
2
#43 opened 6 months ago
by
alanchan808
Prompt Template for RAG
1
#42 opened 6 months ago
by
mox
there is no sliding_window in params.json
1
#41 opened 6 months ago
by
Moses25
Question on further fine-tuning
#40 opened 6 months ago
by
vldvasi
KeyError: 'mixtral' when following the 'run the model' section
2
#39 opened 6 months ago
by
obiwan92
Update config.json
#38 opened 6 months ago
by
medmac01
sliding_window
5
#37 opened 6 months ago
by
issa130
Update config.json
#36 opened 6 months ago
by
issa130
Fix regression
2
#34 opened 6 months ago
by
TimeRobber
SageMaker generation speed, timeouts
1
#33 opened 6 months ago
by
elanmarkowitz
System Prompt Template
#32 opened 6 months ago
by
efei
Update config.json
2
#31 opened 6 months ago
by
tourist800
Facing issue while loading model - KeyError: 'mixtral'
7
#30 opened 6 months ago
by
swapnil3597
Update README.md
#29 opened 6 months ago
by
LPFLEO
Text Generation Inference?
3
#28 opened 6 months ago
by
silvacarl
ValueError: Error raised by inference API: Model is overloaded
2
#25 opened 6 months ago
by
alemaooo
Increase `sliding_window` to 32k
1
#24 opened 6 months ago
by
alpindale
Intuition for quality decrease after quantization
4
#23 opened 6 months ago
by
krumeto
Test Mixtral Model loading ERROR!
3
#21 opened 6 months ago
by
ExceedZhang
Discuss benefits of this work
1
#20 opened 6 months ago
by
Starlento
Mixtral not generating anything for some prompts
4
#19 opened 6 months ago
by
csgxy2022
Inquiry about Generation Speed
5
#17 opened 6 months ago
by
Boyue27
Error when deploying sagemaker endpoint: Unsupported model type mixtral
7
#16 opened 6 months ago
by
harryneal
GPU requirements
5
#14 opened 6 months ago
by
YorelNation
🚀 Torrent File for AI Model Download 🚀
#12 opened 6 months ago
by
Nondzu