Таблица скроллится влево
Задача | Результат | Метрика |
---|---|---|
LCS | 0.178 | Accuracy |
RCB | 0.598 / 0.603 | Avg. F1 / Accuracy |
USE | 0.266 | Grade Norm |
RWSD | 0.665 | Accuracy |
PARus | 0.884 | Accuracy |
ruTiE | 0.674 | Accuracy |
MultiQ | 0.247 / 0.171 | F1-score/EM |
CheGeKa | 0.05 / 0.022 | F1 / EM |
ruModAr | 0.949 | EM |
ruMultiAr | 0.337 | EM |
MathLogicQA | 0.589 | Accuracy |
ruWorldTree | 0.872 / 0.872 | Avg. F1 / Accuracy |
ruOpenBookQA | 0.813 / 0.813 | Avg. F1 / Accuracy |
Таблица скроллится влево
Задача | Результат | Метрика | ||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
BPS | 0.23 | Accuracy | ||||||||||||||||||||||||
ruMMLU | 0.704 | Accuracy | ||||||||||||||||||||||||
SimpleAr | 0.986 | EM | ||||||||||||||||||||||||
ruHumanEval | 0.023 / 0.113 / 0.226 | pass@k | ||||||||||||||||||||||||
ruHHH |
0.781
|
Accuracy | ||||||||||||||||||||||||
ruHateSpeech |
0.736
|
Accuracy | ||||||||||||||||||||||||
ruDetox |
|
Общая средняя оценка (J) Оценка сохранения смысла (SIM) Оценка натуральности (FL) Точность переноса стиля (STA) |
||||||||||||||||||||||||
ruEthics |
Результаты таблицы:
[[-0.368, -0.405
, -0.403, -0.309
, -0.335], |
5 MCC |
MTS AI
MTS AI Chat Medium
This model is a specific architecture stay tuned for the paper
This model is trained with SFT only
-
Stay tuned for the paper
Proprietary model developed by MTS AI
Code version v.1.1.0 All the parameters were not changed. Inference details: torch 2.0.0 + Cuda 11.7.
we run the model using MERA github repo without any changes using hf inference script