Volgen
Nikita Balagansky
Nikita Balagansky
Tinkoff
Geverifieerd e-mailadres voor tinkoff.ru
Titel
Geciteerd door
Geciteerd door
Jaar
PALBERT: Teaching ALBERT to Ponder
N Balagansky, D Gavrilov
NeurIPS 2022 35, 14002--14012, 2022
42022
Democratized Diffusion Language Model
SMLC Vaina, N Balagansky, D Gavrilov
3*2023
Classifiers are better experts for controllable text generation
A Sitdikov, N Balagansky, D Gavrilov, A Markov
arXiv preprint arXiv:2205.07276, 2022
32022
Weight squeezing: Reparameterization for extreme compression and fast inference
C Artem, G Daniil, B Nikita, K Pavel
arXiv: 2010.06993, 2020
22020
Ahead-of-Time P-Tuning
D Gavrilov, N Balagansky
arXiv preprint arXiv:2305.10835, 2023
12023
Linear interpolation in parameter space is good enough for fine-tuned language models
M Rofin, N Balagansky, D Gavrilov
arXiv preprint arXiv:2211.12092, 2022
12022
Learn Your Reference Model for Real Good Alignment
A Gorbatovski, B Shaposhnikov, A Malakhov, N Surnachev, Y Aksenov, ...
arXiv preprint arXiv:2404.09656, 2024
2024
Linear Transformers with Learnable Kernel Functions are Better In-Context Models
Y Aksenov, N Balagansky, SMLC Vaina, B Shaposhnikov, A Gorbatovski, ...
arXiv preprint arXiv:2402.10644, 2024
2024
Diffusion Language Models Generation Can Be Halted Early
SM Lo Cicero Vaina, N Balagansky, D Gavrilov
arXiv e-prints, arXiv: 2305.10818, 2023
2023
Het systeem kan de bewerking nu niet uitvoeren. Probeer het later opnieuw.
Artikelen 1–9