Recurrent memory transformer A Bulatov, Y Kuratov, M Burtsev Advances in Neural Information Processing Systems 35, 11079-11091, 2022 | 80 | 2022 |
Scaling transformer to 1m tokens and beyond with rmt A Bulatov, Y Kuratov, Y Kapushev, MS Burtsev arXiv preprint arXiv:2304.11062, 2023 | 48 | 2023 |
In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs Miss Y Kuratov, A Bulatov, P Anokhin, D Sorokin, A Sorokin, M Burtsev arXiv preprint arXiv:2402.10790, 2024 | 3 | 2024 |
Beyond Attention: Breaking the Limits of Transformer Context Length with Recurrent Memory A Bulatov, Y Kuratov, Y Kapushev, M Burtsev Proceedings of the AAAI Conference on Artificial Intelligence 38 (16), 17700 …, 2024 | | 2024 |
Better Together: Enhancing Generative Knowledge Graph Completion with Language Models and Neighborhood Information A Chepurova, A Bulatov, Y Kuratov, M Burtsev arXiv preprint arXiv:2311.01326, 2023 | | 2023 |