Dmitry V. Dylov
2025
CLARITY: Clinical Assistant for Routing, Inference, and Triage
Vladimir Shaposhnikov | Alexandr Nesterov | Ilia Kopanichuk | Ivan Bakulin | Zhelvakov Egor | Ruslan Abramov | Tsapieva Ekaterina Olegovna | Iaroslav Radionovich Bespalov | Dmitry V. Dylov | Ivan Oseledets
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Vladimir Shaposhnikov | Alexandr Nesterov | Ilia Kopanichuk | Ivan Bakulin | Zhelvakov Egor | Ruslan Abramov | Tsapieva Ekaterina Olegovna | Iaroslav Radionovich Bespalov | Dmitry V. Dylov | Ivan Oseledets
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
We present CLARITY (Clinical Assistant for Routing, Inference and Triage), an AI-driven platform designed to facilitate patient-to-specialist routing, clinical consultations, and severity assessment of patient conditions. Its hybrid architecture combines a Finite State Machine (FSM) for structured dialogue flows with collaborative agents that employ Large Language Model (LLM) to analyze symptoms and prioritize referrals to appropriate specialists. Built on a modular microservices framework, CLARITY ensures safe, efficient, and robust performance, flexible and readily scalable to meet the demands of existing workflows and IT solutions in healthcare.We report integration of our clinical assistant into a large-scale national interhospital platform, with more than 55,000 content-rich userdialogues completed within the two months of deployment, 2,500 of which were expert-annotated for subsequent validation. The validation results show that CLARITY surpasses human-level performance in terms of the first-attempt routing precision, naturally requiring up to 3 times shorter duration of the consultation than with a human.
2022
ALToolbox: A Set of Tools for Active Learning Annotation of Natural Language Texts
Akim Tsvigun | Leonid Sanochkin | Daniil Larionov | Gleb Kuzmin | Artem Vazhentsev | Ivan Lazichny | Nikita Khromov | Danil Kireev | Aleksandr Rubashevskii | Olga Shahmatova | Dmitry V. Dylov | Igor Galitskiy | Artem Shelmanov
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations
Akim Tsvigun | Leonid Sanochkin | Daniil Larionov | Gleb Kuzmin | Artem Vazhentsev | Ivan Lazichny | Nikita Khromov | Danil Kireev | Aleksandr Rubashevskii | Olga Shahmatova | Dmitry V. Dylov | Igor Galitskiy | Artem Shelmanov
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations
We present ALToolbox – an open-source framework for active learning (AL) annotation in natural language processing. Currently, the framework supports text classification, sequence tagging, and seq2seq tasks. Besides state-of-the-art query strategies, ALToolbox provides a set of tools that help to reduce computational overhead and duration of AL iterations and increase annotated data reusability. The framework aims to support data scientists and researchers by providing an easy-to-deploy GUI annotation tool directly in the Jupyter IDE and an extensible benchmark for novel AL methods. We prepare a small demonstration of ALToolbox capabilities available online. The code of the framework is published under the MIT license.
2021
Active Learning for Sequence Tagging with Deep Pre-trained Models and Bayesian Uncertainty Estimates
Artem Shelmanov | Dmitri Puzyrev | Lyubov Kupriyanova | Denis Belyakov | Daniil Larionov | Nikita Khromov | Olga Kozlova | Ekaterina Artemova | Dmitry V. Dylov | Alexander Panchenko
Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume
Artem Shelmanov | Dmitri Puzyrev | Lyubov Kupriyanova | Denis Belyakov | Daniil Larionov | Nikita Khromov | Olga Kozlova | Ekaterina Artemova | Dmitry V. Dylov | Alexander Panchenko
Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume
Annotating training data for sequence tagging of texts is usually very time-consuming. Recent advances in transfer learning for natural language processing in conjunction with active learning open the possibility to significantly reduce the necessary annotation budget. We are the first to thoroughly investigate this powerful combination for the sequence tagging task. We conduct an extensive empirical study of various Bayesian uncertainty estimation methods and Monte Carlo dropout options for deep pre-trained models in the active learning framework and find the best combinations for different types of models. Besides, we also demonstrate that to acquire instances during active learning, a full-size Transformer can be substituted with a distilled version, which yields better computational performance and reduces obstacles for applying deep active learning in practice.
Search
Fix author
Co-authors
- Nikita Khromov 2
- Daniil Larionov 2
- Artem Shelmanov 2
- Ruslan Abramov 1
- Ekaterina Artemova 1
- Ivan Bakulin 1
- Denis Belyakov 1
- Iaroslav Radionovich Bespalov 1
- Zhelvakov Egor 1
- Igor Galitskiy 1
- Danil Kireev 1
- Ilia Kopanichuk 1
- Olga Kozlova 1
- Lyubov Kupriyanova 1
- Gleb Kuzmin 1
- Ivan Lazichny 1
- Alexandr Nesterov 1
- Tsapieva Ekaterina Olegovna 1
- Ivan Oseledets 1
- Alexander Panchenko 1
- Dmitri Puzyrev 1
- Aleksandr Rubashevskii 1
- Leonid Sanochkin 1
- Olga Shahmatova 1
- Vladimir Shaposhnikov 1
- Akim Tsvigun 1
- Artem Vazhentsev 1