Data Expert with 15+ years of hands-on experience in designing, developing, and optimizing enterprise-scale analytical systems. Proven record of leading full-cycle DWH projects — from architecture and ETL to visualization and performance tuning. I build systems myself, not just design them on paper — combining deep technical skill with architectural vision to turn complex data into clear business value.
In recent years, I have focused on integrating Artificial Intelligence into analytical and business systems. I design and deploy LLM-powered architectures with multi-agent orchestration, where roles are clearly separated: Architect, Developer, Tester, DevOps, InfoSec, and an SRE agent responsible for incidents and alerts. The AI layer operates on top of the enterprise data warehouse — autonomously answering managers questions, replacing routine analyst tasks, monitoring KPIs, and signaling important deviations in real time.
I take a holistic approach to building analytical systems: I always deliver complete solutions, not just proofs of concept. Every project is unique — designed from the ground up with respect to its business goals, budget, and infrastructure. I’m deeply involved in each step — from code and infrastructure to optimization and visualization. This website itself runs on the data center I personally built, with all its code, deployment, and automation developed entirely by me. I’m passionate about performance tuning, elegant data models, and creating systems that stay fast, stable, and transparent even under heavy load.
I document both your current system and the optimal architecture for your needs, including a detailed migration plan. Designs are confirmed through diagrams before any implementation begins.
I create data warehouses optimized for your requirements. This could be a lightweight database on a single VM or a large-scale, distributed cluster architecture spanning multiple locations.
I turn your data into actionable insights. Data can come from any source — public datasets, Excel, production databases, APIs, or real-time queues. I create dashboards and visualizations that are intuitive, interactive, and mobile-friendly.
I design, optimize, and maintain databases for your workload. From relational to NoSQL and columnar systems, I ensure reliability, scalability, and high performance with advanced query optimization.
I build and maintain automated infrastructure to keep your data systems running smoothly. From containerization and CI/CD to monitoring and security, I ensure stable deployments and seamless scaling.
I create detailed project plans, find and interview top developers, lead the team, monitor progress, and make adjustments throughout development to ensure successful delivery.
Design and integrate AI modules for analytical systems — from RAG and multi-agent orchestration to monitoring and cost control.
Broad technical expertise and a strong drive to learn new technologies enable me to quickly adapt to any challenge. Сonstantly exploring new approaches and tools to deliver effective solutions.
Enterprise DWH built
Of data stored
Analytical data delay
Team raised
2023-Present (Strive (Data analyzing))
My personal achievements:
• Launched a fully operational automated trading system that generates real revenue. • Created a highly scalable architecture capable of processing real-time market data and news sentiment simultaneously. • Developed the system as a foundation for future personal projects and analytics tools. • Integrated AI modules that analyze unstructured financial data and news sentiment with RAG-based retrieval. • Successfully launched and maintained autonomous AI workflows, combining trading logic and generative insights.
My responsibilities:
• Designed the complete system architecture and implemented trading algorithms from scratch. • Built a real-time trading bot analyzing stock market data and news sentiment. • Developed all data pipelines, ETL processes, and analytics infrastructure personally, with no external developers involved. • Managed deployment, monitoring, and performance tuning to ensure reliable operation in production. • Implemented AI-based analysis modules powered by LLMs for data pattern recognition, anomaly detection, and sentiment evaluation. • Built multi-agent orchestration pipelines automating research, coding, testing, and deployment tasks.
Project technologies:
• Data Sources: MySQL, External APIs • ETL: Python, Argo-Workflows • DWH: HPE Vertica, Kimball model • Visualization: Superset, Chart.js
2020-2024 (Sravni.ru (Web))
My personal achievements:
• Built the entire system from scratch, ensuring stable operation. • Designed end-to-end architecture from products to analytics users. • Negotiated optimal pricing for all solution components. • Led migration from legacy slow DWH to new high-performance system. • Developed and enforced company-wide data workflow policies. • Deployed an internal AI assistant that automated analytical requests using RAG and prompt-engineering. • Reduced report preparation time by over 50% and improved decision-making latency company-wide.
My responsibilities:
• Selected and deployed servers and software solutions across the company. • Designed and implemented company-wide data flow architecture, with a focus on DWH. • Defined and enforced data modeling standards and rules. • Monitored and prevented personal data leaks, ensuring data security compliance. • Researched and integrated LLM-based assistants for analytics team — automating data exploration, insight generation, and documentation. • Introduced AI-driven anomaly detection and SQL suggestion modules, improving report delivery speed and accuracy.
Project technologies:
• Data Sources: MSSQL, PostgreSQL, MySQL, MongoDB, External APIs, RabbitMQ, Kafka, CSV, JSON • ETL: Python, TeamCity, Prefect, Kafka, Debezium • DWH: Snowflake, Kimball model • Visualization: PowerBI, Snowsight, Sigma, Grafana
2016-2020 (SuperJob (Web))
My personal achievements:
• Independently developed a stable analytical system from scratch. • Optimized budget and resources by sizing licenses, disk space, and designing a container-based test environment on 2 hosts. • Developed ETL pipelines using Pentaho and Python. • Built an autonomous auto-testing system requiring no manual intervention. • Delivered a Big Data solution covering the full range of analytical tasks. • Designed a Data Vault model enabling fast, high-quality business insights, including data science, ranking, scoring, and MDM. • Collaborated with analysts to release real-time dashboards on centralized storage. • Implemented a 3NF metadata model for centralized data flow coordination and quality monitoring.
My responsibilities:
• Independently built and maintained a fully functional analytical system from the ground up. • Selected and implemented servers and software solutions to optimize performance and cost. • Designed and maintained the Data Vault model and ETL metadata model using PowerDesigner. • Administered an HP Vertica cluster on *nix servers, expanded the cluster, and balanced loads to maintain minimal data latency, ensuring no user requests queued even at peak delays of 40–50 minutes.
Project technologies:
• Data Sources: MariaDB, MongoDB, CSV, JSON, External APIs • ETL: Pentaho DI, Python, Jenkins, Airflow • DWH: Vertica+Clickhouse, Data Vault model • Visualization: Tableau, Grafana
2014-2016 (Svyaznoy (Retail))
My personal achievements:
• Created a unique adaptive data model for a Big Data warehouse using a cluster-based approach. • Developed Python-based automated tests covering all ETL processes. • Designed and implemented a Data Quality (DQ) system that caught most bugs before deployment, ensuring high reliability of ETL processes.
My responsibilities:
• Led the end-to-end development of a large-scale data warehouse project. • Communicated directly with customers to gather requirements and feedback. • Built and implemented data models. • Planned and coordinated team activities, ensuring efficient workflow. • Refactored and validated the majority of ETL processes.
Project technologies:
• Data Sources: Hadoop, Oracle, MSSQL, PostgreSQL, MySQL • ETL: Pentaho DI, Python, Sqoop, Kafka, Oozie • DWH: Hive, Vertica, EAV
2012-2014 (Asseco group, R-Style SoftLab (IT consulting))
My personal achievements:
• Integrated and completed the Data Warehouse (DWH) system in several top-20 Russian banks. • Developed management reporting and IFRS-compliant reports. • Delivered the final data visualization for end users.
My responsibilities:
• Developed multiple large-scale DWH projects from scratch for major Russian banks. • Delivered end-to-end solutions: sales support, analytics, ETL, DWH, reporting, and client acceptance. • Implemented Oracle DWH (RSDH) and integrated it with existing bank systems. • Built ETL pipelines, OLAP reporting, and data validation processes. • Conducted operational training and certification programs for new team members. • Supported analysts with Oracle and RSDH, ensuring reliable reporting for banking operations.
Project technologies:
• Data Sources: Oracle, MSSQL, PostgreSQL, MySQL • ETL: Informatica powercenter • DWH: Oracle Exadata, 3NF Model
2011-2012 (TNS Global (Research))
My personal achievements:
• Designed real-time data quality reports and alerts, highlighting key system issues instantly. • Created simulation schedules for a web-monitoring robot, modeling user behavior by age, location, and gender. • Developed and launched a custom Reporting Service in-house, enabling accurate and timely analytics.
My responsibilities:
• Built and managed an OLAP warehouse from scratch, ensuring high-speed updates of large cubes. • Developed automated data aggregation and reporting processes, improving data accuracy and timeliness. • Optimized SQL procedures, jobs, and triggers for maximum performance. • Implemented SSAS and SSRS solutions, streamlining statistical reporting and analytics.
Project technologies:
• MSSQL, SSMS, SSIS, SSAS
2009-2010 (Allianz, Rosno (Insurance))
My personal achievements:
• Successfully migrated complex datasets from multiple systems into SAP with full data integrity. • Optimized SAP module performance, accelerating operational reporting. • Independently managed the entire migration workflow from extraction to validation.
My responsibilities:
• Created technical specifications and calculation algorithms for SAP modules. • Tested and optimized SAP module performance. • Exported and corrected data from INFIN, CIS, and other corporate systems to ensure smooth migration into SAP. • Prepared and validated data for migration, ensuring both accuracy and completeness.
Project technologies:
• SAP, Oracle, Microsoft Office, SQL, VBA
Strive (Seoul, South Korea)
Developed and deployed a fully automated trading and analytics platform combining real-time data streaming, CDC, and AI-driven insights. The system leverages LLMs for sentiment analysis and anomaly detection, and runs on a Kubernetes-based distributed infrastructure.
Project technologies:
• Data Sources: MySQL, External APIs • ETL: Python, Argo Workflows, Kafka • DWH: HPE Vertica, Kimball model • Visualization: Superset, Chart.js • Hosting: My own data center
Totle (Tel Aviv, Israel)
Personally developed a high-performance API for real-time Ethereum data analytics, querying big data in milliseconds. Acted as a technical co-founder and lead engineer for the project, which later became the starting point for my personal project Strive.
Project technologies:
• Data Sources: Blockchain • ETL: Node.js • DWH: Snowflake, 3NF model • Visualization: d3.js, Node.js API • Hosting: AWS
OpenHub (Davao, Philippines)
OpenHub is a multi-location data center with two locations in Russia and one in the Philippines. Starting from a small GPU-farm with over 100 graphics cards, it has grown into a full-scale data center hosting high-performance analytical and AI inference servers. I personally built and configured all servers, connecting them into a single high-performance network across all sites.
Project technologies:
• Hardware: Supermicro, Intel, RAID, Cisco, Mikrotik • Software: Windows Hyper-V, Ubuntu • Visualization: Grafana, UptimeRobot • Hosting: My own data center
Sravni.ru (Moscow, RF)
Personally developed the company-wide data streaming system into DWH, implementing real-time streaming and CDC. Built the architecture on Kubernetes and introduced a data mesh approach to microservices, minimizing team size by distributing responsibilities to products.
Project technologies:
• Data Sources: MSSQL, PostgreSQL, MySQL, MongoDB, External APIs, RabbitMQ, Kafka, CSV, JSON • ETL: Python, TeamCity, Prefect-Kubernetes, Kafka, Debezium • DWH: Snowflake, Kimball model • Visualization: PowerBI, Snowsight, Superset, Grafana • Hosting: 100% cloud managed
PharmaKey (Moscow, RF)
Completely redesigned and implemented a new pharmacy analytics platform and continue collaborating as CTO. The system supports 10+ clients, each processing over a billion transactions, providing comprehensive production and sales insights nationwide.
Project technologies:
• Portal backend: PHP, JavaScript, MySql Data Sources: MySQL, CSV, External APIs ETL: Migrated from Rundeck and Bash to Pentaho DI, Python, Airflow • DWH: Migrated from MySql and Redshift to Vertica, Snowflake model • Visualization: Tableau, Metabase • Hosting: Migrated from AWS and Hetzner to Selectel dedicated+vmWare
SuperJob (Moscow, RF)
Personally developed the DWH model and data-driven system as Project Owner, supporting 1 billion weekly events and 100,000 dictionary updates per hour, integrating more than 20 diverse data sources to enable enterprise-wide analytics.
Project technologies:
• Data Sources: MariaDB, MongoDB, CSV, JSON, External APIs • ETL: Pentaho DI, Python, Jenkins, Airflow • DWH: Vertica+Clickhouse, real-time ods + Data Vault model + near real-time star • Visualization: Tableau, Grafana • Hosting: Enterprise dedicated
PetroSoft (Pittsburgh, USA)
Successfully designed and implemented over 30 real-time dashboards, handling 1 billion weekly transactions across petrol stations in the USA and Canada, enabling rapid and accurate business insights.
Project technologies:
• Data Sources: MySQL, Enterprise Bus • ETL: Pentaho DI, Python, Jenkins • DWH: Vertica, Snowflake model • Visualization: Tableau, JavaScript • Hosting: Enterprise dedicated
Rafferi (San Francisco, USA)
Delivered an ETL and database cluster solution, integrating data from Amazon, eBay, and Walmart, which enabled fast, accurate dashboards and actionable e-commerce insights.
Project technologies:
• Data Sources: AWS API, MWS API, eBay API • ETL: Pentaho DI, Java • DWH: Vertica, Star model • Visualization: d3.js, Node.js API • Hosting: My own data center
Russia’s oldest and most prestigious university, consistently ranked among the world’s top universities. Admitted to a highly competitive state-funded (budget) place.
Russia’s leading technical university, internationally recognized for producing most strongest in the world engineers. Admitted without entrance exams after winning the Moscow Physics and Mathematics Olympiad.
Finished the course and successfully passed a personal examination with an MIT professor, validating subject mastery.

