Research Buddy V1.1

Motivation

The Research Buddy project was born out of a personal need. After stepping away from research for some time, I recently resumed working in this space and immediately recognized the need for a system to simplify finding papers aligned with my interests. The starting point was an ML-based classifier for paper abstracts, giving me an immediate hint about a paper’s field before I invested time reading it in depth.

However, my requirements soon expanded. I realized that beyond classification, I also needed summarization and keyword extraction—tools that could provide quick insights into a paper’s content. While free tiers of established LLMs such as Perplexity, ChatGPT, or Gemini exist, they come with limitations: they are fragmented across platforms, and often behind paywalls for extended use. What I sought was an integrated system, built for me, all in one place, without additional cost.

This motivation pushed me to leverage my system design and full-stack development skills to create a modular full-stack ML/AI application.

Research Buddy v1.1 — Current Capabilities

The first official release (v1.1) of Research Buddy is capable of the following:

Demo Video: Research Buddy v1.1 in action

Case Study: Model Training and Dataset

For this project, I initially started with the entire arXiv dataset. However, due to the very limited computational capacity of my laptop and the restrictions of free-tier cloud resources, training quickly became impractical and nearly impossible. This resource bottleneck was a major pain point of the project.

To mitigate this, I created a Computer Science–specific subset of the arXiv data, focusing on five categories most closely aligned with my research interests:

This reduced dataset made it feasible to train multiple models and run comparative evaluations.

Model Comparison Evaluation

To illustrate, a randomly selected abstract from the dataset (Index 387, true category cs.CL) was classified across eight different models.

Abstract (excerpt):
Current task-oriented dialog (TOD) systems mostly manage structured knowledge (e.g. databases and tables) to guide goal-oriented conversations. However, they fall short of handling dialogs which also involve unstructured knowledge...

True Category: cs.CL

Model Predictions:

Model Prediction Confidence Notes
SVM cs.CL N/A Correct prediction; lacks probability output
Multinomial NB cs.CL 0.8139 Correct; strong performance with sparse token counts
Random Forest cs.CL 0.8700 Correct; solid probabilistic confidence
AdaBoost cs.CL 0.2036 Correct but very low confidence; consistent with log-odds limitations
KNN cs.CL 0.9412 Correct; very confident, strong discriminative power in TF-IDF space
Feedforward NN cs.CL 0.9945 Correct; highest confidence; dense input works effectively
XGBoost cs.RO 0.4780 Incorrect; confidence near random threshold
BiLSTM cs.CL 0.8149 Correct; captures token context effectively

Observations

Conclusion

This evaluation demonstrates that even with a reduced dataset, the system effectively serves its intended purpose: providing quick, comparative classification results to guide paper exploration. However, to train a generalized, field-wide classifier, much larger resources and more comprehensive data would be necessary.

Design Principles

Technology Stack

Research Buddy v1.1 is built on a modern, modular, and cost-conscious technology stack:

Rationale:
This stack balances low cost, flexibility, and scalability. Free-tier services (Gemini, Colab, Kaggle) lower experimentation costs, while MLflow and modular APIs ensure smooth upgrades when swapping or retraining models. AWS S3 + CloudFront provide reliable content delivery, while MongoDB ensures dynamic user data handling.

Research Buddy Architecture
Figure 1: Research Buddy v1.1 architecture showing backend, frontend, and ML/LLM pipelines

Future Development