I'm Indrajeet Aditya Roy, a full-stack software engineer and graduate student passionate about building innovative scalable solutions, solving complex problems, gaining professional experience, and collaborating with people. I obtained a Bachelor of Science in Software Engineering with a concentration in Data Science from Iowa State University in 2022. I obtained a Master of Science in Computer Engineering specializing in Artificial Intelligence and Machine Learning from Northeastern University in 2025. I am currently pursuing a graduate degree in Applied Mathematics from Northeastern University.
Anonymous communication networks like Tor face critical security vulnerabilities from Sybil and Cellflood attacks. Sybil attacks exploit decentralized trust models by deploying multiple malicious nodes to gain disproportionate network influence, while Cellflood attacks leverage the computational asymmetry between encryption and decryption to overwhelm relay nodes with circuit creation requests. We propose a multi-layered defense framework incorporating enhanced Directory Authority monitoring, IP-based admission control, and client-side cryptographic puzzles. Our approach achieves significant resistance to both attack vectors while preserving the network's fundamental anonymity guarantees and decentralized architecture. Experimental analysis demonstrates that these countermeasures maintain acceptable performance overhead while substantially increasing attack costs.
Document summarization systems face significant challenges from data scarcity, computational constraints, and hallucination in generated outputs. We present a hybrid architecture combining parameter-efficient fine-tuning of large language models with retrieval-augmented generation for scholarly document summarization. Our approach applies LoRA and PEFT to Mistral 7B, achieving effective adaptation with minimal training data, while a RAG pipeline utilizing Neo4j knowledge graphs and FAISS vector stores provides dynamic factual grounding. Comparative evaluation against bidirectional LSTM baselines demonstrates superior performance in coherence, factual accuracy, and cross-domain generalization. The integrated system successfully mitigates hallucination while maintaining contextual fidelity, demonstrating the viability of modern LLM-based approaches in resource-constrained academic settings.
We present a comprehensive survey of vision-language integration in large language models, tracing the architectural evolution from early dual-encoder systems to contemporary unified frameworks. This work systematically analyzes key developments across three dimensions: architectural design patterns for cross-modal alignment, training paradigms including contrastive learning and generative pre-training, and evaluation methodologies for vision-language tasks. We examine representative models spanning the progression from CLIP and ALIGN to recent instruction-tuned multimodal LLMs, identifying critical design choices and performance tradeoffs. The survey synthesizes current understanding of effective vision-language integration and highlights open challenges in scaling, grounding, and compositional reasoning.
Bird strikes pose significant risks to aviation safety, yet existing assessment frameworks fail to adequately quantify uncertainty in risk predictions. We present a Bayesian simulation framework that propagates uncertainty from raw observations through to trajectory-level risk assessments. The system combines Kalman filtering with seasonal behavioral priors for bird tracking, Beta-Binomial conjugate models with Monte Carlo sampling for parameter estimation, and Gaussian Process regression with composite kernels for spatial risk modeling. Our framework generates posterior predictive distributions enabling exceedance probability calculations along flight corridors. Simulation results demonstrate effective risk differentiation across trajectories with quantified uncertainty bounds that decrease as observational evidence accumulates. This probabilistic approach provides rigorous foundations for uncertainty-aware decision support in aviation safety applications.
Mobile edge computing introduces an intermediate computational tier between mobile devices and cloud infrastructure, enabling new tradeoffs in task scheduling for energy-constrained applications. We extend existing mobile cloud computing frameworks with a three-tier architecture incorporating dynamic resource models for battery state, workload-dependent power consumption, and time-varying network conditions. Applications are represented as DAGs with tasks characterized by computational and communication requirements. Our two-phase scheduling approach applies HEFT-based algorithms for makespan minimization, followed by energy optimization via heuristic task migration or Q-learning. Experimental evaluation demonstrates that the three-tier architecture achieves significant energy reduction compared to traditional mobile-cloud systems while satisfying deadline constraints across diverse device states and network conditions.
Application of Natural Language Processing (NLP) techniques to improve the translation accuracy of ancient and low-resource languages, specifically focusing on two primary technical challenges: the development of classification-translation models that facilitate transliterations between Sumerian and English, and a Named Entity Recognition (NER) model designed to bridge lexical gaps in English originating from the unique properties of Sumerian-specific proper nouns. The project aims to enhance translational accuracy and enrich the semantic understanding of translated texts.
View PaperImplementation of a custom TF-IDF vectorization pipeline that enhances sentiment classification through optimized preprocessing techniques. The system implements frequency-based feature selection with information gain thresholds, context-aware n-gram generation (n=1,2,3), and adaptive stopword filtering. Key technical innovations include a hashmap-based term frequency calculator that reduces computational complexity from O(n²) to O(n), and a modified IDF weighting scheme incorporating corpus-specific entropy measurements. Evaluated on the IMDb dataset, the custom implementation demonstrated a 4.2% accuracy improvement over scikit-learn's TfidfVectorizer baseline, with particular gains in identifying nuanced sentiment expressions and negation patterns. The modular architecture allows for configuration-driven feature extraction without retraining underlying models.
Implementation of a hybrid OLTP/OLAP data warehouse utilizing SAX parsers for efficient XML processing with O(1) memory complexity regardless of document size. The architecture employs a custom metadata-driven ETL pipeline with XPath query optimization for hierarchical XML traversal, reducing extract processing time by 68% compared to DOM parsing approaches. Technical implementation includes a snowflake schema database design with slowly changing dimensions (SCD Type 2) for tracking historical sales representative activities, materialized views for aggregation acceleration, and MySQL stored procedures for transactional consistency. Performance optimizations include parallel data loading with configurable thread pools, XML schema validation with custom DTD constraints, and bitmap indexing on high-cardinality attributes. The dual-purpose warehouse architecture utilizes temporal partitioning and columnar compression techniques, achieving sub-second query response for 92% of OLTP operations while simultaneously supporting complex OLAP aggregations with 3.2TB of historical pharmaceutical sales data spanning 5 years of operations.
A cross platform file system abstraction tool implementing a custom B-tree indexing layer with O(log n) query performance. The architecture features a platform-agnostic core that translates between POSIX and Win32 filesystem APIs through an adapter pattern, enabling unified XPath-inspired query execution across heterogeneous systems. Technical implementation includes memory-mapped file access for high-throughput operations, lazy-loading directory traversal to minimize memory footprint, and a custom binary serialization protocol reducing metadata overhead by 78% compared to JSON representation. The query engine supports complex boolean expressions with wildcard pattern matching and supports attribute-based filtering on extended file properties. Benchmarks demonstrated 3.2x faster recursive search performance compared to native OS utilities, with 40% reduced memory consumption during large directory traversals.
Implementation and evaluation of the K-Means clustering algorithm for analyzing text data within the 20 Newsgroups dataset. By utilizing a Term Frequency-Inverse Document Frequency (TF-IDF) vectorization approach, the implementation aims to categorize text documents into clusters, enhancing the understanding of text content similarities. The custom algorithm implementation is benchmarked against the standard Scikit-learn KMeans to assess performance metrics such as Homogeneity, Completeness, V-measure, Adjusted Rand Index, and Silhouette Coefficient. These metrics are crucial for evaluating the custom model's effectiveness in handling complex textual data, with a focus on optimizing clustering quality and execution efficiency.
A C++ implementation of the theoretical framework presented in the Energy and Performance-Aware Task Scheduling in a Mobile Cloud Computing Environment paper. The implementation constructs a Directed Acyclic Graph (DAG) of tasks and assigns local and cloud execution times.The implementation performs an initial phase of task assignment and prioritization to determine an initial schedule that minimizes total completion time, adhering to the paper's outlined equations and heuristics for computing tasks' priority scores and selecting their initial execution units. Once the baseline schedule is established, the code systematically explores migrating tasks between local cores and the cloud to achieve better energy efficiency. Each candidate migration scenario triggers a linear-time kernel rescheduling algorithm, recomputing all task start and finish times while ensuring all precedence constraints and resource capacities remain satisfied. After evaluating the new completion time and total energy consumption resulting from each migration, the code retains changes that reduce energy without exceeding permissible completion time thresholds. This iterative optimization loop directly mirrors the paper's approach to jointly optimizing both performance (makespan) and energy consumption in a mobile cloud computing environment.
GITHUBThis simulation explores the implementation of Q-Learning, SARSA, and Actor-Critic algorithms in a complex stochastic grid-based maze environment, featuring dynamic challenges like walls, bumps, and oil slicks. The maze consists of 248 navigable cells. Each algorithm is evaluated over 10 independent runs, each encompassing 1,000 episodes, aiming to optimize paths and maximize rewards within a defined penalty and reward structure. This testing framework assesses the efficacy of each algorithm in achieving the highest cumulative rewards and successfully navigating to the goal optimally.
View ReportThe p53-Mdm2 negative feedback loop is crucial for cell cycle regulation and tumor suppression, playing a significant role in advancing cancer therapy and developing new treatment strategies. This project aims to employ reinforcement learning to finely adjust the activation of this network's key components, specifically ATM, p53, Wip1, and MDM2. The primary objective is to establish an optimal control policy that effectively manages the p53-Mdm2 feedback loop, maximizing the therapeutic efficacy of maintaining an active p53 pathway. By achieving and sustaining this optimal state, the implementation offers a novel approach to controlling complex cellular networks, potentially inhibiting tumor growth and enhancing cellular repair mechanisms.
Implementation of a Bayesian-optimized MLP architecture for minimum-error classification of multivariate random vectors with non-Gaussian distributions. The system employs custom regularization techniques combining L2 weight decay with Bayesian priors to approximate posterior probability distributions with 97.2% accuracy on test data. Technical innovations include: implementation of an adaptive learning rate scheduler based on validation loss plateaus, conditional batch normalization for non-IID data samples, and an entropy-weighted loss function that penalizes misclassifications proportionally to their posterior probability differences. Architecture optimization utilized a parameterized grid search across 128 model configurations with 5-fold cross-validation, identifying that a single hidden layer with 42 neurons achieved the optimal bias-variance tradeoff. The final model implements a decision-theoretic MAP classifier with Bayes risk minimization, achieving classification error rates within 0.3% of the theoretical Bayes error limit for the given dataset distribution.
Implementation of an ETL pipeline to efficiently process and transform CSV files into data points and features. Responsive to API calls, the system methodically extracts data from the CSV files, applies necessary transformations, and uploads the processed data to a relational database. This pipeline facilitates dynamic data management, supports data analysis and also provides scope for extended API integration.
Implementation of Maximum A Posteriori (MAP) estimation to enhance vehicle localization accuracy in 2D environments. The Bayesian model integrates range measurements to multiple landmarks with existing positional data, allowing for a highly accurate estimation of the vehicle's most probable location. By utilizing MAP estimation, the model effectively addresses noisy distance measurements and prior knowledge of the vehicle's position, ensuring precise navigation despite environmental complexities and measurement inaccuracies.
GITHUBImplementation of a game engine for the classic Snake game developed using VueJS and complemented by vanilla HTML and CSS.
Vue.js streamlines state management and facilitates real-time rendering of the game environment, ensuring dynamic updates are efficiently handled without reloading the browser. Utilized to implement dynamic management of the game state, including the snake's movement, collision detection, and score tracking. HTML is used to structure the core layout of the game, setting up a grid-based playing field where the snake moves. CSS is utilized to style the game board and snake elements implemeting color schemes, borders, and animations.
Implementation of a game engine for the classic Tic-tac-toe game developed using Vanilla JavaScript, HTML and CSS.
The game engine implements the Minimax algorithm to simulate intelligent opponent moves and strategize optimal responses using JavaScript for the game logic, HTML for structuring the game UI, and CSS for response design, styling and animations. The Minimax algorithm extends the game's interactivity by enabling the system opponent to make decisions that are strategically sound, mimicking human-like decision-making processes adding a challenge and realism to the game engine.
Sound processing and song note compilation platform for real-time audio signal processing and synthesis developed using C and C++. It enables the generation of various waveforms—including Sine, Square, Triangle, and Sawtooth—through user-adjustable parameters such as sample rate, frequency, and duration.
The platform supports complex audio signal layering, allowing users to overlay multiple audio signals with precision. This includes detailed amplitude control and modulation, where the amplitude of one signal can dynamically influence others. Advanced digital filters, such as a digital reverb filter, create short-lived echoes to simulate different acoustic environments effectively.
An integral component of the system is its song player functionality, implemented using C to optimize performance. It decodes various audio file formats, arranging the decoded samples into a seamless playback sequence that can be further manipulated for sound design or educational purposes. This functionality highlights the system's capability to handle complex audio processing tasks efficiently.
Summer 2020 Internship Project
Android companion app to interface with the UV light hardware device component of the Purple-Bug UV ride-share cleaning product.
The Android mobile app was developed using Java and Kotlin, providing a stable and responsive UI. Network communications were implemented using Retrofit, which handled data transmission between the mobile app and backend services effectively. Image loading and caching were optimized using Glide to enhance performance and user experience.
Local data storage was implemented using Room, which offers an abstraction layer over SQLite, simplifying database operations. Complex asynchronous tasks were managed using RxJava and Kotlin Coroutines, ensuring smooth operation of the app by executing background tasks without interfering with the user interface.