Let's Connect

contact_info.sh

$ cat contact_details.txt

$ _

Core Competencies

  AI Research & Development

Leading innovation in:

  • Specialist Large Language Models (LLMs)
  • Master Document Understanding
  • Proficient Multi-Modal AI Systems

  Technical Leadership

Driving excellence through:

  • Lead Cross-functional R&D Team Management
  • Principal Project Architecture
  • Strategic Roadmap Planning

  System Architecture

Specialized in:

  • Master Agentic RAG Systems
  • Core Distributed Systems
  • Proficient Cloud Infrastructure

  Product Development

End-to-end expertise in:

  • Lead AI Product Strategy
  • Advanced Solution Design
  • Skilled Production Deployment

Research Timeline

2024 H2

LLM / Embedding Model Serving

Researched scalable and efficient methods for serving LLMs and embedding models in production environments to support high-throughput applications.

Tech Stack: Text Generation Inference, vLLM, SGLang, Text Embeddings Inference.

Multi-Agent RAG

Explored advanced agentic RAG frameworks for scalable tool usage, memory organization, query management, reflection, reasoning, and execution.

Tech Stack: MetaGPT, AutoGen.

LLM Architecture

Conducted a comprehensive survey of fundamental structures and design principles behind various LLM models.

Tech Stack: DeepSeek, LLaMA, Owen, Titans.

Graph-Based RAG

Researched graph-based retrieval paradigms and algorithms to enhance retrieval performance, exploring single- and multi-graph architectures, as well as layout-based and domain-entity graphs.

Tech Stack: Microsoft GraphRAG, Neo4j.

2024 H1

Document Structurization

Built an end-to-end solution to transform unstructured or semi-structured data into structured formats easily interpretable by LLMs. Examples include converting Excel/PDF files into JSON or Markdown.

Tech Stack: Transformers, Gradio, Docling, Unstructured-IO, LlamaParse, Azure AI Document Intelligence.

LLM Reasoning and Acting

Researched various LLM reasoning schemas with Human-in-the-loop approaches, including Chain-of-Thought (CoT), Tree-of-Thought (ToT), Self-Consistency, ReAct, and Retrieval-Augmented Workflow Optimization (ReWOO).

Tech Stack: LangGraph, LangSmith, LlamaIndex, Haystack.

Retrieval-Augmented Generation

Developed an LLM-powered chatbot utilizing Retrieval-Augmented Generation (RAG) to effectively leverage user-provided data for generating accurate and contextually grounded answers.

Tech Stack: Elasticsearch, Milvus, Kotaemon.

Trustworthy LLM

Investigated methods to reduce LLM hallucinations and enhance knowledge grounding for trustworthy AI outputs.

Tech Stack: TruLens, Ragas, DeepEval.

2023 H2

Tech Stack Management

Investigated effective approaches to organize and manage technical assets for streamlined workflows.

Tech Stack: Hugging Face Hub, Hugging Face Spaces.

Table Detection and Recognition

Researched methodologies for extracting table information, exploring specialized models for table detection and recognition.

Tech Stack: DETR, TATR.

ETL Pipeline

Designed an efficient ETL pipeline for performance tracking and continuous improvement of workflows.

Tech Stack: Apache Airflow, Docker Swarm, Kubernetes.

LLM Training

Researched techniques for training and serving large language models (LLMs), covering pretraining, supervised training, preference alignment (RLHF), quantization, and low-rank adaptation methods.

Tech Stack: PEFT, OpenAI, bitsandbytes, W&B, TorchTune, LangChain, Accelerate.

Multi-Label Classification

Researched lightweight CNN and ViT architectures to develop efficient multi-label classifiers for diverse datasets and tasks.

Tech Stack: Dino, PyTorch Image Models.

2023 H1

Lightweight Multimodal Architecture

Developed a CPU-compatible transformer-based architecture for information extraction, integrating image and text inputs while emphasizing spatial information learning. Achieved state-of-the-art performance among lightweight transformer models. Published two accepted papers at ICDAR 2024.

Tech Stack: Transformers, FAISS.

Text Recognition

Researched advanced OCR architectures, focusing on transformer-based encoder-decoder models with innovative image synthesis and augmentation techniques.

Tech Stack: PyTorch Lightning, Albumentations.

Training Strategies

Experimented with training techniques for CNN and transformer-based architectures, including Stochastic Depth, Teacher-Student Knowledge Distillation, Label Smoothing, and learning rate warm-up strategies.

Tech Stack: PyTorch Image Models (timm).

2022 H2

Document Object Detection

Experimented with anchor-based and anchor-free architectures for document object extraction.

Tech Stack: Ultralytics, MMDetection.

Sentence Embedding

Researched lightweight multilingual sentence embedding techniques for improved efficiency and scalability.

Tech Stack: Sentence Transformers (SBERT).

2022 H1

Image Registration

Researched local image feature matching using OpenCV-based methods (e.g., SIFT, ORB) and deep learning solutions (e.g., LoFTR).

Tech Stack: OpenCV, Kornia.

Cloud Deployment

Deployed an Intelligent Document Processing (IDP) pipeline on AWS, experimenting with distributed system designs for scalability and performance.

Tech Stack: Docker, AWS Services, Seldon, Celery, FastAPI, Uvicorn.

2021 H2

Key-Value Pair Extraction

Explored transformer-based architectures for key-value extraction, focusing on BERT-based models and their variations.

Tech Stack: PyTorch, Hugging Face Transformers.

2021 H1

Model Compression

Researched Knowledge Distillation and Post-Training Quantization (PTQ) for lightweight, high-performance models.

Tech Stack: OpenVINO, TorchScript.

Model Serving

Investigated efficient model serving methods using ONNX and TensorRT for optimized deployment.

Tech Stack: TensorRT, ONNX.

2020 H2

Key-Value Pair Extraction

Applied DGCNN and Residual Gated Graph ConvNets for key-value extraction. Enhanced spatial learning with CenterNET and explored metric learning techniques for embedding and clustering.

Tech Stack: DGL, PyTorch Hub, pytorch-metric-learning.

Sentence Segmentation

Developed strategies using Named Entity Recognition (NER) to segment sentences into meaningful pieces. Implemented GRU/LSTM with CRF loss for tokenization and tagging.

Tech Stack: pytorch-crf.

2020 H1

Node Classification

Investigated graph construction and applied Graph Attention Networks (GANs) for document node classification tasks.

Tech Stack: DGL.

Document Embedding

Researched graph-based solutions, including Graph Convolutional Networks (GCNs), for effective document embedding and representation.

Tech Stack: DGL.

2019 H2

Text Detection

Researched CNN backbones and loss functions for encoder-decoder architectures in text detection.

Tech Stack: Sagemaker, S3, TensorFlow, imgaug.

Text Recognition (OCR)

Improved CRNN+CTC architecture for unlimited-length OCR decoding.

Tech Stack: Sagemaker, S3, PyTorch.

2019 H1

Person Re-Identification

Researched face detection and recognition solutions integrated with human re-identification architectures to develop a surveillance application.

Tech Stack: FaceNet, R-CNN, DeepFace, MTCNN, SSD.