I am a Data Scientist and Machine Learning Engineer with over 3 years of experience specializing in Python, SQL, machine learning, natural language processing (NLP), and data pipelines. I have a strong background in building scalable AI-driven applications and analytics solutions that deliver measurable business impact. My expertise includes designing and deploying cloud-based machine learning workflows and production models that automate decision systems.
Throughout my career, I have developed reusable prompt templates, validation checks, and monitoring dashboards to improve model reliability and reduce errors. I am skilled in creating automated data pipelines and real-time dashboards that ensure data consistency and reduce debugging time. I have also designed gamification algorithms to enhance user engagement and conducted A/B testing to optimize feature performance.
My experience spans various industries, including startups, academia, and software development companies. I have automated data extraction pipelines, developed LLM-based financial summarization workflows, and built predictive demand forecasting models to optimize supply chains. Additionally, I have implemented ML inference APIs and CI/CD pipelines to streamline deployment and monitoring processes.
I hold a Master of Science degree in Computer Science from Texas A&M University, where I also contributed to projects involving sentiment analysis using BERT and image classification with CNNs. These projects have demonstrated my ability to improve accuracy and support significant revenue and cost savings.
I am passionate about leveraging advanced machine learning techniques and cloud technologies to solve complex problems and drive business growth. I am continuously expanding my skills through certifications in Generative AI, Large Language Models, Google BigQuery, and Python. I am eager to contribute my expertise to innovative teams and challenging projects.
Designed and implemented a Python-based LLM evaluation framework with structured test datasets and A/B prompt experimentation, improving model reliability by 15% and reducing customer-facing errors. Developed reusable prompt templates, validation checks, and AWS monitoring dashboards, enabling systematic tracking of model experiments via GitHub and improving detection of hallucination and response inconsistencies.
Designed automated data pipelines using Python, AWS S3, Lambda, Athena, and Glue and built real-time dashboards in Power BI, achieving 100% data consistency and reducing debugging time by 30%. Designed and implemented a gamification algorithm (XP, streak tracking) using Python, improving user engagement by 20%. Conducted A/B testing using statistical analysis in Python (SciPy, NumPy) to evaluate feature performance and improve event tracking accuracy, reducing reporting inconsistencies by 10%.
Automated SEC EDGAR data extraction pipelines using Python (Requests, BeautifulSoup, APIs, Pandas), building structured datasets for financial analysis and reporting. Developed LLM-based financial summarization workflows using OpenAI API and Python, generating automated insight reports that reduced manual analysis workload by 35% and achieved 95% stakeholder approval. Performed exploratory data analysis using Python (Pandas, NumPy, Matplotlib) and developed time-series forecasting models (Scikit-learn), improving distribution planning accuracy by 20% and increasing inventory efficiency by 15%. Designed predictive demand forecasting models to optimize pharmaceutical supply chains, improving resource allocation across 20 regions.
Developed ML-based document classification models (scikit-learn) to automatically tag financial documents, reducing manual sorting by 40% and improving document processing throughput by 30%. Built data pipelines for document ingestion, feature extraction, and model training integrating MongoDB, MariaDB, and AWS storage, enabling processing of 100K+ documents per month. Designed ML inference APIs using FastAPI and AWS Lambda enabling real-time document classification in loan processing systems, reducing latency by 25%. Implemented CI/CD pipelines using Docker, Kubernetes, and Jenkins to automate ML deployment and monitoring, reducing release cycles by 50%.
Jobicy
592 professionals pay to access exclusive and experimental features on Jobicy
Free
USD $0/month
For people just getting started
Plus
USD $8/month
Everything in Free, and: