<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en"><generator uri="https://jekyllrb.com/" version="4.4.1">Jekyll</generator><link href="https://michaelenny.github.io/michael-eniolade.github.io/feed.xml" rel="self" type="application/atom+xml"/><link href="https://michaelenny.github.io/michael-eniolade.github.io/" rel="alternate" type="text/html" hreflang="en"/><updated>2026-04-21T15:06:08+00:00</updated><id>https://michaelenny.github.io/michael-eniolade.github.io/feed.xml</id><title type="html">blank</title><subtitle>Software Engineer and Data Scientist with expertise in AI/ML, cloud-native solutions, and data pipelines. PhD student at University of the Cumberlands. </subtitle><entry><title type="html">Integrating LLMs into Electronic Health Records</title><link href="https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/llms-electronic-health-records/" rel="alternate" type="text/html" title="Integrating LLMs into Electronic Health Records"/><published>2026-03-25T10:00:00+00:00</published><updated>2026-03-25T10:00:00+00:00</updated><id>https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/llms-electronic-health-records</id><content type="html" xml:base="https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/llms-electronic-health-records/"><![CDATA[<p>Electronic Health Records (EHRs) have become the backbone of modern healthcare, yet they remain notoriously difficult for patients to understand and for clinicians to navigate efficiently. In my recent research, I explore how large language models can bridge this gap.</p> <h2 id="the-ehr-challenge">The EHR Challenge</h2> <p>Despite decades of digitization, EHR systems still suffer from:</p> <ul> <li><strong>Information overload</strong>: Clinicians spend more time on documentation than patient care</li> <li><strong>Poor patient comprehension</strong>: Medical records are filled with jargon that patients struggle to understand</li> <li><strong>Interoperability issues</strong>: Data silos prevent seamless information sharing between providers</li> </ul> <h2 id="llm-enhanced-ehr-solutions">LLM-Enhanced EHR Solutions</h2> <p>By integrating LLMs into EHR workflows, we can address these challenges through:</p> <h3 id="patient-facing-applications">Patient-Facing Applications</h3> <ul> <li><strong>Automated health summaries</strong> in plain language after each visit</li> <li><strong>Personalized health education</strong> based on individual diagnoses and treatment plans</li> <li><strong>Interactive Q&amp;A</strong> allowing patients to ask questions about their medical records</li> </ul> <h3 id="clinician-facing-applications">Clinician-Facing Applications</h3> <ul> <li><strong>Intelligent documentation assistance</strong> that drafts clinical notes from conversation</li> <li><strong>Decision support</strong> by surfacing relevant patient history and clinical guidelines</li> <li><strong>Cross-record synthesis</strong> that identifies patterns across a patient’s complete medical history</li> </ul> <h2 id="technical-considerations">Technical Considerations</h2> <p>Implementing LLMs in EHR systems requires careful attention to:</p> <ol> <li><strong>HIPAA compliance</strong> and data security</li> <li><strong>Model fine-tuning</strong> on medical terminology and clinical workflows</li> <li><strong>Audit trails</strong> for all AI-generated content</li> <li><strong>Human-in-the-loop</strong> validation for clinical decisions</li> </ol> <p>The future of healthcare informatics lies at the intersection of AI and clinical practice, and LLMs represent a powerful tool for making that intersection productive and patient-centered.</p>]]></content><author><name></name></author><category term="research"/><category term="LLMs"/><category term="healthcare"/><category term="data-engineering"/><summary type="html"><![CDATA[How large language models can enhance EHR systems for better patient engagement and clinical education.]]></summary></entry><entry><title type="html">Generative AI and LLMs in U.S. Healthcare: Opportunities and Challenges</title><link href="https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/generative-ai-healthcare/" rel="alternate" type="text/html" title="Generative AI and LLMs in U.S. Healthcare: Opportunities and Challenges"/><published>2026-03-15T10:00:00+00:00</published><updated>2026-03-15T10:00:00+00:00</updated><id>https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/generative-ai-healthcare</id><content type="html" xml:base="https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/generative-ai-healthcare/"><![CDATA[<p>The integration of generative AI and large language models (LLMs) into healthcare is rapidly transforming how we approach patient engagement, clinical decision support, and health policy development. In my recent research published in <em>Data Science and Big Data Analysis</em>, I examined the potential of these technologies to reshape the U.S. healthcare landscape.</p> <h2 id="the-promise-of-llms-in-healthcare">The Promise of LLMs in Healthcare</h2> <p>Large language models like GPT-4, Gemini, and open-source alternatives have demonstrated remarkable capabilities in understanding and generating human-like text. When applied to healthcare, these models can:</p> <ul> <li><strong>Simplify complex medical information</strong> for patients, improving health literacy</li> <li><strong>Automate clinical documentation</strong>, reducing physician burnout</li> <li><strong>Support policy development</strong> by analyzing vast amounts of healthcare data and literature</li> <li><strong>Enable personalized patient communication</strong> at scale</li> </ul> <h2 id="key-challenges">Key Challenges</h2> <p>Despite the promise, several challenges remain:</p> <ol> <li><strong>Data Privacy and HIPAA Compliance</strong>: Healthcare data is highly sensitive, and deploying LLMs requires careful consideration of privacy regulations.</li> <li><strong>Hallucination and Accuracy</strong>: LLMs can generate plausible but incorrect information, which is particularly dangerous in medical contexts.</li> <li><strong>Bias and Equity</strong>: Models trained on biased data can perpetuate health disparities.</li> <li><strong>Regulatory Frameworks</strong>: The FDA and other regulatory bodies are still developing guidelines for AI in healthcare.</li> </ol> <h2 id="moving-forward">Moving Forward</h2> <p>The key to successful LLM deployment in healthcare lies in responsible AI practices — ensuring transparency, accountability, and continuous evaluation. My research advocates for a framework that balances innovation with patient safety, leveraging the strengths of generative AI while mitigating its risks.</p> <p>As we continue to push the boundaries of what AI can do in healthcare, it is crucial that we maintain a patient-centered approach, ensuring that these powerful tools serve to enhance, not replace, the human elements of care.</p>]]></content><author><name></name></author><category term="research"/><category term="generative-ai"/><category term="healthcare"/><category term="LLMs"/><summary type="html"><![CDATA[Exploring how generative AI and large language models can transform patient engagement and policy development in the U.S. healthcare system.]]></summary></entry><entry><title type="html">Responsible AI for Healthcare Resource Allocation</title><link href="https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/responsible-ai-healthcare/" rel="alternate" type="text/html" title="Responsible AI for Healthcare Resource Allocation"/><published>2026-02-15T10:00:00+00:00</published><updated>2026-02-15T10:00:00+00:00</updated><id>https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/responsible-ai-healthcare</id><content type="html" xml:base="https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/responsible-ai-healthcare/"><![CDATA[<p>Healthcare systems worldwide face a persistent challenge: how to allocate limited resources — hospital beds, ventilators, staff, and medications — in a way that maximizes patient outcomes while maintaining fairness. AI offers powerful optimization tools, but deploying them responsibly is critical.</p> <h2 id="why-responsible-ai-matters-in-healthcare">Why Responsible AI Matters in Healthcare</h2> <p>Healthcare resource allocation decisions directly impact patient lives. An AI system that optimizes for efficiency without considering equity could systematically disadvantage vulnerable populations. Responsible AI in this context means:</p> <ul> <li><strong>Fairness</strong>: Ensuring equitable access to resources across demographic groups</li> <li><strong>Transparency</strong>: Making the decision-making process interpretable to clinicians and administrators</li> <li><strong>Accountability</strong>: Maintaining clear audit trails and human oversight</li> <li><strong>Robustness</strong>: Performing reliably under varying conditions and edge cases</li> </ul> <h2 id="a-java-based-framework">A Java-Based Framework</h2> <p>In my research, I developed a Java-based framework for healthcare resource allocation that incorporates responsible AI principles from the ground up. The framework includes:</p> <ol> <li><strong>Fairness Constraints</strong>: Built-in mechanisms to detect and mitigate bias across protected attributes</li> <li><strong>Explainable Outputs</strong>: Every allocation decision comes with a human-readable justification</li> <li><strong>Scenario Modeling</strong>: Support for what-if analysis to evaluate allocation strategies before deployment</li> <li><strong>Multi-Stakeholder Input</strong>: Integration of clinical, administrative, and patient perspectives</li> </ol> <h2 id="lessons-for-ai-practitioners">Lessons for AI Practitioners</h2> <p>Building responsible AI systems is not just an ethical imperative — it is a practical one. Systems that are fair, transparent, and accountable are more likely to be trusted and adopted by healthcare professionals, ultimately leading to better patient outcomes.</p> <p>The key takeaway: responsible AI is not a constraint on innovation — it is a catalyst for building systems that truly serve their intended purpose.</p>]]></content><author><name></name></author><category term="research"/><category term="responsible-ai"/><category term="healthcare"/><category term="machine-learning"/><summary type="html"><![CDATA[Designing AI systems that optimize healthcare resource allocation while maintaining fairness and transparency.]]></summary></entry><entry><title type="html">StepShield: Rethinking Intervention Strategies for AI Agents</title><link href="https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/stepshield-rogue-agents/" rel="alternate" type="text/html" title="StepShield: Rethinking Intervention Strategies for AI Agents"/><published>2026-01-20T10:00:00+00:00</published><updated>2026-01-20T10:00:00+00:00</updated><id>https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/stepshield-rogue-agents</id><content type="html" xml:base="https://michaelenny.github.io/michael-eniolade.github.io/blog/2026/stepshield-rogue-agents/"><![CDATA[<p>As AI agents become increasingly autonomous, the question of safety and oversight becomes paramount. In our paper <em>StepShield: When, Not Whether to Intervene on Rogue Agents</em> (arXiv:2601.22136), we propose a novel framework for AI agent intervention that shifts the focus from binary control to temporal optimization.</p> <h2 id="the-problem-with-binary-intervention">The Problem with Binary Intervention</h2> <p>Traditional approaches to AI safety often frame intervention as a binary decision: either we stop the agent or we let it run. This all-or-nothing approach has significant drawbacks:</p> <ul> <li><strong>Over-intervention</strong> wastes computational resources and prevents agents from completing beneficial tasks</li> <li><strong>Under-intervention</strong> allows potentially harmful actions to propagate</li> <li><strong>Static policies</strong> cannot adapt to the dynamic nature of agent behavior</li> </ul> <h2 id="the-stepshield-approach">The StepShield Approach</h2> <p>Our framework introduces a step-level monitoring system that continuously evaluates an agent’s trajectory. Rather than asking <em>whether</em> to intervene, we ask <em>when</em> — identifying the optimal intervention point that maximizes safety while minimizing unnecessary disruption.</p> <p>Key components of StepShield include:</p> <ol> <li><strong>Trajectory Analysis</strong>: Monitoring the agent’s actions at each step to detect deviation from expected behavior</li> <li><strong>Risk Scoring</strong>: Assigning dynamic risk scores based on the potential consequences of each action</li> <li><strong>Intervention Timing</strong>: Using these scores to determine the optimal moment for intervention</li> <li><strong>Graceful Recovery</strong>: Allowing the agent to resume from a safe state after intervention</li> </ol> <h2 id="implications-for-the-field">Implications for the Field</h2> <p>The StepShield framework has broad implications for the deployment of autonomous AI systems in high-stakes environments, from healthcare to cybersecurity. By providing a more nuanced approach to agent oversight, we can build systems that are both more capable and more trustworthy.</p> <p>This work represents a collaboration between researchers at Stanford University, University of the Cumberlands, and the Indian Institute of Science.</p>]]></content><author><name></name></author><category term="research"/><category term="ai-safety"/><category term="agents"/><category term="machine-learning"/><summary type="html"><![CDATA[A deep dive into our research on determining when — not whether — to intervene on rogue AI agents.]]></summary></entry><entry><title type="html">Building Production ML Pipelines with PySpark and Airflow</title><link href="https://michaelenny.github.io/michael-eniolade.github.io/blog/2025/building-ml-pipelines/" rel="alternate" type="text/html" title="Building Production ML Pipelines with PySpark and Airflow"/><published>2025-12-10T10:00:00+00:00</published><updated>2025-12-10T10:00:00+00:00</updated><id>https://michaelenny.github.io/michael-eniolade.github.io/blog/2025/building-ml-pipelines</id><content type="html" xml:base="https://michaelenny.github.io/michael-eniolade.github.io/blog/2025/building-ml-pipelines/"><![CDATA[<p>Building machine learning models is one thing — deploying them reliably at scale is another. Over the past several years, I have worked extensively with PySpark, Airflow, and cloud platforms to build production-grade ML pipelines. Here are the key lessons I have learned.</p> <h2 id="architecture-overview">Architecture Overview</h2> <p>A robust ML pipeline typically consists of several stages:</p> <ol> <li><strong>Data Ingestion</strong>: Pulling raw data from various sources (APIs, databases, file systems)</li> <li><strong>Data Transformation</strong>: Cleaning, feature engineering, and preparing training datasets</li> <li><strong>Model Training</strong>: Training and evaluating models with experiment tracking</li> <li><strong>Model Deployment</strong>: Serving models via APIs or batch inference</li> <li><strong>Monitoring</strong>: Tracking model performance and data drift in production</li> </ol> <h2 id="tool-selection">Tool Selection</h2> <p>After working with numerous tools, here is the stack I have found most effective:</p> <table> <thead> <tr> <th>Stage</th> <th>Tool</th> <th>Why</th> </tr> </thead> <tbody> <tr> <td>Orchestration</td> <td>Apache Airflow</td> <td>DAG-based scheduling, rich UI, extensive integrations</td> </tr> <tr> <td>Processing</td> <td>PySpark</td> <td>Distributed computing for large-scale data</td> </tr> <tr> <td>Storage</td> <td>Delta Lake</td> <td>ACID transactions, schema enforcement, time travel</td> </tr> <tr> <td>Transformation</td> <td>dbt</td> <td>SQL-based transformations with version control</td> </tr> <tr> <td>Experiment Tracking</td> <td>MLflow</td> <td>Model versioning, metrics logging, artifact storage</td> </tr> <tr> <td>Containerization</td> <td>Docker + K8s</td> <td>Reproducible environments, scalable deployment</td> </tr> </tbody> </table> <h2 id="key-lessons">Key Lessons</h2> <ul> <li><strong>Start simple</strong>: Begin with a basic pipeline and add complexity as needed</li> <li><strong>Version everything</strong>: Data, code, models, and configurations should all be versioned</li> <li><strong>Monitor early</strong>: Set up monitoring before issues arise in production</li> <li><strong>Automate testing</strong>: Include data validation tests in your pipeline</li> <li><strong>Design for failure</strong>: Build retry logic and alerting into every stage</li> </ul> <p>The goal is not to use the most sophisticated tools, but to build a pipeline that is reliable, maintainable, and scalable.</p>]]></content><author><name></name></author><category term="engineering"/><category term="data-engineering"/><category term="machine-learning"/><category term="cloud"/><summary type="html"><![CDATA[Lessons learned from building scalable machine learning pipelines using PySpark, Airflow, and cloud-native tools.]]></summary></entry></feed>