<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Paper-Conference | Zachary Zhao</title><link>https://bobbed1999.github.io/publication_types/paper-conference/</link><atom:link href="https://bobbed1999.github.io/publication_types/paper-conference/index.xml" rel="self" type="application/rss+xml"/><description>Paper-Conference</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Wed, 01 Oct 2025 00:00:00 +0000</lastBuildDate><image><url>https://bobbed1999.github.io/media/icon_hu_37226ddb91a8677e.png</url><title>Paper-Conference</title><link>https://bobbed1999.github.io/publication_types/paper-conference/</link></image><item><title>Causal Reinforcement Learning based Agent-Patient Interaction with Clinical Domain Knowledge</title><link>https://bobbed1999.github.io/publications/conference-paper/</link><pubDate>Wed, 01 Oct 2025 00:00:00 +0000</pubDate><guid>https://bobbed1999.github.io/publications/conference-paper/</guid><description>&lt;!-- &gt; [!NOTE]
&gt; Click the _Cite_ button above to demo the feature to enable visitors to import publication metadata into their reference management software.
&gt; [!NOTE]
&gt; Create your slides in Markdown - click the _Slides_ button to check out the example.
Add the publication's **full text** or **supplementary notes** here. You can use rich formatting such as including [code, math, and images](https://docs.hugoblox.com/content/writing-markdown-latex/). --&gt;</description></item><item><title>A Distributed Multimodal Robotic Framework for Emotion-Aware Reminiscence Dialogue in Dementia Care</title><link>https://bobbed1999.github.io/publications/conference-paper1/</link><pubDate>Mon, 01 Sep 2025 00:00:00 +0000</pubDate><guid>https://bobbed1999.github.io/publications/conference-paper1/</guid><description>&lt;!-- &gt; [!NOTE]
&gt; Click the _Cite_ button above to demo the feature to enable visitors to import publication metadata into their reference management software.
&gt; [!NOTE]
&gt; Create your slides in Markdown - click the _Slides_ button to check out the example.
&gt; [!NOTE]
&gt; Create your slides in Markdown - click the _Slides_ button to check out the example.
Add the publication's **full text** or **supplementary notes** here. You can use rich formatting such as including [code, math, and images](https://docs.hugoblox.com/content/writing-markdown-latex/). --&gt;
&lt;p&gt;We introduce an embodied robotic implementation of the \textbf{PARTNER} framework (Personalized AI and Robotics to Nurture Engaging Reminiscence), a distributed multimodal architecture for emotion-aware, personalized dialogue in socially assistive contexts. The framework has three components: a secure cloud portal for managing media, a local server for processing multimodal inputs, and an embodied robot client. PARTNER combines auditory, visual, and textual inputs using Whisper for speech transcription and a vision–language model (GPT-4o) that infers implicit affect from facial snapshots and dialogue history, rather than relying on rigid emotion classifiers. To enhance reproducibility and support future model training, PARTNER incorporates a real-time logging pipeline that synchronizes user inputs, sensor streams, and model outputs into a structured dataset.
We provide a system-level evaluation on our robot, measuring end-to-end command–response latency, transcription accuracy, and dialogue coherence under varied sensing and environmental conditions. Our experiments show sub-3,s loop latency on our testbed, robust transcription across various noise environments, and consistent responses during multi-turn dialogues, These findings validate PARTNER as a deployable platform for adaptive human–robot interaction. To our knowledge, PARTNER is the first Socially Assistive Robotics (SAR)-oriented system that (i) unifies a cloud portal for reminiscence media with a locally executed interaction server and an embodied agent, (ii) leverages VLM-based implicit affect cues for dialogue policy, and (iii) offers a real-time multimodal logging substrate to facilitate future domain-specific VLM/LLM fine-tuning.&lt;/p&gt;</description></item></channel></rss>