Local-first AI medical scribe running entirely in the browser using Chrome built-in AI (no backend)
-
Updated
Mar 19, 2026 - JavaScript
Local-first AI medical scribe running entirely in the browser using Chrome built-in AI (no backend)
This project investigates multimodal emotion recognition using a Transformer-based architecture designed to fuse audio and visual data. The primary goal is to capture fine-grained human emotions by modeling both intra-modal temporal dependencies and inter-modal interactions.
Add a description, image, and links to the avt topic page so that developers can more easily learn about it.
To associate your repository with the avt topic, visit your repo's landing page and select "manage topics."