Model Overview
Advanced multimodal analysis system utilizing acoustic features, facial expressions, and physiological markers to detect and measure stress levels, emotions, in real-time with clinical-grade accuracy.
Key Features
- Real-time analysis (300-400ms)
- Multi-feature processing
- Gender-adaptive normalization
- Visual analysis output
- Automatic Speech Recognition
- Emotion Detection with Confidence Scores
- FACS-based Stress Analysis
- Anxiety Detection through Action Units
- Multi-modal Analysis Support
curl -X POST "https://api.dyagnosys.com/analyze/" \ -F "[email protected]" \ -F "[email protected]"
Performance Metrics
Response Time
300ms
Average Accuracy
90%
Clinical validation Max File Size
1MB
WAV format Input Operational
Audio Processing Limitations
- Requires clear audio input
Minimum 44.1kHz sample rate, -23 LUFS loudness, <-60dB noise floor
- Dependent on audio quality for accuracy
Accuracy drops by 15% for each 10dB increase in background noise
System Operation Requirements
- Best for continuous monitoring
Optimal analysis period: 30+ minutes for baseline establishment
Visual Analysis Limitations
- Optimal lighting required for facial analysis
Minimum 300 lux, uniform lighting, avoid strong shadows
Clinical Usage Limitations
- For medical diagnosis (Clinical Trials in progress)
Clinical trials completion expected Q1 2025
Live Demo
Upload a video to analyze facial action units for stress and anxiety indicators
Video Analysis
Upload a video to see processed frames
Implementation Guide
Implementation details and documentation will be displayed here.
Research Papers
Research papers and publications will be listed here.