Tether’s AI research division, Tether Data, has released QVAC MedPsy, a new family of text-only medical language models optimized for edge deployment. The models come in 1.7 billion and 4 billion parameter versions and are designed to run on consumer hardware including smartphones, laptops, and wearable devices while maintaining strong medical reasoning performance.
According to Tether Data, the smaller QVAC MedPsy-1.7B model achieved an average score of 62.62 across seven closed-ended medical benchmarks. The company said this outperformed Google’s MedGemma-1.5-4B-it model by more than 11 points despite using less than half the parameters. It also approached the performance of larger reasoning-focused models such as Qwen3-4B-Thinking-2507.
The larger QVAC MedPsy-4B model reportedly surpassed MedGemma-27B-text-it on several benchmarks tied to practical healthcare reasoning. On HealthBench Hard, which measures performance in more realistic clinical scenarios, Tether Data reported scores of 58.00 for MedPsy-4B compared with 42.00 for Google’s 27-billion-parameter system.
Tether Data also emphasized inference efficiency as a major differentiator. The company said MedPsy-4B generated benchmark answers using an average of roughly 909 tokens, compared with approximately 2,953 tokens for Qwen3-4B-Thinking-2507. Lower token usage reduces latency and compute costs, which is particularly important for real-time deployment on lower-power devices.
The models are being released under the Apache 2.0 license for research and educational use. Tether Data is also publishing GGUF versions compatible with llama.cpp and its own QVAC SDK, including quantized variants designed to reduce storage requirements while maintaining most benchmark performance. The company said some compressed versions cut file size by nearly 70% with minimal performance degradation.
QVAC MedPsy was evaluated across eight benchmark suites covering clinical reasoning, biomedical research, health literacy, and underserved healthcare contexts. These included MedQA-USMLE, MedMCQA, PubMedQA, AfriMedQA, and HealthBench.
Smaller Medical Models Target Real-World Deployment
The release reflects growing demand for medical AI systems that can run locally instead of relying entirely on cloud infrastructure. Most high-performing healthcare language models are too large to deploy directly on edge devices, limiting their use in low-connectivity or privacy-sensitive environments.
By reducing parameter count and token usage while maintaining benchmark performance, Tether Data is targeting practical deployment scenarios such as offline clinical assistance, medical education tools, and decision-support systems operating directly on consumer hardware.
The focus on local inference is also significant for healthcare providers dealing with strict data privacy requirements. Running models directly on devices can reduce the need to transmit sensitive patient information to remote servers, which may simplify compliance and improve response times.
Tether Expands Into AI-Driven Health Technologies
The MedPsy launch is part of a broader expansion of Tether’s AI and health technology efforts. Earlier, the company introduced BrainWhisperer, a brain-computer interface system designed to convert neural activity into text using on-device AI processing. Tether claimed the system achieved up to 98.3% accuracy while keeping neural data local to the device.
Tether has also been expanding into consumer wellness technologies through investment activity. Eight Sleep recently received a strategic investment from Tether Investments at a reported $1.5 billion valuation, with a focus on AI-driven sleep monitoring and personalized health intelligence.