Skip to main content
Back to Newswire
AI Startups

PrismML Emerges From Stealth With $16.25 Million, Claims 1-Bit LLM Matches Full-Precision Performance

PrismML, a startup founded by researchers from Caltech, has come out of stealth with $16.25 million in SAFE and seed funding, claiming its 1-bit large language model technology achieves radical compression of AI models without degrading output quality. The company's core assertion is that its quantization approach can reduce model weights to single-bit representations while preserving the performance characteristics of much larger, full-precision models. If validated at scale, the technology could substantially lower the compute and memory requirements for running frontier AI models, with implications for on-device inference, edge deployment, and the economics of data center operations. The funding round was reported by the Wall Street Journal. PrismML did not disclose the names of investors in the round at the time of publication. 1-bit quantization has been an active area of AI research, with Microsoft publishing work on the technique in 2024 and other labs exploring the tradeoffs between compression and accuracy. Most prior approaches showed performance degradation at extreme compression ratios, particularly on complex reasoning tasks. PrismML claims its method avoids those tradeoffs, though independent benchmarks have not yet been published. The startup enters a crowded field of companies attempting to make AI inference more efficient. Competitors include Groq, which uses custom silicon, as well as software-level approaches from companies like Neural Magic and llama.cpp contributors. PrismML said it plans to use the funding to expand its team and accelerate development of its inference stack.
Sources
Published by Tech & Business, a media brand covering technology and business. This story was sourced from Wall Street Journal and reviewed by the T&B editorial agent team.