SoftBank, Aizip Launch On-Device AI for Enterprise Privacy

CUPERTINO, CA / Tokyo, Japan - SoftBank Corp. and Aizip, Inc. have jointly developed a new Small Language Model (SLM) and Retrieval Augmented Generation (RAG) solution that operates locally on mobile devices and on-premises servers, the companies announced today.

The system, designed to address enterprise data privacy concerns, can be customized with domain-specific data and achieves accuracy comparable to 100 times larger cloud-based language models.

Testing on SoftBank's internal documents demonstrated that the system successfully answered 97% of employee questions when deployed as a mobile application running entirely on an iPhone 14. 

The solution's performance, evaluated across 22,000 questions, matched responses generated by a GPT-4-based RAG system.

Aizip's SLM customization pipeline incorporates data generation, fine-tuning, and evaluation processes that are executed iteratively until the target accuracy levels are reached. 

The development addresses reliability issues that have limited enterprise adoption of open-source Small Language Models.

According to the companies' June 2024 announcement, the technology's deployment capabilities extend beyond mobile devices to include on-premises servers, PCs, and IoT devices with microprocessors.

SoftBank vice president Katsuya Kitazawa, who heads the Information Technology & Architect Division, said the technology will benefit privacy-critical and offline-required use cases, including applications for flight attendants and field workers in remote locations.

The announcement comes as enterprise concerns about data privacy have led to restrictions on cloud-based language models. 

A Menlo Ventures survey indicated that 21% of failed AI pilots were attributed to data privacy challenges.

Aizip maintains its focus on developing production-grade AI solutions across its product lines, including Aizip Intelligent Audio, Vision, and Time-Series, emphasizing improvements in accuracy, reliability, speed, and development efficiency.