Foxconn Unveils FoxBrain—Chinese AI Model Poised for Open Source Release

1 month ago 15
  • Published on March 10, 2025
  • In AI News

FoxBrain’s training process with NVIDIA GPUs finished in just four weeks.

Hon Hai Research Institute, a research centre by Hon Hai Technology Group (Foxconn), launched a traditional Chinese large language model (LLM) code-named FoxBrain on Monday. 

FoxBrain was initially designed for internal group applications, but it is set to become an important engine for Foxconn’s three major platforms: Smart Manufacturing, Smart EV, and Smart City.

It was initially developed for the Group’s internal systems, supporting applications in data analysis, decision support, document collaboration, mathematics, reasoning, problem-solving, and code generation. The institute also mentioned that the model will be open-sourced and shared publicly in the future.

The institute noted that the development was a milestone in Taiwan’s AI technology, where an efficient and lower-cost model training method was completed in just four weeks.

FoxBrain’s quick training process was made possible using 120 NVIDIA H100 GPUs, scaled with NVIDIA Quantum-2 InfiniBand network. NVIDIA  helped during the training process through the Taipei-1 Supercomputer and technical consultation.

The model is based on Meta’s Llama 3.1 architecture with 70B parameters and features 128k tokens for its context window length. It is said that a unique adaptive reasoning reflection technique was used to train the model in autonomous reasoning.

“In recent months, the deepening of reasoning capabilities and the efficient use of GPUs have gradually become the mainstream development in the field of AI. Our FoxBrain model adopted a very efficient training strategy, focusing on optimising the training process rather than blindly accumulating computing power,” said Yung-Hui Li, director of the Artificial Intelligence Research Centre at the Hon Hai Research Institute. “Through carefully designed training methods and resource optimisation, we have successfully built a local AI model with powerful reasoning capabilities.”

The institute mentioned that its performance is close to world-leading standards, but there is a slight gap with DeepSeek’s distillation model.

The institute stated that it outperforms Llama-3-Taiwan-70B on the same scale, with mathematics and logical reasoning as the focus. However, the result of FoxBrain is scheduled to be shared at NVIDIA GTC 2025 in a session talk titled ‘From Open Source to Frontier AI: Build, Customise, and Extend Foundational Models’ on March 20.

Picture of Ankush Das

Ankush Das

I am a tech aficionado and a computer science graduate with a keen interest in AI, Open Source, and Cybersecurity.

Association of Data Scientists

GenAI Corporate Training Programs

India's Biggest Women in Tech Summit

March 20 and 21, 2025 | 📍 NIMHANS Convention Center, Bengaluru

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

Rising 2025 Women in Tech & AI

March 20 - 21, 2025 | 📍 NIMHANS Convention Center, Bengaluru

AI Startups Conference.April 25, 2025 | 📍 Hotel Radisson Blu, Bengaluru, India

Data Engineering Summit 2025

May 15 - 16, 2025 | 📍 Hotel Radisson Blu, Bengaluru

MachineCon GCC Summit 2025

June 20 to 22, 2025 | 📍 ITC Grand, Goa

Sep 17 to 19, 2025 | 📍KTPO, Whitefield, Bengaluru, India

India's Biggest Developers Summit Feb, 2025 | 📍Nimhans Convention Center, Bengaluru

Read Entire Article