The world’s largest contract electronics manufacturer Foxconn has unveiled its first large-scale language model with reasoning capabilities, FoxBrain, which was trained in just four weeks. The new AI model is said to be based on Meta✴’s Llama 3.1 architecture and is optimized for the traditional Mandarin dialect used in Taiwan.

Image source: Foxconn

Foxconn said Nvidia supported the project with Taiwan’s largest supercomputer, Taipei-1, and technical consulting services for training the model. Foxconn’s FoxBrain model was trained using 120 Nvidia H100 accelerators.

Designed for use in the enterprise, FoxBrain’s capabilities include data analysis, mathematical calculations, problem solving, document collaboration, reasoning, and code generation.

The model “prioritizes optimized learning strategies rather than simply using computing power” to solve a problem, said Yung-Hui Li, director of Foxconn’s AI research center.

Foxconn noted that while FoxBrain is slightly inferior in performance to China’s DeepSeek distillation model, its overall performance is close to global standards, allowing the AI ​​model to be used to improve the company’s manufacturing and supply chain management.

In the future, Foxconn plans to open-source the model, which will allow collaboration with technology partners to expand the model’s application areas, share information, and advance AI in manufacturing, supply chain management, and intelligent decision making.

The company said it would reveal more details about FoxBrain at Nvidia’s GTC AI developer conference in mid-March.

Leave a Reply

Your email address will not be published. Required fields are marked *