TAIDE CAN HELP

Automatic Summarization

arrow

Write a Letter

arrow

Write an Article

arrow

Chinese-to-English Translation

arrow

English-to-Chinese Translation

arrow

Automatic Summarization

arrow

Write a Letter

arrow

Write an Article

arrow

Chinese-to-English Translation

arrow

English-to-Chinese Translation

arrow

TAIDE CAN HELP

Automatic Summarization

arrow

Write a Letter

arrow

Write an Article

arrow

Chinese-to-English Translation

arrow

English-to-Chinese Translation

arrow

PHASED GOALS

Model Size 7B

Computational Resources

Dedicated Computing Infrastructure for TAIDE

  • 32 nodes allocated specifically for TAIDE project model training.
  • Each node boasts 8 V100 32GB GPUs, ensuring substantial processing power.

Optimization with DeepSpeed 03

  • Implementing the DeepSpeed 03 framework for efficient and optimized training of large Al models within the TAIDE project.

Scaled Training Capabilities for TAIDE Models

  • For 7B Model Training: Deploying 8 nodes, totaling 64 V100 GPUs.
  • For 13B Model Training: Expanding to 12-16 nodes, utilizing 96-128 V100 GPUs.

Trustworthy Pre-Training Data

  • Importance of starting with reliable data for building a Base Model.
  • Project initiated by the Taiwanese government, led by our team.
  • Access to pre-training data facilitated by government resources.
  • Collaboration with public media and government agencies yielded about 1.7 billion tokens.
  • Achieved a substantial collection of 420,000 pieces of data.
Model Size 13B

Computational Resources

Dedicated Computing Infrastructure for TAIDE

  • 32 nodes allocated specifically for TAIDE project model training.
  • Each node boasts 8 V100 32GB GPUs, ensuring substantial processing power.

Optimization with DeepSpeed 03

  • Implementing the DeepSpeed 03 framework for efficient and optimized training of large Al models within the TAIDE project.

Scaled Training Capabilities for TAIDE Models

  • For 7B Model Training: Deploying 8 nodes, totaling 64 V100 GPUs.
  • For 13B Model Training: Expanding to 12-16 nodes, utilizing 96-128 V100 GPUs.

Trustworthy Pre-Training Data

  • Importance of starting with reliable data for building a Base Model.
  • Project initiated by the Taiwanese government, led by our team.
  • Access to pre-training data facilitated by government resources.
  • Collaboration with public media and government agencies yielded about 1.7 billion tokens.
  • Achieved a substantial collection of 420,000 pieces of data.
stage

TAIDE Team

National Science and Technology Council

Comprising the National Applied Research Laboratories, Ministry of Digital Affairs, Academia Sinica, professors from major universities, and experts from various fields.

Elite Ensemble

Core model training with RLHF group

Data acquisition with annotation team

High-speed computing resources and technical support team

Application service platform construction and management team

Superior Team

International observation with specification group

Evaluation system and Testing Environment Team

Project Supervision Management

Project Execution Management

Project Info

This project is grounded in Taiwanese culture, incorporating unique elements such as Taiwanese language, values, and customs, enabling generative AI to understand and respond to the needs of local users. It aims to create a trustworthy foundational model for generative AI engines.

As AI applications become increasingly widespread, both businesses and the public have higher expectations. Consequently, there is an urgent demand for AI's performance, safety and robustness, as well as fairness and transparency.

Custom-Built for Taiwan

Creating a trustworthy generative AI dialogue engine foundational model specifically for Taiwan, the government and industry can select the model size by computational power according to their services, to train the model themselves and establish internal applications.

Diverse Traditional Chinese Training Materials

Incorporating various thematic textual resources and training materials to create specific domain application examples, enhancing the model's performance in different thematic areas.

Laying the Foundation for Widespread Application

Constructing a computational environment and application service platform, providing the necessary computational power for model development, and laying the groundwork for subsequent promotion and application.

Strengthening the AI Development Environment

Through legal analysis, establishment of testing norms, and development of evaluation tools, the AI development environment is strengthened and public trust is enhanced.

Public-Private Collaboration for Mutual Benefit

Utilizing public-private collaboration to assist industries in integrating foundational models, these can be used for specific industrial applications. Custom adjustments are made with data provided by the businesses.

close

Core model training with RLHF group

蔡宗翰 Tsai, Tzong-Han
arrow

Tsai, Tzong-Han

Professor
Department of Computer Science & Information Engineering, National Central University.
Coordinator
許永真 Hsu, Yung-Jen
arrow

Hsu, Yung-Jen

Professor
Department of Computer Science and Information Engineering, National Taiwan University
Consultant
吳毅成 Wu, I-Chen
arrow

Wu, I-Chen

Distinguished Professor
College of Computer Science, National Yang Ming Chiao Tung University
Consultant
李宏毅 Lee, Hung-Yi
arrow

Lee, Hung-Yi

Associate Professor
Department of Electrical Engineering, National Taiwan University
Expert
廖元甫 Liao, Yuan-Fu
arrow

Liao, Yuan-Fu

Professor
Institute of Artificial Intelligence Innovation, Industry Academia Innovation School, National Yang Ming Chiao Tung University
Consultant
黃瀚萱 Huang, Hen-Hsen
arrow

Huang, Hen-Hsen

Assistant Research Fellow
Institute of Information Science, Academia Sinica,
Consultant
close

Data acquisition with annotation team

莊庭瑞 Chuang, Tyng-Ruey
arrow

Chuang, Tyng-Ruey

Associate Research Fellow
Institute of Information Science, Academia Sinica
Consultant
close

High-speed computing resources and technical support team

張朝亮 Chang, Chau-Lyan

Chang, Chau-Lyan

Director General
National Center for High-performance Computing, NARLabs
Coordinator
close

Application service platform construction and management team

吳俊興 Wu, Chun-Hsin
arrow

Wu, Chun-Hsin

Associate Professor
Department of Computer Science & Information Engineering, Kaohsiung University
Consultant
close

Project Supervision Management

李育杰 Lee, Yuh-Jye

Lee, Yuh-Jye

Research Fellow
Research Center for Information Technology Innovation, Academia Sinica
Consultant
林博文 Lin, Bou-Wen

Lin, Bou-Wen

Vice President
National Applied Research Laboratories
Coordinator
林青嶔 Lin, Ching-Chin

Lin, Ching-Chin

Supervisor
Administration for Digital Industries, moda
Co-Principal Investigator
蕭景燈 Hsiao, Ching-Teng

Hsiao, Ching-Teng

Director General
Office of Science and Technology Policy, NSTC
OSTP Coordinator
close

Project Execution Management

許武龍 Hsu, Wuu-Long

Hsu, Wuu-Long

Founder
Location Aware Sensing System
Project Manager
蕭奕弘 Hsiao, Yi-Hon

Hsiao, Yi-Hon

Attorney-at-Law
Hsiao, Yi-Hon Law Firm
Consultant