Tele-FLM
T
Tele FLM
Overview :
Tele-FLM (also known as FLM-2) is a 52-billion parameter open-source multilingual large language model with a stable and efficient pre-training paradigm and enhanced fact-checking capabilities. Based on a decoder-only transformer architecture, it has been trained on approximately 2 trillion tokens. Tele-FLM exhibits superior performance compared to models of similar size, sometimes even surpassing larger ones. Besides sharing the model weights, we also provide core design, engineering practices, and training details, hoping they will benefit both the academic and industrial communities.
Target Users :
Tele-FLM is primarily aimed at developers and researchers who need to process and generate multilingual text, especially professionals in the field of natural language processing seeking efficient and high-performing models.
Total Visits: 29.7M
Top Region: US(17.94%)
Website Views : 46.1K
Use Cases
Used for generating concise summaries of text in specific domains.
Provides accurate information retrieval and answering capabilities in question-answering systems.
Serves as a backend for chatbots, delivering a smooth conversational experience.
Features
Decoder-only transformer architecture-based model, optimized for fact-checking capabilities.
Supports multiple languages, including English and Chinese.
Provides core design and engineering practices for easy community use and learning.
Training data covers multiple domains, encompassing a wide range of knowledge.
Utilizes 3D parallel training techniques to enhance training efficiency.
Demonstrates good performance on multiple benchmark datasets.
How to Use
1. Import the torch and transformers libraries.
2. Load the tokenizer and model from the pre-trained model using AutoTokenizer and AutoModelForCausalLM.
3. Convert the input text into a format understandable by the model using the tokenizer.
4. Send the converted input data to the model's device.
5. Generate text using the model.generate method.
6. Decode the generated text back into readable format using the tokenizer.decode method.
7. Print the final generated text.
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase