Convert Figma logo to code with AI

FlagAI-Open logoFlagAI

FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.

3,873
419
3,873
22

Top Related Projects

🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.

39,112

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Ongoing research training transformer models at scale

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

11,862

An open-source NLP research library, built on PyTorch.

31,682

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Quick Overview

FlagAI is an open-source AI toolkit developed by BAAI (Beijing Academy of Artificial Intelligence). It provides a comprehensive set of tools and models for natural language processing, computer vision, and multimodal tasks. FlagAI aims to facilitate AI research and application development by offering pre-trained models, efficient training frameworks, and easy-to-use APIs.

Pros

  • Comprehensive toolkit covering multiple AI domains (NLP, CV, multimodal)
  • Offers pre-trained models and efficient training frameworks
  • Supports both PyTorch and TensorFlow backends
  • Provides easy-to-use APIs for quick integration and deployment

Cons

  • Documentation may be limited or not as extensive as some other popular AI libraries
  • Community support might be smaller compared to more established frameworks
  • May have a steeper learning curve for beginners due to its comprehensive nature
  • Some features or models might be more focused on Chinese language processing

Code Examples

  1. Loading a pre-trained BERT model:
from flagai.auto_model.auto_loader import AutoLoader

loader = AutoLoader(task_name="text_classification", model_name="BERT-base-en")
model = loader.get_model()
tokenizer = loader.get_tokenizer()
  1. Performing text classification:
text = "FlagAI is an excellent AI toolkit."
tokens = tokenizer.tokenize(text)
input_ids = tokenizer.convert_tokens_to_ids(tokens)
logits = model(input_ids)
predicted_class = logits.argmax(-1).item()
  1. Fine-tuning a model on a custom dataset:
from flagai.trainer import Trainer

trainer = Trainer(
    env_type="pytorch",
    experiment_name="bert_classification",
    batch_size=16,
    gradient_accumulation_steps=1,
    max_epochs=3,
    num_gpus=1,
    save_interval=1000,
    eval_interval=100,
)

trainer.train(model, train_dataset, valid_dataset)

Getting Started

To get started with FlagAI, follow these steps:

  1. Install FlagAI using pip:
pip install flagai
  1. Import the necessary modules:
from flagai.auto_model.auto_loader import AutoLoader
from flagai.trainer import Trainer
  1. Load a pre-trained model and tokenizer:
loader = AutoLoader(task_name="text_classification", model_name="BERT-base-en")
model = loader.get_model()
tokenizer = loader.get_tokenizer()
  1. Use the model for inference or fine-tuning as shown in the code examples above.

Competitor Comparisons

🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.

Pros of transformers

  • Extensive model support: Covers a wide range of NLP tasks and architectures
  • Large community and ecosystem: Frequent updates, extensive documentation, and third-party integrations
  • Seamless integration with PyTorch and TensorFlow

Cons of transformers

  • Can be complex for beginners due to its extensive features and options
  • Larger library size and potential overhead for simpler projects
  • May require more computational resources for some models

Code Comparison

transformers:

from transformers import AutoTokenizer, AutoModel

tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased")
model = AutoModel.from_pretrained("bert-base-uncased")

FlagAI:

from flagai.auto_model.auto_loader import AutoLoader

auto_loader = AutoLoader("seq2seq", "GLM-large-ch")
model = auto_loader.get_model()
tokenizer = auto_loader.get_tokenizer()

Both libraries offer similar functionality for loading pre-trained models and tokenizers, but transformers provides a more standardized approach across different model architectures. FlagAI focuses on specific models and tasks, potentially offering a more streamlined experience for supported use cases.

39,112

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Pros of DeepSpeed

  • More mature and widely adopted in the industry
  • Extensive documentation and community support
  • Broader range of optimization techniques and features

Cons of DeepSpeed

  • Steeper learning curve for beginners
  • Primarily focused on PyTorch, limiting flexibility for other frameworks
  • Can be complex to configure for specific use cases

Code Comparison

DeepSpeed:

import deepspeed
model_engine, optimizer, _, _ = deepspeed.initialize(
    args=args, model=model, model_parameters=params
)

FlagAI:

from flagai.auto_model.auto_loader import AutoLoader
loader = AutoLoader(task_name="text_classification",
                    model_name="BERT-base-en")
model = loader.get_model()

DeepSpeed offers more fine-grained control over optimization and distributed training, while FlagAI provides a simpler, more user-friendly interface for common NLP tasks. DeepSpeed is better suited for large-scale, performance-critical applications, whereas FlagAI is more accessible for quick prototyping and smaller projects.

Ongoing research training transformer models at scale

Pros of Megatron-LM

  • Highly optimized for NVIDIA GPUs, offering excellent performance for large-scale language models
  • Supports advanced parallelism techniques like tensor, pipeline, and sequence parallelism
  • Extensive documentation and examples for training and fine-tuning various model architectures

Cons of Megatron-LM

  • Limited flexibility for non-NVIDIA hardware or cloud environments
  • Steeper learning curve due to its focus on high-performance, distributed training
  • Less emphasis on easy-to-use APIs for downstream tasks and applications

Code Comparison

Megatron-LM (model initialization):

model = get_language_model(
    attention_mask_func, num_tokentypes=num_tokentypes, add_pooler=add_pooler,
    init_method=init_method, scaled_init_method=scaled_init_method)

FlagAI (model initialization):

model = BaseModel.from_pretrained(model_name)
model.to(device)

Megatron-LM focuses on distributed training and optimization, while FlagAI emphasizes ease of use and quick deployment. Megatron-LM's code is more complex, reflecting its advanced features, while FlagAI's API is more straightforward for common tasks. Both projects aim to facilitate large-scale language model development but cater to different user needs and hardware setups.

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

Pros of gpt-neox

  • Specialized for training large language models, particularly GPT-style models
  • Extensive documentation and community support
  • Highly optimized for distributed training on multiple GPUs

Cons of gpt-neox

  • Limited flexibility for other AI tasks beyond language modeling
  • Steeper learning curve for users new to large-scale language model training
  • Requires significant computational resources for optimal performance

Code Comparison

gpt-neox:

from megatron.neox_arguments import NeoXArgs
from megatron.global_vars import set_global_variables, get_tokenizer
from megatron.training import pretrain

args = NeoXArgs.from_ymls("configs/your_config.yml")
set_global_variables(args)

FlagAI:

from flagai.auto_model.auto_loader import AutoLoader
from flagai.trainer import Trainer

auto_loader = AutoLoader("lm", model_name="GLM-large-ch")
model = auto_loader.get_model()
trainer = Trainer(env_type="pytorch", pytorch_device="cuda")

The code snippets demonstrate the different approaches to model initialization and training setup. gpt-neox focuses on large-scale distributed training, while FlagAI offers a more user-friendly interface for various AI tasks.

11,862

An open-source NLP research library, built on PyTorch.

Pros of AllenNLP

  • More established and widely used in the NLP research community
  • Extensive documentation and tutorials available
  • Strong integration with PyTorch and support for various NLP tasks

Cons of AllenNLP

  • Steeper learning curve for beginners
  • Less focus on large-scale language models and multi-modal tasks
  • May require more setup and configuration for certain tasks

Code Comparison

AllenNLP:

from allennlp.predictors import Predictor

predictor = Predictor.from_path("https://storage.googleapis.com/allennlp-public-models/bert-base-srl-2020.03.24.tar.gz")
result = predictor.predict(sentence="Did Uriah honestly think he could beat the game in under three hours?")

FlagAI:

from flagai.auto_model.auto_loader import AutoLoader

loader = AutoLoader("seq2seq", "THUDM/chatglm-6b", use_cache=True)
model = loader.get_model()
tokenizer = loader.get_tokenizer()

Both libraries offer easy-to-use interfaces for loading and using pre-trained models, but FlagAI seems to have a more streamlined approach for large language models.

31,682

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Pros of fairseq

  • More established and widely used in the research community
  • Extensive documentation and examples for various NLP tasks
  • Supports a broader range of architectures and models

Cons of fairseq

  • Steeper learning curve for beginners
  • Less focus on Chinese language models and tasks
  • Requires more setup and configuration for specific use cases

Code Comparison

FlagAI:

from flagai.auto_model.auto_loader import AutoLoader

loader = AutoLoader("seq2seq", "GLM-large-ch")
model = loader.get_model()
tokenizer = loader.get_tokenizer()

fairseq:

from fairseq.models.transformer import TransformerModel

model = TransformerModel.from_pretrained('/path/to/model')
tokenizer = model.encode('Hello world')

FlagAI focuses on simplifying the process of loading and using pre-trained models, especially for Chinese language tasks. It provides a more streamlined API for common use cases.

fairseq offers more flexibility and control over model architecture and training process, but requires more code and configuration to set up and use models.

Both libraries support various NLP tasks, but FlagAI has a stronger emphasis on Chinese language models and applications, while fairseq covers a broader range of languages and architectures.

Convert Figma logo designs to code with AI

Visual Copilot

Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.

Try Visual Copilot

README

FlagAI CII Best Practices Python application GitHub release (release name instead of tag name)

简体中文


FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model. Our goal is to support training, fine-tuning, and deployment of large-scale models on various downstream tasks with multi-modality.

Why should I use FlagAI?

  1. Quickly Download Models via API

    FlagAI provides an API that allows you to quickly download pre-trained models and fine-tune them on a wide range of datasets collected from SuperGLUE and CLUE benchmarks for both Chinese and English text.

    FlagAI now supports over 30 mainstream models, including Language Model Aquila, multilingual text and image representation model AltCLIP, text-to-image generation model AltDiffusion Huggingface space, WuDao GLM (with a maximum of 10 billion parameters), EVA-CLIP, OPT, BERT, RoBERTa, GPT2, T5, ALM, and models from Huggingface Transformers, etc.

  2. Parallel train with fewer than 10 lines of code

    Backed by the four most popular data/model parallel libraries -- PyTorch, Deepspeed, Megatron-LM, BMTrain -- FlagAI allows for seamless integration between them, enabling users to parallel their training/testing process with fewer than ten lines of code.

  3. Conveniently use the few-shot learning toolkits

    FlagAI also provides prompt-learning toolkit for few-shot tasks.

  4. Particularly good at Chinese tasks

    These models can be applied to (Chinese/English) Text, for tasks like text classification, information extraction, question answering, summarization, and text generation, with a particular focus on Chinese tasks.

Toolkits and Pre-trained Models

The code is partially based on GLM, Transformers,timm and DeepSpeedExamples.

Toolkits

NameDescriptionExamples
GLM_custom_pvpCustomizing PET templatesREADME.md
GLM_ptuningp-tuning tool——
BMInf-generateAccelerating generationREADME.md

Pre-trained Models

ModelTaskTrainFinetuneInference/GenerateExamples
AquilaNatural Language Processing✅✅✅README.md
ALMArabic Text Generation✅❌✅README.md
AltCLIPImage-Text Matching✅✅✅README.md
AltCLIP-m18Image-Text Matching✅✅✅README.md
AltDiffusionText-to-Image Generation❌❌✅README.md
AltDiffusion-m18Text-to-Image Generation,supporting 18 languages❌❌✅README.md
BERT-title-generation-englishEnglish Title Generation✅❌✅README.md
CLIPImage-Text Matching✅❌✅——
CPM3-finetuneText Continuation❌✅❌——
CPM3-generateText Continuation❌❌✅——
CPM3_pretrainText Continuation✅❌❌——
CPM_1Text Continuation❌❌✅README.md
EVA-CLIPImage-Text Matching✅✅✅README.md
GalacticaText Continuation❌❌✅——
GLM-large-ch-blank-fillingBlank Filling❌❌✅TUTORIAL
GLM-large-ch-poetry-generationPoetry Generation✅❌✅TUTORIAL
GLM-large-ch-title-generationTitle Generation✅❌✅TUTORIAL
GLM-pretrainPre-Train✅❌❌——
GLM-seq2seqGeneration✅❌✅——
GLM-superglueClassification✅❌❌——
GPT-2-text-writtingText Continuation❌❌✅TUTORIAL
GPT2-text-writtingText Continuation❌❌✅——
GPT2-title-generationTitle Generation❌❌✅——
OPTText Continuation❌❌✅README.md
RoBERTa-base-ch-nerNamed Entity Recognition✅❌✅TUTORIAL
RoBERTa-base-ch-semantic-matchingSemantic Similarity Matching✅❌✅TUTORIAL
RoBERTa-base-ch-title-generationTitle Generation✅❌✅TUTORIAL
RoBERTa-faqQuestion-Answer❌❌✅README.md
Swinv1Image Classification✅❌✅——
Swinv2Image Classification✅❌✅——
T5-huggingface-11bTrain✅❌❌TUTORIAL
T5-title-generationTitle Generation❌❌✅TUTORIAL
T5-flagai-11bPre-Train✅❌❌——
ViT-cifar100Pre-Train✅❌❌——

Contributing

Thanks for your interest in contributing! There are many ways to get involved; start with our contributor guidelines and then check these open issues for specific tasks.

Contact us

Welcome to raise your questions or feature requests on GitHub Issues , and share your experience on the Discussions board.

Quick Start

We provide many models which are trained to perform different tasks. You can load these models by AutoLoader to make prediction. See more in FlagAI/quickstart.

Requirements and Installation

  • Python version >= 3.8
  • PyTorch version >= 1.8.0
  • [Optional] For training/testing models on GPUs, you'll also need to install CUDA and NCCL
  • To install FlagAI with pip:
pip install -U flagai
  • [Optional] To install FlagAI and develop locally:
git clone https://github.com/FlagAI-Open/FlagAI.git
python setup.py install
  • [Optional] For faster training, install NVIDIA's apex
git clone https://github.com/NVIDIA/apex
cd apex
pip install -v --disable-pip-version-check --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" ./
  • [Optional] For ZeRO optimizers, install DEEPSPEED (>= 0.7.7)
git clone https://github.com/microsoft/DeepSpeed
cd DeepSpeed
DS_BUILD_CPU_ADAM=1 DS_BUILD_AIO=1 DS_BUILD_UTILS=1 pip install -e .
ds_report # check the deespeed status
  • [Optional] For BMTrain training, install BMTrain (>= 0.2.2)
git clone https://github.com/OpenBMB/BMTrain
cd BMTrain
python setup.py install
  • [Optional] For BMInf low-resource inference, install BMInf
pip install bminf

pip install flash-attn
  • [Tips] For single-node docker environments, we need to set up ports for your ssh. e.g., root@127.0.0.1 with port 711
>>> vim ~/.ssh/config
Host 127.0.0.1
    Hostname 127.0.0.1
    Port 7110
    User root
  • [Tips] For multi-node docker environments, generate ssh keys and copy the public key to all nodes (in ~/.ssh/)
>>> ssh-keygen -t rsa -C "xxx@xxx.com"

Load model and tokenizer

We provide the AutoLoad class to load the model and tokenizer quickly, for example:

from flagai.auto_model.auto_loader import AutoLoader

auto_loader = AutoLoader(
    task_name="title-generation",
    model_name="BERT-base-en"
)
model = auto_loader.get_model()
tokenizer = auto_loader.get_tokenizer()

This example is for the title_generation task, and you can also model other tasks by modifying the task_name. Then you can use the model and tokenizer to fine-tune or test.

Examples

1. Predictor

We provide the Predictor class to predict for different tasks, for example:

from flagai.model.predictor.predictor import Predictor
predictor = Predictor(model, tokenizer)
test_data = [
    "Four minutes after the red card, Emerson Royal nodded a corner into the path of the unmarked Kane at the far post, who nudged the ball in for his 12th goal in 17 North London derby appearances. Arteta's misery was compounded two minutes after half-time when Kane held the ball up in front of goal and teed up Son to smash a shot beyond a crowd of defenders to make it 3-0.The goal moved the South Korea talisman a goal behind Premier League top scorer Mohamed Salah on 21 for the season, and he looked perturbed when he was hauled off with 18 minutes remaining, receiving words of consolation from Pierre-Emile Hojbjerg.Once his frustrations have eased, Son and Spurs will look ahead to two final games in which they only need a point more than Arsenal to finish fourth.",
]

for text in test_data:
    print(
        predictor.predict_generate_beamsearch(text,
                                              out_max_length=50,
                                              beam_size=3))

This example is for the seq2seq task, where we can get beam-search results by calling the predict_generate_beamsearch function. In addition, we also support prediction for tasks such as NER and title generate.

2. NER

from flagai.auto_model.auto_loader import AutoLoader
from flagai.model.predictor.predictor import Predictor

task_name = "ner"
model_name = "RoBERTa-base-ch"
target = ["O", "B-LOC", "I-LOC", "B-ORG", "I-ORG", "B-PER", "I-PER"]
maxlen = 256

auto_loader = AutoLoader(task_name,
                         model_name=model_name,
                         load_pretrain_params=True,
                         class_num=len(target))

model = auto_loader.get_model()
tokenizer = auto_loader.get_tokenizer()

predictor = Predictor(model, tokenizer)

test_data = [
    "6月15日,河南省文物考古研究所曹操高陵文物队公开发表声明承认:“从来没有说过出土的珠子是墓主人的",
    "4月8日,北京冬奥会、冬残奥会总结表彰大会在人民大会堂隆重举行。习近平总书记出席大会并发表重要讲话。在讲话中,总书记充分肯定了北京冬奥会、冬残奥会取得的优异成绩,全面回顾了7年筹办备赛的不凡历程,深入总结了筹备举办北京冬奥会、冬残奥会的宝贵经验,深刻阐释了北京冬奥精神,对运用好冬奥遗产推动高质量发展提出明确要求。",
    "当地时间8日,欧盟委员会表示,欧盟各成员国政府现已冻结共计约300亿欧元与俄罗斯寡头及其他被制裁的俄方人员有关的资产。",
    "这一盘口状态下英国必发公司亚洲盘交易数据显示博洛尼亚热。而从欧赔投注看,也是主队热。巴勒莫两连败,",
]

for t in test_data:
    entities = predictor.predict_ner(t, target, maxlen=maxlen)
    result = {}
    for e in entities:
        if e[2] not in result:
            result[e[2]] = [t[e[0]:e[1] + 1]]
        else:
            result[e[2]].append(t[e[0]:e[1] + 1])
    print(f"result is {result}")

3. Semantic Matching example

from flagai.auto_model.auto_loader import AutoLoader
from flagai.model.predictor.predictor import Predictor

maxlen = 256

auto_loader = AutoLoader("semantic-matching",
                         model_name="RoBERTa-base-ch",
                         load_pretrain_params=True,
                         class_num=2)
model = auto_loader.get_model()
tokenizer = auto_loader.get_tokenizer()

predictor = Predictor(model, tokenizer)

test_data = [["后悔了吗", "你有没有后悔"], ["打开自动横屏", "开启移动数据"],
             ["我觉得你很聪明", "你聪明我是这么觉得"]]

for text_pair in test_data:
    print(predictor.predict_cls_classifier(text_pair))

LICENSE

The majority of FlagAI is licensed under the Apache 2.0 license, however portions of the project are available under separate license terms:

News

  • [9 June 2023] release v1.7.0, Support Aquila #324;
  • [31 Mar 2023] release v1.6.3, Support AltCLIP-m18 #303 and AltDiffusion-m18 #302;
  • [17 Mar 2023] release v1.6.2, Support application of new optimizers #266, and added a new gpt model name 'GPT2-base-en' for English;
  • [2 Mar 2023] release v1.6.1, Support Galactica model #234; BMInf, a low-resource inference package #238, and examples for p-tuning #227
  • [12 Jan 2023] release v1.6.0, support a new parallel lib called BMTrain and integate Flash Attention to speedup training of BERT and ViT models, examples in FlashAttentionBERT and FlashAttentionViT. Also add the contrastive search based text generation method SimCTG and DreamBooth finetuning based on AltDiffusion, examples in AltDiffusionNaruto.
  • [28 Nov 2022] release v1.5.0, support 1.1B EVA-CLIP and [ALM: A large Arabic Language Model based on GLM], examples in ALM
  • [10 Nov 2022] release v1.4.0, support AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities, examples in AltCLIP and AltDiffusion
  • [29 Aug 2022] release v1.3.0, Added CLIP module and redesigned tokenizer APIs in #81
  • [21 Jul 2022] release v1.2.0, ViTs are supported in #71
  • [29 Jun 2022] release v1.1.0, support OPTs downloading and inference/fine-tuning #63
  • [17 May 2022] made our first contribution in #1

Platforms supported

Misc

↳ Stargazers, thank you for your support!

Stargazers repo roster for @FlagAI-Open/FlagAI

↳ Forkers, thank you for your support!

Forkers repo roster for @FlagAI-Open/FlagAI

↳ Star History

Star History Chart]