Generative AI/Language Model

[Large Language Model] FLAN-T5

데이터 세상 2023. 7. 11. 13:38
728x90
반응형

FLAN-T5

https://huggingface.co/docs/transformers/model_doc/flan-t5

 

FLAN-T5

Reinforcement learning models

huggingface.co

 

논문: https://arxiv.org/pdf/2210.11416.pdf

 

An encoder-decoder model based on the T5 

 Scaling Instruction Fine-tuned Language Models

여러 타스크를 통해 fine-tunedT5의 향상된 버전

 


Flan

프롬프팅을 기반으로 하는 사전 교육 방법


FLAN-T5-XL

https://huggingface.co/google/flan-t5-x

 

google/flan-t5-xl · Hugging Face

If you already know T5, FLAN-T5 is just better at everything. For the same number of parameters, these models have been fine-tuned on more than 1000 additional tasks covering also more languages. As mentioned in the first few lines of the abstract : Flan-P

huggingface.co

60 Languages including Korean

taskmaster2, djaym7/wiki_dialog, deepmind/code_contests, lambada, gsm8k, aqua_rat, esnli, quasc qed 포함하는 데이터 세트의 Flan 컬렉션에서 훈련된 T5


FLAN-T5-XXL

https://huggingface.co/google/flan-t5-xxl

 

google/flan-t5-xxl · Hugging Face

If you already know T5, FLAN-T5 is just better at everything. For the same number of parameters, these models have been fine-tuned on more than 1000 additional tasks covering also more languages. As mentioned in the first few lines of the abstract : Flan-P

huggingface.co

 

an 11 billion parameter model based on the Flan-T5 family

Language(s) (NLP): English, German, French

 

728x90
반응형