Skip to content

Latest commit

 

History

History
39 lines (24 loc) · 2.68 KB

model_card.md

File metadata and controls

39 lines (24 loc) · 2.68 KB

PAGnol: French Generative Models -- Model Card

This model card was inspired by Model Cards for Model Reporting, Mitchell et al. 2018.

Model details

  • Model description: PAGNol is a series of large-scale generative models for the French language;
  • Model architecture: GPT-like, trained with a language modelling objective, with up to 1.5B parameters;
  • Model history: PAGnol models were trained over the month of April 2021, and trained on data cutting off in 2019;
  • Dataset: trained on a dataset built with CCNet, similar to the one used for CamemBERT-Large, some research models trained with OSCAR;
  • Organization: LightOn, in cooperation with the ALMAnaCH team of Inria;
  • Licence: PAGnol models and inference code are available under the MIT Licence;
  • Paper: coming soon, more information available here;
  • Contact: [email protected].

Intended uses

PAGnol is geared towards free-form text generation. It is best used for creative writing, without strong constraints on its outputs. It may be fine-tuned to specific forms and styles of text generation. However, in line with previous work (T5, Raffel et al. 2019), we found it is not competitive when finetuned on specific tasks (classification, question answering, etc.)

PAGnol trained on OSCAR is downloadable only for research purposes.

We encourage further research on PAGnol zero-shot abilities as well as bias and fairness issues. If you are interested in these topics, you can get in touch with us.

Limitations and bias

PAGnol is not grounded, and cannot distinguish between facts and fiction.

To enhance output quality, we trained PAGnol on CCNet, a dataset filtered by a language model to match Wikipedia-like writing. However, this does not prevent PAGnol from generating offensive, suggestive, or biased content. Additional filtering of its outputs is recommended for most use cases.

Use judgement and discretion before deploying PAGnol. In particular, we recommend performing a study of use-case specific biases and limitations.

PAGnol trained on OSCAR is significantly more likely to produce explicit and offensive content.

Evaluation

We will share additional data on PAGnol end-task performance in a variety of contexts (discriminative/generative tasks, fine-tuning/few-shot) soon.