PGB: Permutation and Grouping for BERT Pruning 


Vol. 50,  No. 6, pp. 503-510, Jun.  2023
10.5626/JOK.2023.50.6.503


PDF

  Abstract

Recently, pre-trained Transformer-based models have been actively used for various artificial intelligence tasks, such as natural language processing and image recognition. However, these models have billions of parameters, which require significant computation for inference, and may be subject to many limitations for use in resource-limited environments. To address this problem, we propose PGB(Permutation Grouped BERT pruning), a new group-based structured pruning method for Transformer models. PGB effectively finds a way to change the optimal attention order according to resource constraints, and prunes unnecessary heads based on the importance of the heads to minimize the information loss in the model. Through various comparison experiments, PGB shows better performance in terms of inference speed and accuracy loss than the other existing structured pruning methods for the pre-trained BERT model.


  Statistics
Cumulative Counts from November, 2022
Multiple requests among the same browser session are counted as one view. If you mouse over a chart, the values of data points will be shown.


  Cite this article

[IEEE Style]

H. Lim and D. Choi, "PGB: Permutation and Grouping for BERT Pruning," Journal of KIISE, JOK, vol. 50, no. 6, pp. 503-510, 2023. DOI: 10.5626/JOK.2023.50.6.503.


[ACM Style]

Hye-Min Lim and Dong-Wan Choi. 2023. PGB: Permutation and Grouping for BERT Pruning. Journal of KIISE, JOK, 50, 6, (2023), 503-510. DOI: 10.5626/JOK.2023.50.6.503.


[KCI Style]

임혜민, 최동완, "PGB: BERT 프루닝을 위한 순서 변경 규칙 및 그룹화," 한국정보과학회 논문지, 제50권, 제6호, 503~510쪽, 2023. DOI: 10.5626/JOK.2023.50.6.503.


[Endnote/Zotero/Mendeley (RIS)]  Download


[BibTeX]  Download



Search




Journal of KIISE

  • ISSN : 2383-630X(Print)
  • ISSN : 2383-6296(Electronic)
  • KCI Accredited Journal

Editorial Office

  • Tel. +82-2-588-9240
  • Fax. +82-2-521-1352
  • E-mail. chwoo@kiise.or.kr