|
About Language Model
|
|
0
|
997
|
August 28, 2020
|
|
探讨下实现文本语法/语义完整性的可能
|
|
1
|
964
|
August 13, 2024
|
|
“哥哥和姐姐“”的抽象意义表示
|
|
0
|
583
|
April 7, 2024
|
|
Dialogue-oriented Pre-training
|
|
0
|
707
|
April 6, 2023
|
|
DialoGPT: Large-Scale Generative Pre-training for Conversational Response
|
|
0
|
867
|
April 5, 2023
|
|
An Embarrassingly Simple Method to Mitigate Undesirable Properties of
|
|
0
|
934
|
December 27, 2022
|
|
Deduplicating Training Data Makes Language Models Better
|
|
0
|
2343
|
December 25, 2022
|
|
Prompt-free and Efficient Few-shot Learning with Language Models
|
|
0
|
924
|
December 15, 2022
|
|
Making Transformers Solve Compositional Tasks
|
|
0
|
1128
|
November 16, 2022
|
|
Adaptive Testing and Debugging of NLP Models
|
|
0
|
1207
|
September 26, 2022
|
|
如何实现拼音转汉字
|
|
0
|
1030
|
August 25, 2022
|
|
Sparse Progressive Distillation: Resolving Overfitting under
|
|
0
|
1068
|
May 29, 2022
|
|
SKEP: Sentiment Knowledge Enhanced Pre-training for Sentiment Analysis
|
|
1
|
1397
|
February 17, 2022
|
|
Phrase-BERT: Improved Phrase Embeddings from BERT with an Application to...
|
|
0
|
987
|
January 24, 2022
|
|
Finetuning Pretrained Transformers into RNNs
|
|
0
|
810
|
January 16, 2022
|
|
Block Pruning For Faster Transformers
|
|
0
|
967
|
January 16, 2022
|
|
What’s in Your Head? Emergent Behaviour in Multi-Task Transformer Models
|
|
0
|
928
|
December 30, 2021
|
|
AdapterDrop: On the Efficiency of Adapters in Transformers
|
|
0
|
1391
|
December 30, 2021
|
|
Frustratingly Simple Pretraining Alternatives to Masked Language Modeling
|
|
0
|
1002
|
November 30, 2021
|
|
Condenser: a Pre-training Architecture for Dense Retrieval
|
|
0
|
1654
|
November 21, 2021
|
|
How to Train BERT with an Academic Budget
|
|
0
|
1009
|
November 17, 2021
|
|
The Power of Scale for Parameter-Efficient Prompt Tuning
|
|
0
|
1070
|
November 15, 2021
|
|
Constrained Language Models Yield Few-Shot Semantic Parsers
|
|
0
|
1076
|
November 15, 2021
|
|
ConvFiT: Conversational Fine-Tuning of Pretrained Language Models
|
|
0
|
1018
|
November 10, 2021
|
|
#EMNLP21#干细胞假说:神经网络也具备干细胞难成全才
|
|
0
|
1012
|
November 6, 2021
|
|
Improving Sequence-to-Sequence Pre-training via Sequence Span Rewriting
|
|
0
|
832
|
November 6, 2021
|
|
Lower Perplexity is Not Always Human-Like
|
|
0
|
930
|
November 5, 2021
|
|
Bird’s Eye: Probing for Linguistic Graph Structures with a Simple Information-Theoretic Approach
|
|
0
|
779
|
November 2, 2021
|
|
When Do You Need Billions of Words of Pretraining Data?
|
|
0
|
893
|
October 26, 2021
|
|
CPT: A Pre-Trained Unbalanced Transformer for Both Chinese Language...
|
|
1
|
1035
|
October 18, 2021
|