You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
作者您好。心中有几个疑问,希望您能不吝赐教
1.pre-train上来就是一堆超参(这些超参在哪个文件里面的);pre-train部分的最后一句是训练把,而且后面带了一堆参数?到底我要输入什么指令从而接下去运行。
2.我的服务器只有一个gpu,要运行你的代码,是不是要改一些配置?但是到底要改哪些参数
3.数据集的路径在哪个文件,没看到有
4."we use the English Wikipedia corpus and BookCorpus (Zhu et al., 2015) for pre-training. By concatenating these two datasets, we obtain a corpus with roughly 16GB in size. We set the vocabulary size (sub-word tokens) as 32,768. We use the GLUE (General Language Understanding Evaluation) dataset (Wang et al., 2018) as the downstream tasks to evaluate the performance of the pre-trained models".这个是原文中的内容,为什么要concatenating 两个数据集,具体怎么concat
The text was updated successfully, but these errors were encountered:
作者您好。心中有几个疑问,希望您能不吝赐教
1.pre-train上来就是一堆超参(这些超参在哪个文件里面的);pre-train部分的最后一句是训练把,而且后面带了一堆参数?到底我要输入什么指令从而接下去运行。
2.我的服务器只有一个gpu,要运行你的代码,是不是要改一些配置?但是到底要改哪些参数
3.数据集的路径在哪个文件,没看到有
4."we use the English Wikipedia corpus and BookCorpus (Zhu et al., 2015) for pre-training. By concatenating these two datasets, we obtain a corpus with roughly 16GB in size. We set the vocabulary size (sub-word tokens) as 32,768. We use the GLUE (General Language Understanding Evaluation) dataset (Wang et al., 2018) as the downstream tasks to evaluate the performance of the pre-trained models".这个是原文中的内容,为什么要concatenating 两个数据集,具体怎么concat
The text was updated successfully, but these errors were encountered: