tests/testthat/_snaps/dataset_bert.md

dataset_bert works

Code
  test_result_df
Output
  <bert_dataset>
    Inherits from: <dataset>
    Public:
      .getitem: function (index) 
      .length: function () 
      clone: function (deep = FALSE) 
      initialize: function (x, y = NULL, tokenizer = tokenize_bert, n_tokens = 128L) 
      token_types: 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2  ...
      tokenized_text: 102 102 2071 2063 3794 3794 103 103 2146 2037 2063 2179  ...
      y: 1 2
Code
  test_result_df$token_types
Output
       [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14]
  [1,]    1    1    1    1    2    2    2    2    2     2     2     2     2     2
  [2,]    1    1    1    1    2    2    2    2    2     2     2     2     2     2
       [,15] [,16] [,17] [,18] [,19] [,20] [,21] [,22] [,23] [,24] [,25] [,26]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,27] [,28] [,29] [,30] [,31] [,32] [,33] [,34] [,35] [,36] [,37] [,38]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,39] [,40] [,41] [,42] [,43] [,44] [,45] [,46] [,47] [,48] [,49] [,50]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,51] [,52] [,53] [,54] [,55] [,56] [,57] [,58] [,59] [,60] [,61] [,62]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,63] [,64] [,65] [,66] [,67] [,68] [,69] [,70] [,71] [,72] [,73] [,74]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,75] [,76] [,77] [,78] [,79] [,80] [,81] [,82] [,83] [,84] [,85] [,86]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,87] [,88] [,89] [,90] [,91] [,92] [,93] [,94] [,95] [,96] [,97] [,98]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,99] [,100] [,101] [,102] [,103] [,104] [,105] [,106] [,107] [,108]
  [1,]     2      2      2      2      2      2      2      2      2      2
  [2,]     2      2      2      2      2      2      2      2      2      2
       [,109] [,110] [,111] [,112] [,113] [,114] [,115] [,116] [,117] [,118]
  [1,]      2      2      2      2      2      2      2      2      2      2
  [2,]      2      2      2      2      2      2      2      2      2      2
       [,119] [,120] [,121] [,122] [,123] [,124] [,125] [,126] [,127] [,128]
  [1,]      2      2      2      2      2      2      2      2      2      2
  [2,]      2      2      2      2      2      2      2      2      2      2
Code
  test_result_df$tokenized_text
Output
       [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14]
  [1,]  102 2071 3794  103 2146 2063  103    1    1     1     1     1     1     1
  [2,]  102 2063 3794  103 2037 2179  103    1    1     1     1     1     1     1
       [,15] [,16] [,17] [,18] [,19] [,20] [,21] [,22] [,23] [,24] [,25] [,26]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,27] [,28] [,29] [,30] [,31] [,32] [,33] [,34] [,35] [,36] [,37] [,38]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,39] [,40] [,41] [,42] [,43] [,44] [,45] [,46] [,47] [,48] [,49] [,50]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,51] [,52] [,53] [,54] [,55] [,56] [,57] [,58] [,59] [,60] [,61] [,62]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,63] [,64] [,65] [,66] [,67] [,68] [,69] [,70] [,71] [,72] [,73] [,74]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,75] [,76] [,77] [,78] [,79] [,80] [,81] [,82] [,83] [,84] [,85] [,86]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,87] [,88] [,89] [,90] [,91] [,92] [,93] [,94] [,95] [,96] [,97] [,98]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,99] [,100] [,101] [,102] [,103] [,104] [,105] [,106] [,107] [,108]
  [1,]     1      1      1      1      1      1      1      1      1      1
  [2,]     1      1      1      1      1      1      1      1      1      1
       [,109] [,110] [,111] [,112] [,113] [,114] [,115] [,116] [,117] [,118]
  [1,]      1      1      1      1      1      1      1      1      1      1
  [2,]      1      1      1      1      1      1      1      1      1      1
       [,119] [,120] [,121] [,122] [,123] [,124] [,125] [,126] [,127] [,128]
  [1,]      1      1      1      1      1      1      1      1      1      1
  [2,]      1      1      1      1      1      1      1      1      1      1
Code
  test_result_df$y
Output
  [1] 1 2
Code
  test_result_factor
Output
  <bert_dataset>
    Inherits from: <dataset>
    Public:
      .getitem: function (index) 
      .length: function () 
      clone: function (deep = FALSE) 
      initialize: function (x, y = NULL, tokenizer = tokenize_bert, n_tokens = 128L) 
      token_types: 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2  ...
      tokenized_text: 102 102 2071 2063 3794 3794 103 103 2146 2037 2063 2179  ...
      y: 1 2
Code
  test_result_factor$token_types
Output
       [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14]
  [1,]    1    1    1    1    2    2    2    2    2     2     2     2     2     2
  [2,]    1    1    1    1    2    2    2    2    2     2     2     2     2     2
       [,15] [,16] [,17] [,18] [,19] [,20] [,21] [,22] [,23] [,24] [,25] [,26]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,27] [,28] [,29] [,30] [,31] [,32] [,33] [,34] [,35] [,36] [,37] [,38]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,39] [,40] [,41] [,42] [,43] [,44] [,45] [,46] [,47] [,48] [,49] [,50]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,51] [,52] [,53] [,54] [,55] [,56] [,57] [,58] [,59] [,60] [,61] [,62]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,63] [,64] [,65] [,66] [,67] [,68] [,69] [,70] [,71] [,72] [,73] [,74]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,75] [,76] [,77] [,78] [,79] [,80] [,81] [,82] [,83] [,84] [,85] [,86]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,87] [,88] [,89] [,90] [,91] [,92] [,93] [,94] [,95] [,96] [,97] [,98]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,99] [,100] [,101] [,102] [,103] [,104] [,105] [,106] [,107] [,108]
  [1,]     2      2      2      2      2      2      2      2      2      2
  [2,]     2      2      2      2      2      2      2      2      2      2
       [,109] [,110] [,111] [,112] [,113] [,114] [,115] [,116] [,117] [,118]
  [1,]      2      2      2      2      2      2      2      2      2      2
  [2,]      2      2      2      2      2      2      2      2      2      2
       [,119] [,120] [,121] [,122] [,123] [,124] [,125] [,126] [,127] [,128]
  [1,]      2      2      2      2      2      2      2      2      2      2
  [2,]      2      2      2      2      2      2      2      2      2      2
Code
  test_result_factor$tokenized_text
Output
       [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14]
  [1,]  102 2071 3794  103 2146 2063  103    1    1     1     1     1     1     1
  [2,]  102 2063 3794  103 2037 2179  103    1    1     1     1     1     1     1
       [,15] [,16] [,17] [,18] [,19] [,20] [,21] [,22] [,23] [,24] [,25] [,26]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,27] [,28] [,29] [,30] [,31] [,32] [,33] [,34] [,35] [,36] [,37] [,38]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,39] [,40] [,41] [,42] [,43] [,44] [,45] [,46] [,47] [,48] [,49] [,50]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,51] [,52] [,53] [,54] [,55] [,56] [,57] [,58] [,59] [,60] [,61] [,62]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,63] [,64] [,65] [,66] [,67] [,68] [,69] [,70] [,71] [,72] [,73] [,74]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,75] [,76] [,77] [,78] [,79] [,80] [,81] [,82] [,83] [,84] [,85] [,86]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,87] [,88] [,89] [,90] [,91] [,92] [,93] [,94] [,95] [,96] [,97] [,98]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,99] [,100] [,101] [,102] [,103] [,104] [,105] [,106] [,107] [,108]
  [1,]     1      1      1      1      1      1      1      1      1      1
  [2,]     1      1      1      1      1      1      1      1      1      1
       [,109] [,110] [,111] [,112] [,113] [,114] [,115] [,116] [,117] [,118]
  [1,]      1      1      1      1      1      1      1      1      1      1
  [2,]      1      1      1      1      1      1      1      1      1      1
       [,119] [,120] [,121] [,122] [,123] [,124] [,125] [,126] [,127] [,128]
  [1,]      1      1      1      1      1      1      1      1      1      1
  [2,]      1      1      1      1      1      1      1      1      1      1
Code
  test_result_factor$y
Output
  [1] 1 2
Code
  test_result_null
Output
  <bert_dataset>
    Inherits from: <dataset>
    Public:
      .getitem: function (index) 
      .length: function () 
      clone: function (deep = FALSE) 
      initialize: function (x, y = NULL, tokenizer = tokenize_bert, n_tokens = 128L) 
      token_types: 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2  ...
      tokenized_text: 102 102 2071 2063 3794 3794 103 103 2146 2037 2063 2179  ...
      y:
Code
  test_result_null$token_types
Output
       [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14]
  [1,]    1    1    1    1    2    2    2    2    2     2     2     2     2     2
  [2,]    1    1    1    1    2    2    2    2    2     2     2     2     2     2
       [,15] [,16] [,17] [,18] [,19] [,20] [,21] [,22] [,23] [,24] [,25] [,26]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,27] [,28] [,29] [,30] [,31] [,32] [,33] [,34] [,35] [,36] [,37] [,38]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,39] [,40] [,41] [,42] [,43] [,44] [,45] [,46] [,47] [,48] [,49] [,50]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,51] [,52] [,53] [,54] [,55] [,56] [,57] [,58] [,59] [,60] [,61] [,62]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,63] [,64] [,65] [,66] [,67] [,68] [,69] [,70] [,71] [,72] [,73] [,74]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,75] [,76] [,77] [,78] [,79] [,80] [,81] [,82] [,83] [,84] [,85] [,86]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,87] [,88] [,89] [,90] [,91] [,92] [,93] [,94] [,95] [,96] [,97] [,98]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,99] [,100] [,101] [,102] [,103] [,104] [,105] [,106] [,107] [,108]
  [1,]     2      2      2      2      2      2      2      2      2      2
  [2,]     2      2      2      2      2      2      2      2      2      2
       [,109] [,110] [,111] [,112] [,113] [,114] [,115] [,116] [,117] [,118]
  [1,]      2      2      2      2      2      2      2      2      2      2
  [2,]      2      2      2      2      2      2      2      2      2      2
       [,119] [,120] [,121] [,122] [,123] [,124] [,125] [,126] [,127] [,128]
  [1,]      2      2      2      2      2      2      2      2      2      2
  [2,]      2      2      2      2      2      2      2      2      2      2
Code
  test_result_null$tokenized_text
Output
       [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14]
  [1,]  102 2071 3794  103 2146 2063  103    1    1     1     1     1     1     1
  [2,]  102 2063 3794  103 2037 2179  103    1    1     1     1     1     1     1
       [,15] [,16] [,17] [,18] [,19] [,20] [,21] [,22] [,23] [,24] [,25] [,26]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,27] [,28] [,29] [,30] [,31] [,32] [,33] [,34] [,35] [,36] [,37] [,38]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,39] [,40] [,41] [,42] [,43] [,44] [,45] [,46] [,47] [,48] [,49] [,50]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,51] [,52] [,53] [,54] [,55] [,56] [,57] [,58] [,59] [,60] [,61] [,62]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,63] [,64] [,65] [,66] [,67] [,68] [,69] [,70] [,71] [,72] [,73] [,74]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,75] [,76] [,77] [,78] [,79] [,80] [,81] [,82] [,83] [,84] [,85] [,86]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,87] [,88] [,89] [,90] [,91] [,92] [,93] [,94] [,95] [,96] [,97] [,98]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,99] [,100] [,101] [,102] [,103] [,104] [,105] [,106] [,107] [,108]
  [1,]     1      1      1      1      1      1      1      1      1      1
  [2,]     1      1      1      1      1      1      1      1      1      1
       [,109] [,110] [,111] [,112] [,113] [,114] [,115] [,116] [,117] [,118]
  [1,]      1      1      1      1      1      1      1      1      1      1
  [2,]      1      1      1      1      1      1      1      1      1      1
       [,119] [,120] [,121] [,122] [,123] [,124] [,125] [,126] [,127] [,128]
  [1,]      1      1      1      1      1      1      1      1      1      1
  [2,]      1      1      1      1      1      1      1      1      1      1
Code
  test_result_null$y
Output
  integer(0)
Code
  test_result_tokens
Output
  <bert_dataset>
    Inherits from: <dataset>
    Public:
      .getitem: function (index) 
      .length: function () 
      clone: function (deep = FALSE) 
      initialize: function (x, y = NULL, tokenizer = tokenize_bert, n_tokens = 128L) 
      token_types: 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2  ...
      tokenized_text: 102 102 2071 2063 3794 3794 103 103 2146 2037 2063 2179  ...
      y: 1 2
Code
  test_result_tokens$token_types
Output
       [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14]
  [1,]    1    1    1    1    2    2    2    2    2     2     2     2     2     2
  [2,]    1    1    1    1    2    2    2    2    2     2     2     2     2     2
       [,15] [,16] [,17] [,18] [,19] [,20] [,21] [,22] [,23] [,24] [,25] [,26]
  [1,]     2     2     2     2     2     2     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2     2     2     2     2     2     2
       [,27] [,28] [,29] [,30] [,31] [,32]
  [1,]     2     2     2     2     2     2
  [2,]     2     2     2     2     2     2
Code
  test_result_tokens$tokenized_text
Output
       [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14]
  [1,]  102 2071 3794  103 2146 2063  103    1    1     1     1     1     1     1
  [2,]  102 2063 3794  103 2037 2179  103    1    1     1     1     1     1     1
       [,15] [,16] [,17] [,18] [,19] [,20] [,21] [,22] [,23] [,24] [,25] [,26]
  [1,]     1     1     1     1     1     1     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1     1     1     1     1     1     1
       [,27] [,28] [,29] [,30] [,31] [,32]
  [1,]     1     1     1     1     1     1
  [2,]     1     1     1     1     1     1
Code
  test_result_tokens$y
Output
  [1] 1 2


macmillancontentscience/torchtransformers documentation built on Aug. 6, 2023, 5:35 a.m.