research
          
      
      ∙
      05/17/2023
    Token-wise Decomposition of Autoregressive Language Model Hidden States for Analyzing Model Predictions
While there is much recent interest in studying why Transformer-based la...
          
            research
          
      
      ∙
      04/22/2023
    Transformer-Based LM Surprisal Predicts Human Reading Times Best with About Two Billion Training Tokens
Recent psycholinguistic studies have drawn conflicting conclusions about...
          
            research
          
      
      ∙
      12/23/2022
    Why Does Surprisal From Larger Transformer-Based Language Models Provide a Poorer Fit to Human Reading Times?
This work presents a detailed linguistic analysis into why larger Transf...
          
            research
          
      
      ∙
      12/21/2022
     
             
  
  
     
                             share
 share