Print Email Facebook Twitter Evaluating Large Language Model Performance on User and Language Defined Elements in Code Title Evaluating Large Language Model Performance on User and Language Defined Elements in Code Author Mekkes, Erik (TU Delft Electrical Engineering, Mathematics and Computer Science) Contributor van Deursen, A. (mentor) Izadi, M. (mentor) Katzy, J.B. (mentor) Nadeem, A. (graduation committee) Degree granting institution Delft University of Technology Programme Computer Science and Engineering Project CSE3000 Research Project Date 2023-06-28 Abstract Large Language Models of code have seen significant jumps in performance recently. However, these jumps tend to accompany a notable and perhaps concerning increase in scale and costs. We contribute an evaluation of prediction performance with respect to model size by assessing the layer-wise progression for language and user-defined elements in code, using a new technique of Tuned Lenses. We show that language-defined elements can be predicted more accurately in earlier layers of the PolyCoder model than user-defined elements and contribute an evaluation of the attention mechanism, which shows patterns that explain such aspects of performance and indicate areas of missed potential. These findings encourage research into the internal prediction performance for other characteristic aspects of code and could lead to the introduction of new methods that make use of these characteristics to improve performance without relying on scaling. Subject Large Language ModelsCode CompletionPerformance EvaluationGPTTransformersAttention MechanismTuned LensToken Classification To reference this document use: http://resolver.tudelft.nl/uuid:5d3350fb-ad8e-4975-9ce2-c541a3ec64a5 Part of collection Student theses Document type bachelor thesis Rights © 2023 Erik Mekkes Files PDF E.J.Mekkes_Evaluating_Lar ... _Code..pdf 791.66 KB Close viewer /islandora/object/uuid:5d3350fb-ad8e-4975-9ce2-c541a3ec64a5/datastream/OBJ/view