Microsoft and Tsinghua University have developed a 7B-parameter AI coding model that outperforms 14B rivals using only ...
Like all AI models based on the Transformer architecture, the large language models (LLMs) that underpin today’s coding ...
Abstract: Large language models (LLMs) have made significant progress in the field of natural language processing, but research on MATLAB code generation remains relatively scarce. As a programming ...
Abstract: Deep code models are vulnerable to adversarial attacks, making it possible for semantically identical inputs to trigger different responses. Current black-box attack methods typically ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results