AI
AI Papers
LLM
CMMLU: Measuring massive multitask language understanding in Chinese
CodeFuse-13B: A Pretrained Multi-lingual Code Large Language Model
CROSSCODEEVAL: A Diverse and Multilingual Benchmark for Cross-File Code Completion
CRUXEval-X: A Benchmark for Multilingual Code Reasoning, Understanding and Execution
CRUXEval: A Benchmark for Code Reasoning, Understanding and Execution
DeepSeek-Coder: When the Large Language Model Meets Programming – The Rise of Code Intelligence
DemoCraft: Using In-Context Learning to Improve Code Generation in Large Language Models
Efficient Training of Language Models to Fill in the Middle
Evaluating Large Language Models Trained on Code
Evaluation of LLMs on Syntax-Aware Code Fill-in-the-Middle Tasks
InterTrans: Leveraging Transitive Intermediate Translations to Enhance LLM-based Code Translation
LLM-Eval: Unified Multi-Dimensional Automatic Evaluation for Open-Domain Conversations with Large Language Models
McEval: Massively Multilingual Code Evaluation
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
Qwen2.5-Coder Technical Report
SGLang: Efficient Execution of Structured Language Model Programs
Neural Network
Intention Recognition with Recurrent Neural Networks for Dynamic Human-Robot Collaboration
Transformer Model
Attention is All You Need
(
Reading Notes
)
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Transformed-based Map Matching Model with Limited Ground-Truth Data using Transfer-Learning Approach
Transformer++
Transformers: State-of-the-Art Natural Language Processing