新闻动态
成员信息
科学研究
联系我们
中文 (简体)
中文 (简体)
English
Kaiyan Zhang
最新
Fourier Position Embedding: Enhancing Attentions Periodic Extension for Length Generalization
Free Process Rewards without Process Labels
How to Synthesize Text Data without Model Collapse?
MedXpertQA: Benchmarking Expert-Level Medical Reasoning and Understanding
GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning
Technologies on Effectiveness and Efficiency: A Survey of State Spaces Models
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling
Process reinforcement through implicit rewards.
OpenPRM: Building Open-domain Process-based Reward Models with Preference Trees
Automating exploratory proteomics research via language models.
Retrieval-Augmented Visual Question Answering via Built-in Autoregressive Search Engines.
Efficient Diffusion Models: A Comprehensive Survey from Principles to Practices
Scalable Efficient Training of Large Language Models with Low-dimensional Projected Attention
Towards Building Specialized Generalist AI with System 1 and System 2 Fusion
UltraMedical: Building Specialized Generalists in Biomedicine
Fast and Slow Generating: An Empirical Study on Large and Small Language Models Collaborative Decoding
Intuitive Fine-Tuning: Towards Simplifying Alignment into a Single Process
Online DPO: Online Direct Preference Optimization with Fast-Slow Chasing
CoGenesis: A Framework Collaborating Large and Small Language Models for Secure Context-Aware Instruction Following
SMR: State Memory Replay for Long Sequence Modeling
PaD: Program-aided Distillation Specializes Large Models in Reasoning
Large Language Models as Biomedical Hypothesis Generators: A Comprehensive Evaluation
Generative Multi-Modal Knowledge Retrieval with Large Language Models
Large Language Models are Zero Shot Hypothesis Proposers
CRaSh: Clustering, Removing, and Sharing Enhance Fine-tuning without Full Large Language Model
引用
×