新闻动态
成员信息
科学研究
联系我们
中文 (简体)
中文 (简体)
English
Bowen Zhou
最新
Fourier Position Embedding: Enhancing Attentions Periodic Extension for Length Generalization
Free Process Rewards without Process Labels
How to Synthesize Text Data without Model Collapse?
MedXpertQA: Benchmarking Expert-Level Medical Reasoning and Understanding
GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning
A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond.
Technologies on Effectiveness and Efficiency: A Survey of State Spaces Models
Less is More: Efficient Model Merging with Binary Task Switch
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling
Process reinforcement through implicit rewards.
Advancing LLM Reasoning Generalists with Preference Trees
OpenPRM: Building Open-domain Process-based Reward Models with Preference Trees
Towards AI-45^{\circ} Law: A Roadmap to Trustworthy AGI
Automating exploratory proteomics research via language models.
Retrieval-Augmented Visual Question Answering via Built-in Autoregressive Search Engines.
Empowering private tutoring by chaining large language models.
Neural Residual Diffusion Models for Deep Scalable Vision Generation
Efficient Diffusion Models: A Comprehensive Survey from Principles to Practices
MSI-Agent: Incorporating Multi-Scale Insight into Embodied Agents for Superior Planning and Decision-Making
On the token distance modeling ability of higher RoPE attention dimension
Scalable Efficient Training of Large Language Models with Low-dimensional Projected Attention
Safe-SD: Safe and Traceable Stable Diffusion with Text Prompt Trigger for Invisible Generative Watermarking
Towards Building Specialized Generalist AI with System 1 and System 2 Fusion
Exploring Adversarial Robustness of Deep State Space Models
UltraMedical: Building Specialized Generalists in Biomedicine
Enhancing Adversarial Transferability via Information Bottleneck Constraints
Fast and Slow Generating: An Empirical Study on Large and Small Language Models Collaborative Decoding
Intuitive Fine-Tuning: Towards Simplifying Alignment into a Single Process
Online DPO: Online Direct Preference Optimization with Fast-Slow Chasing
CoGenesis: A Framework Collaborating Large and Small Language Models for Secure Context-Aware Instruction Following
SMR: State Memory Replay for Long Sequence Modeling
Trust in Internal or External Knowledge? Generative Multi-Modal Entity Linking with Knowledge Retriever
On Large Language Models' Hallucination with Regard to Known Facts
PaD: Program-aided Distillation Specializes Large Models in Reasoning
Contrastive Augmented Graph2Graph Memory Interaction for Few Shot Continual Learning
Investigating Deep Watermark Security: An Adversarial Transferability Perspective
Large Language Models as Biomedical Hypothesis Generators: A Comprehensive Evaluation
Interactive Continual Learning: Fast and Slow Thinking
LAKE-RED: Camouflaged Images Generation by Latent Background Knowledge Retrieval-Augmented Diffusion
Generative AI for Complex Scenarios: Language Models are Sequence Processors
AdapEdit: Spatio-Temporal Guided Adaptive Editing Algorithm for Text-Based Continuity-Sensitive Image Editing
Generative Multi-Modal Knowledge Retrieval with Large Language Models
LMD: Faster Image Reconstruction with Latent Masking Diffusion
Large Language Models are Zero Shot Hypothesis Proposers
CRaSh: Clustering, Removing, and Sharing Enhance Fine-tuning without Full Large Language Model
Enhancing Chat Language Models by Scaling High-quality Instructional Conversations
Sparse Low-rank Adaptation of Pre-trained Language Models
Improving Robustness of Intent Detection Under Adversarial Attacks: A Geometric Constraint Perspective
SAM Struggles in Concealed Scenes -- Empirical Study on Segment Anything
Trustworthy AI: From Principles to Practices
引用
×