Gemini 2.5: Pushing the Frontier with Advanced Reasoning, Multimodality, Long Context, and Next Generation Agentic Capabilities 06-23
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition 07-08
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence 06-30
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence 06-30
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence 06-30
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence 06-30
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence 06-30
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence 06-30
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models 06-30
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models 06-30
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models 06-30
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models 06-30
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models 06-30
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models 06-30
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models 06-30
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search 07-01
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation 07-02
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition 07-08
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition 07-08
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition 07-08
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition 07-08
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition 07-08
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition 07-08
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08
Helix Parallelism: Rethinking Sharding Strategies for Interactive Multi-Million-Token LLM Decoding 07-08