Picture for Neel Jain

Neel Jain

Democratizing AI: Open-source Scalable LLM Training on GPU-based Supercomputers

Add code
Feb 12, 2025
Viaarxiv icon

Exploiting Sparsity for Long Context Inference: Million Token Contexts on Commodity GPUs

Add code
Feb 10, 2025
Figure 1 for Exploiting Sparsity for Long Context Inference: Million Token Contexts on Commodity GPUs
Figure 2 for Exploiting Sparsity for Long Context Inference: Million Token Contexts on Commodity GPUs
Figure 3 for Exploiting Sparsity for Long Context Inference: Million Token Contexts on Commodity GPUs
Figure 4 for Exploiting Sparsity for Long Context Inference: Million Token Contexts on Commodity GPUs
Viaarxiv icon

Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach

Add code
Feb 07, 2025
Viaarxiv icon

Refusal Tokens: A Simple Way to Calibrate Refusals in Large Language Models

Add code
Dec 09, 2024
Figure 1 for Refusal Tokens: A Simple Way to Calibrate Refusals in Large Language Models
Figure 2 for Refusal Tokens: A Simple Way to Calibrate Refusals in Large Language Models
Figure 3 for Refusal Tokens: A Simple Way to Calibrate Refusals in Large Language Models
Figure 4 for Refusal Tokens: A Simple Way to Calibrate Refusals in Large Language Models
Viaarxiv icon

LiveBench: A Challenging, Contamination-Free LLM Benchmark

Add code
Jun 27, 2024
Figure 1 for LiveBench: A Challenging, Contamination-Free LLM Benchmark
Figure 2 for LiveBench: A Challenging, Contamination-Free LLM Benchmark
Figure 3 for LiveBench: A Challenging, Contamination-Free LLM Benchmark
Figure 4 for LiveBench: A Challenging, Contamination-Free LLM Benchmark
Viaarxiv icon

GenQA: Generating Millions of Instructions from a Handful of Prompts

Add code
Jun 14, 2024
Figure 1 for GenQA: Generating Millions of Instructions from a Handful of Prompts
Figure 2 for GenQA: Generating Millions of Instructions from a Handful of Prompts
Figure 3 for GenQA: Generating Millions of Instructions from a Handful of Prompts
Figure 4 for GenQA: Generating Millions of Instructions from a Handful of Prompts
Viaarxiv icon

Be like a Goldfish, Don't Memorize! Mitigating Memorization in Generative LLMs

Add code
Jun 14, 2024
Figure 1 for Be like a Goldfish, Don't Memorize! Mitigating Memorization in Generative LLMs
Figure 2 for Be like a Goldfish, Don't Memorize! Mitigating Memorization in Generative LLMs
Figure 3 for Be like a Goldfish, Don't Memorize! Mitigating Memorization in Generative LLMs
Figure 4 for Be like a Goldfish, Don't Memorize! Mitigating Memorization in Generative LLMs
Viaarxiv icon

Transformers Can Do Arithmetic with the Right Embeddings

Add code
May 27, 2024
Figure 1 for Transformers Can Do Arithmetic with the Right Embeddings
Figure 2 for Transformers Can Do Arithmetic with the Right Embeddings
Figure 3 for Transformers Can Do Arithmetic with the Right Embeddings
Figure 4 for Transformers Can Do Arithmetic with the Right Embeddings
Viaarxiv icon

NEFTune: Noisy Embeddings Improve Instruction Finetuning

Add code
Oct 10, 2023
Viaarxiv icon

Baseline Defenses for Adversarial Attacks Against Aligned Language Models

Add code
Sep 04, 2023
Figure 1 for Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Figure 2 for Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Figure 3 for Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Figure 4 for Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Viaarxiv icon