Shivchander Sudalairaj
Shivchander Sudalairaj's contributions
Article
Building domain-specific LLMs with synthetic data and SDG Hub
Shivchander Sudalairaj
+2
Use the open source SDG Hub to quickly create custom synthetic data pipelines. Train and evaluate your models faster and more efficiently.
Article
SDG Hub: Building synthetic data pipelines with modular blocks
Aditi Saluja
+2
Discover SDG Hub, an open framework for building, composing, and scaling synthetic data pipelines for large language models.
Article
Lessons on reproducing R1-like reasoning in small LLMs
Akash Srivastava
+8
Learn about an efficient inference scaling method that can improve your model's reasoning ability and performance at runtime while saving on compute costs.
Article
On reasoning versus inference-time scaling
Akash Srivastava
+8
Progress in small LLM reasoning: Our Qwen-32B model, using particle filtering, now surpasses o1-preview on Math500.
Article
Granite, LIMO, and small LLM reasoning
Akash Srivastava
+8
On reproducing R1-like reasoning in small LLMs: LIMO dataset ineffective for Llama/Granite; synthetic data generation shows promise but fine-tuning is tricky.
Article
How particle filtering makes small LLMs think big
Akash Srivastava
+8
An update on reproducing R1-like reasoning in small LLMs: Granite models show big gains with particle filtering, outperforming GPT-4o on benchmarks.
Building domain-specific LLMs with synthetic data and SDG Hub
Use the open source SDG Hub to quickly create custom synthetic data pipelines. Train and evaluate your models faster and more efficiently.
SDG Hub: Building synthetic data pipelines with modular blocks
Discover SDG Hub, an open framework for building, composing, and scaling synthetic data pipelines for large language models.
Lessons on reproducing R1-like reasoning in small LLMs
Learn about an efficient inference scaling method that can improve your model's reasoning ability and performance at runtime while saving on compute costs.
On reasoning versus inference-time scaling
Progress in small LLM reasoning: Our Qwen-32B model, using particle filtering, now surpasses o1-preview on Math500.
Granite, LIMO, and small LLM reasoning
On reproducing R1-like reasoning in small LLMs: LIMO dataset ineffective for Llama/Granite; synthetic data generation shows promise but fine-tuning is tricky.
How particle filtering makes small LLMs think big
An update on reproducing R1-like reasoning in small LLMs: Granite models show big gains with particle filtering, outperforming GPT-4o on benchmarks.