Oxen.ai Blog

Welcome to the Oxen.ai blog 🐂

The team at Oxen.ai is dedicated to helping AI practictioners go from research to production. To help enable this, we host a research paper club on Fridays called ArXiv Dives, where we go over state of the art research and how you can apply it to your own work.

Take a look at our Arxiv Dives, Practical ML Dives as well as a treasure trove of content on how to go from raw datasets to production ready AI/ML systems. We cover everything from prompt engineering, fine-tuning, computer vision, natural language understanding, generative ai, data engineering, to best practices when versioning your data. So, dive in and explore – we're excited to share our journey and learnings with you 🚀

The Best AI Data Version Control Tools [2025]
The Best AI Data Version Control Tools [2025]

Data is often seen as static. It's common to just dump your data into S3 buckets in tarballs or upload to Hugging Face and leave it at that. Yet nowadays, data needs to evolve and ...

Greg Schoeninger
Greg Schoeninger
Dec 27, 2024
6 min read
OpenCoder: The OPEN Cookbook For Top-Tier Code LLMs
OpenCoder: The OPEN Cookbook For Top-Tier Code LLMs

Welcome to the last arXiv Dive of 2024! Every other week we have been diving into interesting research papers in AI/ML. In this blog we’ll be diving into Open Coder, a paper and co...

Mathias
Mathias
Dec 24, 2024
- Arxiv Dives
14 min read
LLaVA-CoT: Let Vision Language Models Reason Step-By-Step
LLaVA-CoT: Let Vision Language Models Reason Step-By-Step

When it comes to large language models, it is still the early innings. Many of them still hallucinate, fail to follow instructions, or generally don’t work. The only way to combat ...

Greg Schoeninger
Greg Schoeninger
Dec 10, 2024
- Arxiv Dives
12 min read
How Upcycling MoEs Beat Dense LLMs
How Upcycling MoEs Beat Dense LLMs

In this Arxiv Dive, Nvidia researcher, Ethan He, presents his co-authored work Upcycling LLMs in Mixture of Experts (MoE). He goes into what a MoE is, the challenges behind upcycli...

Greg Schoeninger
Greg Schoeninger
Nov 19, 2024
- Arxiv Dives
1 min read
Thinking LLMs: General Instruction Following with Thought Generation
Thinking LLMs: General Instruction Following with Thought Generation

The release of OpenAI-O1 has motivated a lot of people to think deeply about…thoughts 💭. Thinking before you speak is a skill that some people have better than others 😉, but a sk...

Greg Schoeninger
Greg Schoeninger
Nov 11, 2024
- Arxiv Dives
14 min read
The Prompt Report Part 2: Plan and Solve, Tree of Thought, and Decomposition Prompting
The Prompt Report Part 2: Plan and Solve, Tree of Thought, and Decomposition Prompting

In the last blog, we went over prompting techniques 1-3 of The Prompt Report. This arXiv Dive, we were lucky to have the authors of the paper join us to go through some of the more...

Greg Schoeninger
Greg Schoeninger
Oct 31, 2024
- Arxiv Dives
17 min read
The Prompt Report Part 1: A Systematic Survey of Prompting Techniques
The Prompt Report Part 1: A Systematic Survey of Prompting Techniques

For this blog we are switching it up a bit. In past Arxiv Dives, we have gone deep into the underlying model architectures and techniques that make large language models and other ...

Mathias
Mathias
Oct 9, 2024
- Arxiv Dives
12 min read
arXiv Dive: How Flux and Rectified Flow Transformers Work
arXiv Dive: How Flux and Rectified Flow Transformers Work

Flux made quite a splash with its release on August 1st, 2024 as the new state of the art generative image model outperforming SDXL, SDXL-Turbo, Pixart, and DALL-E. While the model...

Mathias
Mathias
Sep 18, 2024
- Arxiv Dives
9 min read
How Well Can Llama 3.1 8B Detect Political Spam? [4/4]
How Well Can Llama 3.1 8B Detect Political Spam? [4/4]

It only took about 11 minutes to fine-tuned Llama 3.1 8B on our political spam synthetic dataset using ReFT. While this is extremely fast, beating out our previous record of 14 min...

Eric Laurence
Eric Laurence
Sep 14, 2024
3 min read
Fine-Tuning Llama 3.1 8B in Under 12 Minutes [3/4]
Fine-Tuning Llama 3.1 8B in Under 12 Minutes [3/4]

Meta has recently released Llama 3.1, including their 405 billion parameter model which is the most capable open model to date and the first open model on the same level as GPT 4. ...

Eric Laurence
Eric Laurence
Sep 5, 2024
3 min read