PeerLM logoPeerLM
Back to Blog
open-sourcellm-evaluationqwengemmaliquidai

Best New Open-Source Models You Haven't Tried Yet: Qwen3.5 vs Gemma 4 vs LiquidAI LFM2

PeerLM TeamMay 11, 2026

The State of Open-Source LLMs in 2026

The landscape of open-source artificial intelligence is evolving faster than ever. While frontier proprietary models often dominate the headlines, developers and AI practitioners are increasingly turning to open-source and open-weights models to balance cost, control, and performance. As of May 2026, a new wave of models has emerged that offers impressive capabilities previously reserved for the largest, most expensive API-based solutions.

In this post, we explore three standout model families that you might have missed: Qwen3.5, Gemma 4, and LiquidAI LFM2. These models represent the current pinnacle of accessible AI, providing high context windows and specialized parameter counts that are ideal for production-grade applications.

Why Switch to Open-Source?

Beyond the obvious benefits of data privacy and reduced vendor lock-in, modern open-source models now provide a compelling price-to-performance ratio. When evaluating these models, we look at three primary metrics: Cost per Million Tokens (M), Context Length, and Parameter efficiency.

Model Comparison Table

ModelInput Cost/MOutput Cost/MContext LengthParams
LiquidAI LFM2-24B-A2B$0.03$0.1233K13b-30b
Qwen3.5-9B$0.05$0.15256K7b-13b
Gemma 4 26B A4B$0.06$0.33262K13b-30b
Gemma 4 31B$0.13$0.38262K30b-70b
Qwen3.5-35B-A3B$0.16$1.30262K30b-70b

Deep Dive: The Top Contenders

1. LiquidAI: LFM2-24B-A2B

The LiquidAI LFM2 series has set a new benchmark for cost-efficiency. At just $0.03 per million input tokens, it is one of the most economical choices for high-volume inference tasks. Its 33K context window is sufficient for most standard RAG (Retrieval-Augmented Generation) pipelines, making it a perfect "workhorse" model for lightweight data processing.

2. Qwen3.5-9B: The Context King

For developers working with massive documents or long-form codebase analysis, the Qwen3.5-9B is a revelation. Offering a 256K context window at a price point of $0.05 per million input tokens, it punches significantly above its weight class. Its parameter efficiency (7b-13b range) allows for rapid inference, which is critical for real-time applications.

3. Gemma 4: Google's Open-Weights Powerhouse

Gemma 4 26B and 31B represent a significant leap in reasoning capabilities for the open-weights community. With a standard 262K context window, these models are designed to handle complex logic and multi-step reasoning tasks. While slightly more expensive than the Qwen or LiquidAI offerings, the quality of output for creative and analytical tasks makes it a strong contender for production environments where accuracy is paramount.

Practical Recommendations

  • For Budget-Conscious RAG: Use LiquidAI LFM2-24B-A2B. It provides the best cost-to-performance ratio for standard retrieval tasks.
  • For Long-Context Analysis: The Qwen3.5-9B is unmatched in its ability to process large amounts of data at a very low cost.
  • For Complex Reasoning: Deploy Gemma 4 31B. Its architecture is better suited for tasks requiring nuance, deep understanding, and structured output.

Conclusion

The gap between proprietary models and open-source alternatives is closing. By experimenting with models like the Qwen3.5 series or Google's Gemma 4, developers can significantly reduce their dependency on expensive frontier APIs without sacrificing the quality of their applications. We recommend benchmarking these models against your specific use case using the PeerLM platform to see which architecture fits your latency and accuracy requirements best.

Ready to find the best model for your use case?

Run blind evaluations with your real prompts. Free to start, results in minutes.