PeerLM logoPeerLM
All Comparisons

Amazon: Nova Pro 1.0 vs Google: Gemini 3.1 Pro Preview: Coding Performance with 10 Evaluators

We evaluate Amazon: Nova Pro 1.0 and Google: Gemini 3.1 Pro Preview to determine the leader in Coding Performance with 10 Evaluators.

Amazon: Nova Pro 1.0

1.6

/ 10

vs

Google: Gemini 3.1 Pro Preview

8.4

/ 10

Key Findings

Overall PerformanceGoogle: Gemini 3.1 Pro Preview

Ranked #1 with an overall score of 8.42, significantly outperforming the competition.

Cost EfficiencyAmazon: Nova Pro 1.0

Offers a much lower cost per output token at $0.004953 compared to Gemini's $0.01227.

Instruction FollowingGoogle: Gemini 3.1 Pro Preview

Demonstrated superior capability in adhering to complex coding constraints.

Specifications

SpecAmazon: Nova Pro 1.0Google: Gemini 3.1 Pro Preview
Provideramazongoogle
Context Length300K1.0M
Input Price (per 1M tokens)$0.80$2.00
Output Price (per 1M tokens)$3.20$12.00
Max Output Tokens5,12065,536
Tierstandardadvanced

Our Verdict

Google: Gemini 3.1 Pro Preview is the superior choice for high-stakes coding, offering unmatched accuracy and instruction following. While Amazon: Nova Pro 1.0 provides a more budget-friendly option for simpler tasks, it struggles to maintain the same level of performance in complex coding scenarios.

Overview

In the rapidly evolving landscape of generative AI, selecting the right model for software engineering tasks is critical. This comparative analysis examines Amazon: Nova Pro 1.0 vs Google: Gemini 3.1 Pro Preview, specifically focusing on their capabilities within a Coding Performance with 10 Evaluators benchmark suite. By utilizing PeerLM’s comparative ranking methodology, we provide an objective look at how these models handle complex code generation and instruction adherence.

Benchmark Results

The evaluation reveals a significant performance gap between the two models. Google: Gemini 3.1 Pro Preview secured the top position, demonstrating a substantial lead in overall coding efficacy compared to Amazon: Nova Pro 1.0. The following table summarizes the performance metrics observed during this run.

ModelRankOverall ScoreAvg Completion TokensCost per Output Token
Google: Gemini 3.1 Pro Preview18.421612$0.01227
Amazon: Nova Pro 1.021.58100$0.004953

Criteria Breakdown

The evaluation focused on two core pillars essential for development workflows: Accuracy and Instruction Following. In comparative testing, Google: Gemini 3.1 Pro Preview consistently outperformed its peer, showing a greater ability to translate complex requirements into functional, accurate code blocks. Amazon: Nova Pro 1.0, while highly efficient in terms of raw resource consumption, struggled to match the depth and precision required by the 10 evaluators in this specific coding context.

Accuracy

Google: Gemini 3.1 Pro Preview demonstrated a superior grasp of syntax and logic, resulting in highly accurate code generation that required minimal revision. Amazon: Nova Pro 1.0 showed lower accuracy, suggesting challenges with the specific complexity of the coding prompts provided in this suite.

Instruction Following

Coding tasks often require strict adherence to style guides, specific libraries, or architectural constraints. Gemini 3.1 Pro Preview proved to be the more reliable model for these structured requirements, whereas Nova Pro 1.0 showed difficulty in consistently maintaining alignment with the provided instructions.

Cost & Latency

Cost efficiency is a major factor for enterprise-scale coding operations. While Amazon: Nova Pro 1.0 is significantly cheaper per output token at $0.004953, the trade-off is reflected in the lower output token count and lower overall score. In contrast, Google: Gemini 3.1 Pro Preview commands a higher price per token ($0.01227) but yields a much higher volume of complex, high-quality completion tokens, making it a more powerful tool for intensive coding tasks.

Use Cases

  • Google: Gemini 3.1 Pro Preview: Ideal for complex software architecture design, large-scale refactoring, and tasks requiring high levels of reasoning and precise instruction adherence.
  • Amazon: Nova Pro 1.0: Suited for lightweight coding tasks, simple script generation, or scenarios where budget constraints are the primary driver and token volume is kept low.

Verdict

For developers prioritizing coding quality and instruction adherence, Google: Gemini 3.1 Pro Preview is the clear winner. While Amazon: Nova Pro 1.0 offers lower costs, its performance in this specific benchmark suggests it is not yet ready to compete with the high-reasoning capabilities of the Gemini 3.1 Pro Preview in rigorous coding environments.

Backed by real data

View the Full Evaluation Report

See every response, score, and evaluator judgment behind this comparison. All data from PeerLM's blind evaluation pipeline.

View Report

Run your own comparison

Test Amazon: Nova Pro 1.0 vs Google: Gemini 3.1 Pro Preview with your own prompts and criteria. Get results in minutes.

Start Free

Get a free managed report

We'll run a full evaluation with your real prompts and deliver a detailed recommendation. Free for qualified teams.

Request Report

Methodology

Evaluated using PeerLM's blind evaluation pipeline with 4 responses per model across 2 criteria.