Skip to content

docs: bring README up to date with v0.6.0 / v0.6.1 merges#783

Closed
anandgupta42 wants to merge 1 commit intomainfrom
docs/readme-changelog-v0.6.x
Closed

docs: bring README up to date with v0.6.0 / v0.6.1 merges#783
anandgupta42 wants to merge 1 commit intomainfrom
docs/readme-changelog-v0.6.x

Conversation

@anandgupta42
Copy link
Copy Markdown
Contributor

@anandgupta42 anandgupta42 commented May 3, 2026

What does this PR do?

Updates README.md so it reflects what actually shipped between v0.5.12 and v0.6.1 on main. The Changelog section had stalled at v0.5.11 (March 2026), and the warehouse / LLM provider lists were missing additions from the same period.

Changelog section — added one-liners for the missing releases, trimmed three lower-signal v0.5.x entries (v0.5.5, v0.5.3, v0.5.1) to keep the section length steady:

Release Highlight
v0.6.1 BigQuery finops fixes (INFORMATION_SCHEMA columns, multi-region), advisory anti-slop workflow
v0.6.0 Cross-warehouse data parity (data_diff + /data-parity), Microsoft Fabric / MSSQL, Databricks AI Gateway, Bedrock custom-endpoints guide
v0.5.21 Automated dbt unit test generation, manifest parse cache, dialect-aware sql_explain
v0.5.20 Altimate model auto-selection, password URI percent-encoding, trace list pagination
v0.5.18 Native GitLab MR review, Altimate LLM Gateway provider, MCP env-var interpolation
v0.5.16 ClickHouse driver support, agent loop detection, error classification overhaul
v0.5.14 MongoDB driver support, skill follow-up suggestions

Key Features — new "Cross-Warehouse Data Parity" entry for the data_diff tool and /data-parity skill (v0.6.0). This is a major user-visible capability that wasn't surfaced anywhere on the landing README.

Supported Warehouses — added Microsoft Fabric (v0.6.0).

Works with Any LLM — added Databricks AI Gateway (v0.6.0, #649), Snowflake Cortex (v0.5.6), and LM Studio (v0.5.7). The latter two were already supported but had been missed in earlier README sweeps.

No code changes — this is README only.

Type of change

  • Bug fix (non-breaking change which fixes an issue)
  • New feature (non-breaking change which adds functionality)
  • Breaking change (fix or feature that would cause existing functionality to not work as expected)
  • Documentation only (no code changes; updates README.md)

Issue for this PR

N/A — documentation refresh; no specific tracking issue. The changes are derived directly from CHANGELOG.md entries for v0.5.14 → v0.6.1 already merged to main.

How did you verify your code works?

  • Diff review — confirmed every added bullet maps to an existing entry in CHANGELOG.md.
  • Cross-checks against the codebase / merged PRs:
  • Markdown render — visually scanned the diff for table / bullet / heading correctness; no broken markdown.
  • No code paths touched — typecheck still cache-hits on push (run via the pre-push hook): 5 successful, 4 cached, 1 cache miss (altimate-code; passes).
  • Marker Guard — N/A, README is not under packages/opencode/src/.

Checklist

  • My code follows the style guidelines of this project
  • I have performed a self-review of my own code
  • I have commented my code, particularly in hard-to-understand areas (N/A — docs only)
  • I have made corresponding changes to the documentation (this is the documentation change)
  • My changes generate no new warnings
  • I have added tests that prove my fix is effective or that my feature works (N/A — docs only)

Summary by cubic

Updates README.md to reflect v0.5.14–v0.6.1: adds the Cross‑Warehouse Data Parity feature (data_diff, /data-parity), updates supported warehouses and LLM providers, and extends the changelog through v0.6.1.
Adds Microsoft Fabric and the Databricks AI Gateway, Snowflake Cortex, and LM Studio; trims a few lower‑signal v0.5.x changelog entries to keep the section concise.

Written for commit ea643a7. Summary will update on new commits.

Summary by CodeRabbit

  • New Features

    • Cross-warehouse data parity: Compare rows and columns across multiple warehouses with flexible execution algorithms and a profile-only mode designed for sensitive environments.
  • Documentation

    • Extended warehouse support to include Microsoft Fabric and Oracle.
    • Expanded the list of supported LLM providers available for integration.
    • Updated changelog with latest release information.

The README's changelog stopped at v0.5.11 and the warehouse / provider lists
were missing additions shipped in the v0.5.14–v0.6.1 range. This catches
them up:

- Changelog: added v0.6.1, v0.6.0, v0.5.21, v0.5.20, v0.5.18, v0.5.16,
  v0.5.14; trimmed v0.5.5 / v0.5.3 / v0.5.1 to keep the section length
  steady.
- Key Features: new "Cross-Warehouse Data Parity" entry for the `data_diff`
  tool and `/data-parity` skill shipped in v0.6.0 (#493, #705).
- Supported Warehouses: added Microsoft Fabric (v0.6.0, via T-SQL dialect
  + `tedious` / Entra ID auth in `data_diff` MSSQL/Fabric path).
- Works with Any LLM: added Databricks AI Gateway (v0.6.0, #649),
  Snowflake Cortex (v0.5.6), and LM Studio (v0.5.7) — the last two
  were also missing.

No code changes; README only.
Copy link
Copy Markdown

@claude claude Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Claude Code Review

This repository is configured for manual code reviews. Comment @claude review to trigger a review and subscribe this PR to future pushes, or @claude review once for a one-time review.

Tip: disable this comment in your organization's Code Review settings.

@coderabbitai
Copy link
Copy Markdown

coderabbitai Bot commented May 3, 2026

📝 Walkthrough

Walkthrough

README documentation is updated with a new cross-warehouse data parity feature section, extended warehouse support (Microsoft Fabric, Oracle), expanded LLM provider integrations (Databricks AI Gateway, Snowflake Cortex, others), and refreshed changelog entries for recent releases.

Changes

Documentation Updates

Layer / File(s) Summary
New Feature Documentation
README.md (lines 125–127)
Introduces "Cross-Warehouse Data Parity" subsection describing row/column diffing across 12 warehouses, data_diff tool, /data-parity skill, and multiple execution algorithms (auto, joindiff, hashdiff, profile, cascade) with profile-only mode for sensitive environments.
Supported Infrastructure
README.md (lines 157–165)
Extends "Supported Warehouses" list to include Microsoft Fabric and Oracle; expands "Works with Any LLM" provider list with Databricks AI Gateway, Snowflake Cortex, LM Studio, and other integrations.
Release History
README.md (lines 184–193)
Refreshes "Changelog" section with newer release versions (v0.6.1, v0.6.0, and v0.5.x entries), replacing prior visible bullet set.

Estimated Code Review Effort

🎯 1 (Trivial) | ⏱️ ~3 minutes


🐰 A README hops with glee,
New warehouses dance—from one to three!
Parity checked, and LLMs align,
Changelog sparkles with versions so fine! ✨
The docs now shine, oh what a sight!

🚥 Pre-merge checks | ✅ 4 | ❌ 1

❌ Failed checks (1 inconclusive)

Check name Status Explanation Resolution
Description check ❓ Inconclusive The description is comprehensive and mostly follows the template structure with detailed 'What does this PR do?' and 'Type of change' sections, but is missing the required 'PINEAPPLE' marker at the top and lacks a formal 'Test Plan' section. Add 'PINEAPPLE' at the very top of the description as required for AI-generated contributions, and include an explicit 'Test Plan' section.
✅ Passed checks (4 passed)
Check name Status Explanation
Title check ✅ Passed The title accurately summarizes the main change—updating README.md to reflect v0.6.0/v0.6.1 releases—and is concise and specific.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.
Linked Issues check ✅ Passed Check skipped because no linked issues were found for this pull request.
Out of Scope Changes check ✅ Passed Check skipped because no linked issues were found for this pull request.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Commit unit tests in branch docs/readme-changelog-v0.6.x

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share
Review rate limit: 7/8 reviews remaining, refill in 7 minutes and 30 seconds.

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link
Copy Markdown

@coderabbitai coderabbitai Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

🧹 Nitpick comments (1)
README.md (1)

126-126: ⚡ Quick win

Clarify the “12 warehouses” scope vs overall supported warehouse list.

At Line 126, data parity is described as “across 12 warehouses,” while Line 157 lists 13 supported warehouses overall. A short qualifier like “12 SQL warehouses for parity” (or naming exclusions) would prevent confusion.

Also applies to: 157-157

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@README.md` at line 126, The README's summary sentence "Row-level and
column-level diffing across 12 warehouses..." conflicts with the full supported
list (13 warehouses); update that sentence to clarify scope by specifying "12
SQL warehouses for parity" or similar and reference the full supported list,
e.g., change the phrase to "Row-level and column-level diffing across 12 SQL
warehouses for parity (see full list of 13 supported warehouses below)" or
explicitly note which warehouse is excluded from parity; edit the sentence
containing `data_diff` and `/data-parity` to add this qualifier so the earlier
summary and the later supported-warehouses list are consistent.
🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@README.md`:
- Line 126: README currently claims "partitioned execution for 100M+ row tables"
while the tool docs in packages/opencode/src/altimate/tools/data-diff.ts
recommend partitioning for ">10M rows"; update one or the other so both are
consistent: either change the README phrasing to match the tool doc (e.g.,
"recommended for >10M rows") or update the tool doc to justify and match "100M+
rows" and add a brief citation to benchmark context; reference the strings
"100M+ row tables" in README and the threshold text ">10M rows" in data-diff.ts
(and mention `data_diff`/`/data-parity` in the doc update) when making the
change.

---

Nitpick comments:
In `@README.md`:
- Line 126: The README's summary sentence "Row-level and column-level diffing
across 12 warehouses..." conflicts with the full supported list (13 warehouses);
update that sentence to clarify scope by specifying "12 SQL warehouses for
parity" or similar and reference the full supported list, e.g., change the
phrase to "Row-level and column-level diffing across 12 SQL warehouses for
parity (see full list of 13 supported warehouses below)" or explicitly note
which warehouse is excluded from parity; edit the sentence containing
`data_diff` and `/data-parity` to add this qualifier so the earlier summary and
the later supported-warehouses list are consistent.
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: Repository UI

Review profile: CHILL

Plan: Pro

Run ID: bae6fcdc-62cd-475c-82d5-e8d28a4e59da

📥 Commits

Reviewing files that changed from the base of the PR and between a7dea3d and ea643a7.

📒 Files selected for processing (1)
  • README.md

Comment thread README.md
Transpile SQL between Snowflake, BigQuery, Databricks, Redshift, PostgreSQL, MySQL, SQL Server, and DuckDB.

### Cross-Warehouse Data Parity
Row-level and column-level diffing across 12 warehouses (including cross-dialect pairs like Postgres ↔ Snowflake or Databricks ↔ Fabric) via the `data_diff` tool and `/data-parity` skill. Five algorithms — `auto`, `joindiff`, `hashdiff`, `profile`, and `cascade` — partitioned execution for 100M+ row tables, and a `profile`-only mode for PII / PHI / PCI environments.
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor | ⚡ Quick win

Align scale claim with tool docs to avoid overstating capability.

At Line 126, README says partitioned execution is for “100M+ row tables,” but the tool description in packages/opencode/src/altimate/tools/data-diff.ts frames partitioning as recommended for “>10M rows.” Please make these thresholds consistent (or cite benchmark context if 100M+ is intentional).

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@README.md` at line 126, README currently claims "partitioned execution for
100M+ row tables" while the tool docs in
packages/opencode/src/altimate/tools/data-diff.ts recommend partitioning for
">10M rows"; update one or the other so both are consistent: either change the
README phrasing to match the tool doc (e.g., "recommended for >10M rows") or
update the tool doc to justify and match "100M+ rows" and add a brief citation
to benchmark context; reference the strings "100M+ row tables" in README and the
threshold text ">10M rows" in data-diff.ts (and mention
`data_diff`/`/data-parity` in the doc update) when making the change.

Copy link
Copy Markdown

@cubic-dev-ai cubic-dev-ai Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

1 issue found across 1 file

Prompt for AI agents (unresolved issues)

Check if these issues are valid — if so, understand the root cause of each and fix them. If appropriate, use sub-agents to investigate and fix each issue separately.


<file name="README.md">

<violation number="1" location="README.md:126">
P3: The threshold "100M+ row tables" appears inconsistent with the tool implementation in `data-diff.ts`, which recommends partitioned execution for >10M rows. Consider aligning this claim with the actual tool threshold to avoid overstating the capability.</violation>
</file>

Reply with feedback, questions, or to request a fix. Tag @cubic-dev-ai to re-run a review.

Comment thread README.md
Transpile SQL between Snowflake, BigQuery, Databricks, Redshift, PostgreSQL, MySQL, SQL Server, and DuckDB.

### Cross-Warehouse Data Parity
Row-level and column-level diffing across 12 warehouses (including cross-dialect pairs like Postgres ↔ Snowflake or Databricks ↔ Fabric) via the `data_diff` tool and `/data-parity` skill. Five algorithms — `auto`, `joindiff`, `hashdiff`, `profile`, and `cascade` — partitioned execution for 100M+ row tables, and a `profile`-only mode for PII / PHI / PCI environments.
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

P3: The threshold "100M+ row tables" appears inconsistent with the tool implementation in data-diff.ts, which recommends partitioned execution for >10M rows. Consider aligning this claim with the actual tool threshold to avoid overstating the capability.

Prompt for AI agents
Check if this issue is valid — if so, understand the root cause and fix it. At README.md, line 126:

<comment>The threshold "100M+ row tables" appears inconsistent with the tool implementation in `data-diff.ts`, which recommends partitioned execution for >10M rows. Consider aligning this claim with the actual tool threshold to avoid overstating the capability.</comment>

<file context>
@@ -122,6 +122,9 @@ Credit analysis, expensive query detection, warehouse right-sizing, unused resou
 Transpile SQL between Snowflake, BigQuery, Databricks, Redshift, PostgreSQL, MySQL, SQL Server, and DuckDB.
 
+### Cross-Warehouse Data Parity
+Row-level and column-level diffing across 12 warehouses (including cross-dialect pairs like Postgres ↔ Snowflake or Databricks ↔ Fabric) via the `data_diff` tool and `/data-parity` skill. Five algorithms — `auto`, `joindiff`, `hashdiff`, `profile`, and `cascade` — partitioned execution for 100M+ row tables, and a `profile`-only mode for PII / PHI / PCI environments.
+
 ### PII Detection & Safety
</file context>
Suggested change
Row-level and column-level diffing across 12 warehouses (including cross-dialect pairs like Postgres ↔ Snowflake or Databricks ↔ Fabric) via the `data_diff` tool and `/data-parity` skill. Five algorithms — `auto`, `joindiff`, `hashdiff`, `profile`, and `cascade` — partitioned execution for 100M+ row tables, and a `profile`-only mode for PII / PHI / PCI environments.
Row-level and column-level diffing across 12 warehouses (including cross-dialect pairs like Postgres ↔ Snowflake or Databricks ↔ Fabric) via the `data_diff` tool and `/data-parity` skill. Five algorithms — `auto`, `joindiff`, `hashdiff`, `profile`, and `cascade` — partitioned execution for 10M+ row tables, and a `profile`-only mode for PII / PHI / PCI environments.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant