Skip to content

[FEATURE REQ] Contribution for Spark 4.0 OLTP Connector for Cosmos DB #47607

@pramodpatil07

Description

@pramodpatil07

I am planning to upgrade our production data landscape (Databricks) to DBR 17.3 LTS, which runs on Apache Spark 4.0. Currently, the azure-cosmos-spark connector only supports up to Spark 3.5.

I am aware of issue #46290 stating a planned release for Q1 2026. However, due to our internal upgrade timelines, we need this support sooner.

Proposal I am a Senior Data Engineer with extensive Scala/Spark experience and I am willing to contribute the development of the Spark 4.0 module to this repository.

My proposed plan:

Create a new module sdk/cosmos/azure-cosmos-spark_4-0_2-13.

Port the existing 3.5 logic to support Scala 2.13 (required by Spark 4.0).

Address breaking changes in the Spark 4.0 DataSource V2 API.

Question for Maintainers (@kushagraThapar / @TheovanKraay)

Has internal work already started on this?

If yes, is there a feature branch I can contribute to?

If no, are you open to accepting a PR for this new module if I follow the contribution guidelines?

Setup

OS: Windows 11

IDE: IntelliJ / VS Code

Build: Maven 3.8+ / Java 17

Metadata

Metadata

Labels

ClientThis issue points to a problem in the data-plane of the library.CosmosService AttentionWorkflow: This issue is responsible by Azure service team.customer-reportedIssues that are reported by GitHub users external to the Azure organization.feature-requestThis issue requires a new behavior in the product in order be resolved.needs-team-attentionWorkflow: This issue needs attention from Azure service team or SDK team

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions