For most common read and write operations on Delta tables, you can use Apache Spark reader and writer APIs (see Table batch reads and writes and Table streaming reads and writes). However, there are some operations that are specific to Delta Lake and you must use Delta Lake programmatic APIs. This article describes these programmatic APIs.
Some programmatic APIs are still evolving and are indicated with the Evolving qualifier in the API docs.
Databricks ensures binary compatibility between the Delta Lake project and Delta Lake in Databricks Runtime. Compatibility matrixes lists the Delta Lake API version packaged in each Databricks Runtime version and a link to the respective API documentation.