About 30,300 results
Open links in new tab
  1. Spark SQL — PySpark master documentation - Databricks

    This page gives an overview of all public Spark SQL API.

  2. Apache Spark overview - Azure Databricks | Microsoft Learn

    Jan 16, 2026 · Apache Spark is the technology powering compute clusters and SQL warehouses in Azure Databricks. This page provides an overview of the documentation in this section. Get started …

  3. PySpark on Databricks | Databricks on AWS

    Jan 16, 2026 · With Spark DataFrames, you can efficiently read, write, transform, and analyze data using Python and SQL, which means you are always leveraging the full power of Spark.

  4. PySpark Overview — PySpark 4.1.0 documentation - Apache Spark

    Dec 11, 2025 · PySpark combines Python’s learnability and ease of use with the power of Apache Spark to enable processing and analysis of data at any size for everyone familiar with Python. PySpark …

  5. SQL language reference - Databricks on AWS

    Jan 13, 2026 · This is a SQL command reference for Databricks SQL and Databricks Runtime. For information about using SQL with Lakeflow Spark Declarative Pipelines, see Pipeline SQL language …

  6. Ultimate Guide To Databricks - livedocs.com

    Jan 13, 2026 · Databricks SQL (formerly SQL Analytics) provides a query editor and dashboard builder for people who just want to ask questions of their data without learning Spark APIs.

  7. What is Spark SQL? Intro, Features & Benefits | Databricks

    Spark SQL is a Spark module for structured data processing. It provides a programming abstraction called DataFrames and can also act as a distributed SQL query engine. It enables unmodified …

  8. GitHub - apache/spark: Apache Spark - A unified analytics engine for ...

    Spark is a unified analytics engine for large-scale data processing. It provides high-level APIs in Scala, Java, Python, and R (Deprecated), and an optimized engine that supports general computation …

  9. PySpark on Azure Databricks - Azure Databricks | Microsoft Learn

    Jan 16, 2026 · DataFrames are the primary objects in Apache Spark. A DataFrame is a dataset organized into named columns. You can think of a DataFrame like a spreadsheet or a SQL table, a …

  10. PySpark reference - Azure Databricks | Microsoft Learn

    Jan 16, 2026 · Main classes for working with PySpark SQL, including SparkSession and DataFrame fundamentals. The entry point for reading data and executing SQL queries in PySpark applications. …