pyspark.sql.Catalog.analyzeTable#

Catalog.analyzeTable(tableName, noScan=False)[source]#

Computes table statistics (same as SQL ANALYZE TABLE COMPUTE STATISTICS).

New in version 4.2.0.

Parameters
tableNamestr

Table or view name. May be qualified with catalog and database (namespace).

noScanbool, optional

If True, use NOSCAN mode (reuse existing column statistics where possible).

Examples

>>> _ = spark.sql("DROP TABLE IF EXISTS tbl_an_doc")
>>> _ = spark.sql("CREATE TABLE tbl_an_doc (id INT) USING csv")
>>> _ = spark.sql("INSERT INTO tbl_an_doc VALUES (1)")
>>> spark.catalog.analyzeTable("tbl_an_doc")
>>> spark.catalog.analyzeTable("tbl_an_doc", noScan=True)
>>> _ = spark.sql("DROP TABLE tbl_an_doc")