Kontext Copilot - An AI assistant for data analytics. Learn more
Expression of Interest
Find Number of Rows of Hive Table via Scala
insights Stats
warning Please login first to view stats information.
Kontext
Code Snippets & Tips
Code snippets and tips for various programming languages/frameworks. All code examples are under MIT or Apache 2.0 license unless specified otherwise.
Code description
To find the number of rows/records in a Hive table, we can use Spark SQL
count
aggregation function: Hive SQL - Aggregate Functions Overview with Examples.This code snippet provide example of Scala code to implement the same. spark-shell
is used directly for simplicity. The code snippet can also run Jupyter Notebooks or Zeppelin with Spark kernel. Alternatively, it can be compiled to jar file and then submit as job via spark-submit
.
Code snippet
val sql="select count(*) from test_db.test_table;"; //sql: String = select count(*) from test_db.test_table; val df = spark.sql(sql); //df: org.apache.spark.sql.DataFrame = [count(1): bigint] df.show() /* +--------+ |count(1)| +--------+ | 5| +--------+ */ val df1 = spark.sql("select * from test_db.test_table;"); //df1: org.apache.spark.sql.DataFrame = [id: int, attr: string] df1.groupBy().count() //res5: org.apache.spark.sql.DataFrame = [count: bigint] println(df1.groupBy().count()) //[count: bigint] df1.groupBy().count().show() /*+-----+ |count| +-----+ | 5| +-----+ */
copyright
This page is subject to Site terms.
comment Comments
No comments yet.