DBT (Data Build Tool) quoting refers to the process of wrapping a string or identifier…
Category: article
DBT : Best practice for SQL developers using DBT
When working with DBT, it’s important to follow best practices to ensure that your code is maintainable, scalable, and easy…
DBT : Explain on DBT Project configuration in dbt_project.yml
In this article, we’ll dive into the various settings you can configure in this file, including project name, version, DBT…
DBT : Explain on DBTs models, snapshots, seeds, tests, macros, docs, sources, exposures, metrics, analysis
In this article, we’ll dive into the key concepts of DBT, including models, snapshots, seeds, tests, macros, docs, sources, exposures,…
DBT : Concatenating two or more arrays into a single array in DBT macro.
DBT macros are reusable code snippets that can be used to create custom functions for SQL transformations. In this article,…
DBT : How do you use DBT to document your data pipeline?
DBT helps maintain a clear and detailed documentation of the entire data pipeline, making it easier for team members to…
DBT : How do you use DBT to version and deploy your models
With DBT, you can automate the process of transforming raw data into business insights, and version and deploy those models…
PySpark: Explanation of PySpark Full Outer Join with example.
One of the most commonly used operations in PySpark is joining two dataframes together. Full outer join is one of…
PySpark : Reading from multiple files , how to get the file which contain each record in PySpark [input_file_name]
pyspark.sql.functions.input_file_name One of the most useful features of PySpark is the ability to access metadata about the input files being…
PySpark : Exploding a column of arrays or maps into multiple rows in a Spark DataFrame [posexplode_outer]
pyspark.sql.functions.posexplode_outer The posexplode_outer function in PySpark is part of the pyspark.sql.functions module and is used to explode a column of…
PySpark : Transforming a column of arrays or maps into multiple columns, with one row for each element in the array or map [posexplode]
pyspark.sql.functions.posexplode The posexplode function in PySpark is part of the pyspark.sql.functions module and is used to transform a column of…