Utilizing SQL with Python: SQLAlchemy and Pandas

Date:

Share post:


Picture by Writer

 

As a knowledge scientist, you want Python for detailed knowledge evaluation, knowledge visualization, and modeling. Nonetheless, when your knowledge is saved in a relational database, you could use SQL (Structured Question Language) to extract and manipulate the information. However how do you combine SQL with Python to unlock the total potential of your knowledge?

On this tutorial, we are going to study to mix the facility of SQL with the flexibleness of Python utilizing SQLAlchemy and Pandas. We’ll discover ways to hook up with databases, execute SQL queries utilizing SQLAlchemy, and analyze and visualize knowledge utilizing Pandas.

Set up Pandas and SQLAlchemy utilizing:

pip set up pandas sqlalchemy

 

1. Saving the Pandas DataFrame as an SQL Desk

 

To create the SQL desk utilizing the CSV dataset, we are going to:

  1. Create a SQLite database utilizing the SQLAlchemy.
  2. Load the CSV dataset utilizing the Pandas. The countries_poluation dataset consists of the Air High quality Index (AQI) for all nations on the earth from 2017 to 2023. 
  3. Convert all of the AQI columns from object to numerical and drop row with lacking values.
# Import crucial packages
import pandas as pd
import psycopg2
from sqlalchemy import create_engine
 
# creating the brand new db
engine = create_engine(
    "sqlite:///kdnuggets.db")
 
# learn the CSV dataset
knowledge = pd.read_csv("/work/air_pollution new.csv")

col = ['2017', '2018', '2019', '2020', '2021', '2022', '2023']

for s in col:
    knowledge[s] = pd.to_numeric(knowledge[s], errors="coerce")

    knowledge = knowledge.dropna(subset=[s])

 

  1. Save the Pandas dataframe as a SQL desk. The `to_sql` perform requires a desk identify and the engine object.  
# save the dataframe as a SQLite desk
knowledge.to_sql('countries_poluation', engine, if_exists="replace")

 

In consequence, your SQLite database is saved in your file listing. 

 

Deepnote file manager

 

Observe: I’m utilizing Deepnote for this tutorial to run the Python code seamlessly. Deepnote is a free AI Cloud Pocket book that can enable you to rapidly run any knowledge science code. 

 

2. Loading the SQL Desk utilizing Pandas

 

To load all the desk from the SQL database as a Pandas dataframe, we are going to:

  1. Set up the reference to our database by offering the database URL.
  2. Use the `pd.read_sql_table` perform to load all the desk and convert it right into a Pandas dataframe. The perform requires desk anime, engine objects, and column names. 
  3. Show the highest 5 rows. 
import pandas as pd
import psycopg2
from sqlalchemy import create_engine
 
# set up a reference to the database
engine = create_engine("sqlite:///kdnuggets.db")
 
# learn the sqlite desk
table_df = pd.read_sql_table(
    "countries_poluation",
    con=engine,
    columns=['city', 'country', '2017', '2018', '2019', '2020', '2021', '2022',
       '2023']
)
 
table_df.head()

 

The SQL desk has been efficiently loaded as a dataframe. This implies you can now use it to carry out knowledge evaluation and visualization utilizing standard Python packages reminiscent of Seaborn, Matplotlib, Scipy, Numpy, and extra.

 

countries air pollution pandas dataframe

 

3. Operating the SQL Question utilizing Pandas

 

As an alternative of limiting ourselves to 1 desk, we are able to entry all the database through the use of the `pd.read_sql` perform. Simply write a easy SQL question and supply it with the engine object.

The SQL question will show two columns from the “countries_population” desk, type it by the “2023” column, and show the highest 5 outcomes.

# learn desk knowledge utilizing sql question
sql_df = pd.read_sql(
    "SELECT city,[2023] FROM countries_poluation ORDER BY [2023] DESC LIMIT 5",
    con=engine
)
 
print(sql_df)

 

We received to the highest 5 cities on the earth with the worst air high quality. 

         metropolis  2023
0       Lahore  97.4
1        Hotan  95.0
2      Bhiwadi  93.3
3  Delhi (NCT)  92.7
4     Peshawar  91.9

 

4. Utilizing the SQL Question End result with Pandas

 

We will additionally use the outcomes from SQL question and carry out additional evaluation. For instance, calculate the typical of the highest 5 cities utilizing Pandas. 

average_air = sql_df['2023'].imply()
print(f"The average of top 5 cities: {average_air:.2f}")

 

Output:

The common of high 5 cities: 94.06

 

Or, create a bar plot by specifying the x and y arguments and the kind of plot. 

sql_df.plot(x="city",y="2023",type = "barh");

 

data visualization using pandas

 

Conclusion

 

The probabilities of utilizing SQLAlchemy with Pandas are limitless. You’ll be able to carry out easy knowledge evaluation utilizing the SQL question, however to visualise the outcomes and even practice the machine studying mannequin, you need to convert it right into a Pandas dataframe. 

On this tutorial, now we have realized the right way to load a SQL database into Python, carry out knowledge evaluation, and create visualizations. If you happen to loved this information, additionally, you will respect ‘A Information to Working with SQLite Databases in Python‘, which offers an in-depth exploration of utilizing Python’s built-in sqlite3 module.
 
 

Abid Ali Awan (@1abidaliawan) is a licensed knowledge scientist skilled who loves constructing machine studying fashions. At the moment, he’s specializing in content material creation and writing technical blogs on machine studying and knowledge science applied sciences. Abid holds a Grasp’s diploma in know-how administration and a bachelor’s diploma in telecommunication engineering. His imaginative and prescient is to construct an AI product utilizing a graph neural community for college kids fighting psychological sickness.

Related articles

How They’re Altering Distant Work

Distant work has change into part of on a regular basis life for many people. Whether or not...

David Maher, CTO of Intertrust – Interview Sequence

David Maher serves as Intertrust’s Govt Vice President and Chief Know-how Officer. With over 30 years of expertise in...

Is It Google’s Largest Rival But?

For years, Google has been the go-to place for locating something on the web. Whether or not you’re...

Meshy AI Overview: How I Generated 3D Fashions in One Minute

Have you ever ever spent hours (and even days) painstakingly creating 3D fashions, solely to really feel just...