Picture by Editor | ChatGPT
Knowledge analytics has modified. It’s not ample to know instruments like Python, SQL, and Excel to be an information analyst.
As an information skilled at a tech firm, I’m experiencing firsthand the combination of AI into each worker’s workflow. There’s an ocean of AI instruments that may now entry and analyze your total database and enable you to construct knowledge analytics tasks, machine studying fashions, and internet purposes in minutes.
In case you are an aspiring knowledge skilled and aren’t utilizing these AI instruments, you might be shedding out. And shortly, you’ll be surpassed by different knowledge analysts; people who find themselves utilizing AI to optimize their workflows.
On this article, I’ll stroll you thru AI instruments that can enable you to keep forward of the competitors and 10X your knowledge analytics workflows.
With these instruments, you possibly can:
- Construct and deploy artistic portfolio tasks to get employed as an information analyst
- Use plain English to create end-to-end knowledge analytics purposes
- Velocity up your knowledge workflows and develop into a extra environment friendly knowledge analyst
Moreover, this text shall be a step-by-step information on tips on how to use AI instruments to construct knowledge analytics purposes. We’ll concentrate on two AI instruments specifically – Cursor and Pandas AI.
For a video model of this text, watch this:
AI Instrument 1: Cursor
Cursor is an AI code editor that has entry to your total codebase. You simply must kind a immediate into Cursor’s chat interface, and it’ll entry all of the information in your listing and edit code for you.
In case you are a newbie and may’t write a single line of code, you possibly can even begin with an empty code folder and ask Cursor to construct one thing for you. The AI software will then comply with your directions and create code information in response to your necessities.
Here’s a information on how you should utilize Cursor to construct an end-to-end knowledge analytics challenge with out writing a single line of code.
Step 1: Cursor Set up and Setup
Let’s see how we will use Cursor AI for knowledge analytics.
To put in Cursor, simply go to www.cursor.com, obtain the model that’s appropriate together with your OS, comply with the set up directions, and you’ll be arrange in seconds.
Right here’s what the Cursor interface seems like:

Cursor AI Interface
To comply with alongside to this tutorial, obtain the practice.csv
file from the Sentiment Evaluation Dataset on Kaggle.
Then create a folder named “Sentiment Evaluation Mission” and transfer the downloaded practice.csv file into it.
Lastly, create an empty file named app.py
. Your challenge folder ought to now appear to be this:

Sentiment Evaluation Mission Folder
This shall be our working listing.
Now, open this folder in Cursor by navigating to File -> Open Folder.
The best facet of the display has a chat interface the place you possibly can kind prompts into Cursor. Discover that there are a couple of alternatives right here. Let’s choose “Agent” within the drop-down.
This tells Cursor to discover your codebase and act as an AI assistant that can refactor and debug your code.
Moreover, you possibly can select which language mannequin you’d like to make use of with Cursor (GPT-4o, Gemini-2.5-Professional, and so forth). I recommend utilizing Claude-4-Sonnet, a mannequin that’s well-known for its superior coding capabilities.
Step 2: Prompting Cursor to Construct an Software
Let’s now kind this immediate into Cursor, asking it to construct an end-to-end sentiment evaluation mannequin utilizing the coaching dataset in our codebase:
Create a sentiment evaluation internet app that:
1. Makes use of a pre-trained DistilBERT mannequin to investigate the sentiment of textual content (constructive, destructive, or impartial)
2. Has a easy internet interface the place customers can enter textual content and see outcomes
3. Exhibits the sentiment outcome with applicable colours (inexperienced for constructive, pink for destructive)
4. Runs instantly with no need any coaching
Please join all of the information correctly in order that after I enter textual content and click on analyze, it exhibits me the sentiment outcome straight away.
After you enter this immediate into Cursor, it should robotically generate code information to construct the sentiment evaluation software.
Step 3: Accepting Modifications and Operating Instructions
As Cursor creates new information and generates code, it’s good to click on on “Settle for” to substantiate the modifications made by the AI agent.
After Cursor writes out all of the code, it’d immediate you to run some instructions on the terminal. Executing these instructions will will let you set up the required dependencies and run the net software.
Simply click on on “Run,” which permits Cursor to run these instructions for us:

Run Command Cursor
As soon as Cursor has constructed the applying, it should let you know to repeat and paste this hyperlink into your browser:

Cursor App Hyperlink
Doing so will lead you to the sentiment evaluation internet software, which seems like this:

Sentiment Evaluation App with Cursor
It is a fully-fledged internet software that employers can work together with. You’ll be able to paste any sentence into this app and it’ll predict the sentiment, returning a outcome to you.
I discover instruments like Cursor to be extremely highly effective if you’re a newbie within the discipline and wish to productionize your tasks.
Most knowledge professionals don’t know front-end programming languages like HTML and CSS, because of which we’re unable to showcase our tasks in an interactive software.
Our code usually sits in Kaggle notebooks, which doesn’t give us a aggressive benefit over lots of of different candidates doing the very same factor.
A software like Cursor, nonetheless, can set you aside from the competitors. It may possibly enable you to flip your concepts into actuality by coding out precisely what you inform it to.
AI Instrument 2: Pandas AI
Pandas AI helps you to manipulate and analyze Pandas knowledge frames with out writing any code.
You simply must kind prompts in plain English, which reduces the complexity that comes with performing knowledge preprocessing and EDA.
For those who don’t already know, Pandas is a Python library that you should utilize to investigate and manipulate knowledge.
You learn knowledge into one thing often known as a Pandas knowledge body, which then means that you can carry out operations in your knowledge.
Let’s undergo an instance of how one can carry out knowledge preprocessing, manipulation, and evaluation with Pandas AI.
For this demo, I shall be utilizing the Titanic Survival Prediction dataset on Kaggle (obtain the practice.csv
file).
For this evaluation, I recommend utilizing a Python pocket book surroundings, like a Jupyter Pocket book, a Kaggle Pocket book, or Google Colab. The whole code for this evaluation could be present in this Kaggle Pocket book.
Step 1: Pandas AI Set up and Setup
After getting your pocket book surroundings prepared, kind the command under to put in Pandas AI:
!pip set up pandasai
Subsequent, load the Titanic dataframe with the next strains of code:
import pandas as pd
train_data = pd.read_csv('/kaggle/enter/titanic/practice.csv')
Now let’s import the next libraries:
import os
from pandasai import SmartDataframe
from pandasai.llm.openai import OpenAI
Subsequent, we should create a Pandas AI object to investigate the Titanic practice dataset.
Right here’s what this implies:
Pandas AI is a library that connects your Pandas knowledge body to a Giant Language Mannequin. You should use Pandas AI to hook up with GPT-4o, Claude-3.5, and different LLMs.
By default, Pandas AI makes use of a language mannequin known as Bamboo LLM. To attach Pandas AI to the language mannequin, you possibly can go to this web site to get an API key.
Then, enter the API key into this block of code to create a Pandas AI object:
# Set the PandasAI API key
# By default, until you select a special LLM, it should use BambooLLM.
# You will get your free API key by signing up at https://app.pandabi.ai
os.environ['PANDASAI_API_KEY'] = 'your-pandasai-api-key' # Substitute together with your precise key
# Create SmartDataframe with default LLM (Bamboo)
smart_df = SmartDataframe(train_data)
Personally, I confronted some points in retrieving the Bamboo LLM API key. Attributable to this, I made a decision to get an API key from OpenAI as an alternative. Then, I used the GPT-4o mannequin for this evaluation.
One caveat to this method is that OpenAI’s API keys aren’t free. You have to buy OpenAI’s API tokens to make use of these fashions.
To do that, navigate to Open AI’s web site and buy tokens from the billings web page. Then you possibly can go to the “API keys” web page and create your API key.
Now that you’ve got the OpenAI API key, it’s good to enter it into this block of code to attach the GPT-4o mannequin to Pandas AI:
# Set your OpenAI API key
os.environ["OPENAI_API_KEY"] = "YOUR_API_KEY"
# Initialize OpenAI LLM
llm = OpenAI(api_token=os.environ["OPENAI_API_KEY"], mannequin="gpt-4o")
config = {
"llm": llm,
"enable_cache": False,
"verbose": False,
"save_logs": True
}
# Create SmartDataframe with express configuration
smart_df = SmartDataframe(train_data, config=config)
We will now use this Pandas AI object to investigate the Titanic dataset.
Step 2: EDA and Knowledge Preprocessing with Pandas AI
First, let’s begin with a easy immediate asking Pandas AI to explain this dataset:
smart_df.chat("Are you able to describe this dataset and supply a abstract, format the output as a desk.")
You will note a outcome that appears like this, with a primary statistical abstract of the dataset:

Titanic Dataset Description
Usually we’d write some code to get a abstract like this. With Pandas AI, nonetheless, we simply want to write down a immediate.
It will prevent a ton of time if you happen to’re a newbie who needs to investigate some knowledge however don’t know tips on how to write Python code.
Subsequent, let’s carry out some exploratory knowledge evaluation with Pandas AI:
I’m asking it to present me the connection between the “Survived” variable within the Titanic dataset, together with another variables within the dataset:
smart_df.chat("Are there correlations between Survived and the next variables: Age, Intercourse, Ticket Fare. Format this output as a desk.")
The above immediate ought to offer you a correlation coefficient between “Survived” and the opposite variables within the dataset.
Subsequent, let’s ask Pandas AI to assist us visualize the connection between these variables:
1. Survived and Age
smart_df.chat("Are you able to visualize the connection between the Survived and Age columns?")
The above immediate ought to provide you with a histogram that appears like this:

Titanic Dataset Age Distribution
This visible tells us that youthful passengers have been extra prone to survive the crash.
2. Survived and Gender
smart_df.chat("Are you able to visualize the connection between the Survived and Intercourse")
You must get a bar chart showcasing the connection between “Survived” and “Gender.”
3. Survived and Fare
smart_df.chat("Are you able to visualize the connection between the Survived and Fare")
The above immediate rendered a field plot, telling me that passengers who paid increased fare costs have been extra prone to survive the Titanic crash.
Be aware that LLMs are non-deterministic, which implies that the output you’ll get would possibly differ from mine. Nevertheless, you’ll nonetheless get a response that can enable you to higher perceive the dataset.
Subsequent, we will carry out some knowledge preprocessing with prompts like these:
Immediate Instance 1
smart_df.chat("Analyze the standard of this dataset. Determine lacking values, outliers, and potential knowledge points that may should be addressed earlier than we construct a mannequin to foretell survival.")
Immediate Instance 2
smart_df.chat("Let's drop the cabin column from the dataframe because it has too many lacking values.")
Immediate Instance 3
smart_df.chat("Let's impute the Age column with the median worth.")
For those who’d prefer to undergo all of the preprocessing steps I used to wash this dataset with Pandas AI, you could find the whole prompts and code in my Kaggle pocket book.
In lower than 5 minutes, I used to be capable of preprocess this dataset by dealing with lacking values, encoding categorical variables, and creating new options. This was completed with out writing a lot Python code, which is particularly useful if you’re new to programming.
How one can Be taught AI for Knowledge Analytics: Subsequent Steps
For my part, the primary promoting level of instruments like Cursor and Pandas AI is that they will let you analyze knowledge and make code edits inside your programming interface.
This is much better than having to repeat and paste code out of your programming IDE into an interface like ChatGPT.
Moreover, as your codebase grows (i.e. you probably have hundreds of strains of code and over 10 datasets), it’s extremely helpful to have an built-in AI software that has all of the context and may perceive the connection between these code information.
For those who’re trying to be taught AI for knowledge analytics, listed here are some extra instruments that I’ve discovered useful:
- GitHub Copilot: This software is much like Cursor. You should use it inside your programming IDE to generate code options, and it even has a chat interface you possibly can work together with.
- Microsoft Copilot in Excel: This AI software helps you robotically analyze knowledge in your spreadsheets.
- Python in Excel: That is an extension that means that you can run Python code inside Excel. Whereas this isn’t an AI software, I’ve discovered it extremely helpful because it means that you can centralize your knowledge evaluation with out having to modify between totally different purposes.
Natassha Selvaraj is a self-taught knowledge scientist with a ardour for writing. Natassha writes on all the things knowledge science-related, a real grasp of all knowledge subjects. You’ll be able to join together with her on LinkedIn or try her YouTube channel.