Jobs & Careers
How to Learn AI for Data Analytics in 2025
Image by Editor | ChatGPT
Data analytics has changed. It is no longer sufficient to know tools like Python, SQL, and Excel to be a data analyst.
As a data professional at a tech company, I am experiencing firsthand the integration of AI into every employee’s workflow. There is an ocean of AI tools that can now access and analyze your entire database and help you build data analytics projects, machine learning models, and web applications in minutes.
If you are an aspiring data professional and aren’t using these AI tools, you are losing out. And soon, you will be surpassed by other data analysts; people who are using AI to optimize their workflows.
In this article, I will walk you through AI tools that will help you stay ahead of the competition and 10X your data analytics workflows.
With these tools, you can:
- Build and deploy creative portfolio projects to get hired as a data analyst
- Use plain English to create end-to-end data analytics applications
- Speed up your data workflows and become a more efficient data analyst
Additionally, this article will be a step-by-step guide on how to use AI tools to build data analytics applications. We will focus on two AI tools in particular – Cursor and Pandas AI.
For a video version of this article, watch this:
AI Tool 1: Cursor
Cursor is an AI code editor that has access to your entire codebase. You just have to type a prompt into Cursor’s chat interface, and it will access all the files in your directory and edit code for you.
If you are a beginner and can’t write a single line of code, you can even start with an empty code folder and ask Cursor to build something for you. The AI tool will then follow your instructions and create code files according to your requirements.
Here is a guide on how you can use Cursor to build an end-to-end data analytics project without writing a single line of code.
Step 1: Cursor Installation and Setup
Let’s see how we can use Cursor AI for data analytics.
To install Cursor, just go to www.cursor.com, download the version that is compatible with your OS, follow the installation instructions, and you will be set up in seconds.
Here’s what the Cursor interface looks like:
Cursor AI Interface
To follow along to this tutorial, download the train.csv
file from the Sentiment Analysis Dataset on Kaggle.
Then create a folder named “Sentiment Analysis Project” and move the downloaded train.csv file into it.
Finally, create an empty file named app.py
. Your project folder should now look like this:
Sentiment Analysis Project Folder
This will be our working directory.
Now, open this folder in Cursor by navigating to File -> Open Folder.
The right side of the screen has a chat interface where you can type prompts into Cursor. Notice that there are a few selections here. Let’s select “Agent” in the drop-down.
This tells Cursor to explore your codebase and act as an AI assistant that will refactor and debug your code.
Additionally, you can choose which language model you’d like to use with Cursor (GPT-4o, Gemini-2.5-Pro, etc). I suggest using Claude-4-Sonnet, a model that is well-known for its advanced coding capabilities.
Step 2: Prompting Cursor to Build an Application
Let’s now type this prompt into Cursor, asking it to build an end-to-end sentiment analysis model using the training dataset in our codebase:
Create a sentiment analysis web app that:
1. Uses a pre-trained DistilBERT model to analyze the sentiment of text (positive, negative, or neutral)
2. Has a simple web interface where users can enter text and see results
3. Shows the sentiment result with appropriate colors (green for positive, red for negative)
4. Runs immediately without needing any training
Please connect all the files properly so that when I enter text and click analyze, it shows me the sentiment result right away.
After you enter this prompt into Cursor, it will automatically generate code files to build the sentiment analysis application.
Step 3: Accepting Changes and Running Commands
As Cursor creates new files and generates code, you need to click on “Accept” to confirm the changes made by the AI agent.
After Cursor writes out all the code, it might prompt you to run some commands on the terminal. Executing these commands will allow you to install the required dependencies and run the web application.
Just click on “Run,” which allows Cursor to run these commands for us:
Run Command Cursor
Once Cursor has built the application, it will tell you to copy and paste this link into your browser:
Cursor App Link
Doing so will lead you to the sentiment analysis web application, which looks like this:
Sentiment Analysis App with Cursor
This is a fully-fledged web application that employers can interact with. You can paste any sentence into this app and it will predict the sentiment, returning a result to you.
I find tools like Cursor to be incredibly powerful if you are a beginner in the field and want to productionize your projects.
Most data professionals don’t know front-end programming languages like HTML and CSS, due to which we’re unable to showcase our projects in an interactive application.
Our code often sits in Kaggle notebooks, which doesn’t give us a competitive advantage over hundreds of other applicants doing the exact same thing.
A tool like Cursor, however, can set you apart from the competition. It can help you turn your ideas into reality by coding out exactly what you tell it to.
AI Tool 2: Pandas AI
Pandas AI lets you manipulate and analyze Pandas data frames without writing any code.
You just have to type prompts in plain English, which reduces the complexity that comes with performing data preprocessing and EDA.
If you don’t already know, Pandas is a Python library that you can use to analyze and manipulate data.
You read data into something known as a Pandas data frame, which then allows you to perform operations on your data.
Let’s go through an example of how you can perform data preprocessing, manipulation, and analysis with Pandas AI.
For this demo, I will be using the Titanic Survival Prediction dataset on Kaggle (download the train.csv
file).
For this analysis, I suggest using a Python notebook environment, like a Jupyter Notebook, a Kaggle Notebook, or Google Colab. The complete code for this analysis can be found in this Kaggle Notebook.
Step 1: Pandas AI Installation and Setup
Once you have your notebook environment ready, type the command below to install Pandas AI:
!pip install pandasai
Next, load the Titanic dataframe with the following lines of code:
import pandas as pd
train_data = pd.read_csv('/kaggle/input/titanic/train.csv')
Now let’s import the following libraries:
import os
from pandasai import SmartDataframe
from pandasai.llm.openai import OpenAI
Next, we must create a Pandas AI object to analyze the Titanic train dataset.
Here’s what this means:
Pandas AI is a library that connects your Pandas data frame to a Large Language Model. You can use Pandas AI to connect to GPT-4o, Claude-3.5, and other LLMs.
By default, Pandas AI uses a language model called Bamboo LLM. To connect Pandas AI to the language model, you can visit this website to get an API key.
Then, enter the API key into this block of code to create a Pandas AI object:
# Set the PandasAI API key
# By default, unless you choose a different LLM, it will use BambooLLM.
# You can get your free API key by signing up at https://app.pandabi.ai
os.environ['PANDASAI_API_KEY'] = 'your-pandasai-api-key' # Replace with your actual key
# Create SmartDataframe with default LLM (Bamboo)
smart_df = SmartDataframe(train_data)
Personally, I faced some issues in retrieving the Bamboo LLM API key. Due to this, I decided to get an API key from OpenAI instead. Then, I used the GPT-4o model for this analysis.
One caveat to this approach is that OpenAI’s API keys aren’t free. You must purchase OpenAI’s API tokens to use these models.
To do this, navigate to Open AI’s website and purchase tokens from the billings page. Then you can go to the “API keys” page and create your API key.
Now that you have the OpenAI API key, you need to enter it into this block of code to connect the GPT-4o model to Pandas AI:
# Set your OpenAI API key
os.environ["OPENAI_API_KEY"] = "YOUR_API_KEY"
# Initialize OpenAI LLM
llm = OpenAI(api_token=os.environ["OPENAI_API_KEY"], model="gpt-4o")
config = {
"llm": llm,
"enable_cache": False,
"verbose": False,
"save_logs": True
}
# Create SmartDataframe with explicit configuration
smart_df = SmartDataframe(train_data, config=config)
We can now use this Pandas AI object to analyze the Titanic dataset.
Step 2: EDA and Data Preprocessing with Pandas AI
First, let’s start with a simple prompt asking Pandas AI to describe this dataset:
smart_df.chat("Can you describe this dataset and provide a summary, format the output as a table.")
You will see a result that looks like this, with a basic statistical summary of the dataset:
Titanic Dataset Description
Typically we’d write some code to get a summary like this. With Pandas AI, however, we just need to write a prompt.
This will save you a ton of time if you’re a beginner who wants to analyze some data but don’t know how to write Python code.
Next, let’s perform some exploratory data analysis with Pandas AI:
I’m asking it to give me the relationship between the “Survived” variable in the Titanic dataset, along with some other variables in the dataset:
smart_df.chat("Are there correlations between Survived and the following variables: Age, Sex, Ticket Fare. Format this output as a table.")
The above prompt should provide you with a correlation coefficient between “Survived” and the other variables in the dataset.
Next, let’s ask Pandas AI to help us visualize the relationship between these variables:
1. Survived and Age
smart_df.chat("Can you visualize the relationship between the Survived and Age columns?")
The above prompt should give you a histogram that looks like this:
Titanic Dataset Age Distribution
This visual tells us that younger passengers were more likely to survive the crash.
2. Survived and Gender
smart_df.chat("Can you visualize the relationship between the Survived and Sex")
You should get a bar chart showcasing the relationship between “Survived” and “Gender.”
3. Survived and Fare
smart_df.chat("Can you visualize the relationship between the Survived and Fare")
The above prompt rendered a box plot, telling me that passengers who paid higher fare prices were more likely to survive the Titanic crash.
Note that LLMs are non-deterministic, which means that the output you’ll get might differ from mine. However, you will still get a response that will help you better understand the dataset.
Next, we can perform some data preprocessing with prompts like these:
Prompt Example 1
smart_df.chat("Analyze the quality of this dataset. Identify missing values, outliers, and potential data issues that would need to be addressed before we build a model to predict survival.")
Prompt Example 2
smart_df.chat("Let's drop the cabin column from the dataframe as it has too many missing values.")
Prompt Example 3
smart_df.chat("Let's impute the Age column with the median value.")
If you’d like to go through all the preprocessing steps I used to clean this dataset with Pandas AI, you can find the complete prompts and code in my Kaggle notebook.
In less than 5 minutes, I was able to preprocess this dataset by handling missing values, encoding categorical variables, and creating new features. This was done without writing much Python code, which is especially helpful if you are new to programming.
How to Learn AI for Data Analytics: Next Steps
In my opinion, the main selling point of tools like Cursor and Pandas AI is that they allow you to analyze data and make code edits within your programming interface.
This is far better than having to copy and paste code out of your programming IDE into an interface like ChatGPT.
Additionally, as your codebase grows (i.e. if you have thousands of lines of code and over 10 datasets), it is incredibly useful to have an integrated AI tool that has all the context and can understand the connection between these code files.
If you’re looking to learn AI for data analytics, here are some more tools that I’ve found helpful:
- GitHub Copilot: This tool is similar to Cursor. You can use it within your programming IDE to generate code suggestions, and it even has a chat interface you can interact with.
- Microsoft Copilot in Excel: This AI tool helps you automatically analyze data in your spreadsheets.
- Python in Excel: This is an extension that allows you to run Python code within Excel. While this isn’t an AI tool, I’ve found it incredibly useful as it allows you to centralize your data analysis without having to switch between different applications.
Natassha Selvaraj is a self-taught data scientist with a passion for writing. Natassha writes on everything data science-related, a true master of all data topics. You can connect with her on LinkedIn or check out her YouTube channel.
Jobs & Careers
HCLSoftware Launches Domino 14.5 With Focus on Data Privacy and Sovereign AI
HCLSoftware, a global enterprise software leader, launched HCL Domino 14.5 on July 7 as a major upgrade, specifically targeting governments and organisations operating in regulated sectors that are concerned about data privacy and digital independence.
A key feature of the new release is Domino IQ, a sovereign AI extension built into the Domino platform. This new tool gives organisations full control over their AI models and data, helping them comply with regulations such as the European AI Act.
It also removes dependence on foreign cloud services, making it easier for public sector bodies and banks to protect sensitive information.
“The importance of data sovereignty and avoiding unnecessary foreign government influence extends beyond SaaS solutions and AI. Specifically for collaboration – the sensitive data within email, chat, video recordings and documents. With the launch of Domino+ 14.5, HCLSoftware is helping over 200+ government agencies safeguard their sensitive data,” said Richard Jefts, executive vice president and general manager at HCLSoftware
The updated Domino+ collaboration suite now includes enhanced features for secure messaging, meetings, and file sharing. These tools are ready to deploy and meet the needs of organisations that handle highly confidential data.
The platform is supported by IONOS, a leading European cloud provider. Achim Weiss, CEO of IONOS, added, “Today, more than ever, true digital sovereignty is the key to Europe’s digital future. That’s why at IONOS we are proud to provide the sovereign cloud infrastructure for HCL’s sovereign collaboration solutions.”
Other key updates in Domino 14.5 include achieving BSI certification for information security, the integration of security event and incident management (SEIM) tools to enhance threat detection and response, and full compliance with the European Accessibility Act, ensuring that all web-based user experiences are inclusive and accessible to everyone.
With the launch of Domino 14.5, HCLSoftware is aiming to be a trusted technology partner for public sector and highly regulated organisations seeking control, security, and compliance in their digital operations.
Jobs & Careers
Mitsubishi Electric Invests in AI-Assisted PLM Systems Startup ‘Things’
Mitsubishi Electric Corporation announced on July 7 that its ME Innovation Fund has invested in Things, a Japan-based startup that develops and provides AI-assisted product lifecycle management (PLM) systems for the manufacturing industry.
This startup specialises in comprehensive document management, covering everything from product planning and development to disposal. According to the company, this marks the 12th investment made by Mitsubishi’s fund to date.
Through this investment, Mitsubishi Electric aims to combine its extensive manufacturing and control expertise with Things’ generative AI technology. The goal is to accelerate the development of digital transformation (DX) solutions that tackle various challenges facing the manufacturing industry.
In recent years, Japan’s manufacturing sector has encountered several challenges, including labour shortages and the ageing of skilled technicians, which hinder the transfer of expertise. In response, DX initiatives, such as the implementation of PLM and other digital systems, have progressed rapidly. However, these initiatives have faced challenges related to development time, cost, usability, and scalability.
Komi Matsubara, an executive officer at Mitsubishi Electric Corporation, stated, “Through our collaboration with Things, we expect to generate new value by integrating our manufacturing expertise with Things’ generative AI technology. We aim to leverage this initiative to enhance the overall competitiveness of the Mitsubishi Electric group.”
Things launched its ‘PRISM’ PLM system in May 2023, utilising generative AI to improve the structure and usage of information in manufacturing. PRISM offers significant cost and scalability advantages, enhancing user interfaces and experiences while effectively implementing proofs of concept across a wide range of companies.
Atsuya Suzuki, CEO of Things, said, “We are pleased to establish a partnership with Mitsubishi Electric through the ME Innovation Fund. By combining our technology with Mitsubishi Electric’s expertise in manufacturing and control, we aim to accelerate the global implementation of pioneering DX solutions for manufacturing.”
Jobs & Careers
AI to Track Facial Expressions to Detect PTSD Symptoms in Children
A research team from the University of South Florida (USF) has developed an AI system that can identify post-traumatic stress disorder (PTSD) in children.
The project addresses a longstanding clinical dilemma: diagnosing PTSD in children who may not have the emotional vocabulary, cognitive development or comfort to articulate their distress. Traditional methods such as subjective interviews and self-reported questionnaires often fall short. This is where AI steps in.
“Even when they weren’t saying much, you could see what they were going through on their faces,” Alison Salloum, professor at the USF School of Social Work, reportedly said. Her observations during trauma interviews laid the foundation for collaboration with Shaun Canavan, an expert in facial analysis at USF’s Bellini College of Artificial Intelligence, Cybersecurity, and Computing.
The study introduces a privacy-first, context-aware classification model that analyses subtle facial muscle movements. However, instead of using raw footage, the system extracts non-identifiable metrics such as eye gaze, mouth curvature, and head position, ensuring ethical boundaries are respected when working with vulnerable populations.
“We don’t use raw video. We completely get rid of subject identification and only keep data about facial movement,” Canavan reportedly emphasised. The AI also accounts for conversational context, whether a child is speaking to a parent or a therapist, which significantly influences emotional expressivity.
Across 18 therapy sessions, with over 100 minutes of footage per child and approximately 185,000 frames each, the AI identified consistent facial expression patterns in children diagnosed with PTSD. Notably, children were more expressive with clinicians than with parents; a finding that aligns with psychological literature suggesting shame or emotional avoidance often inhibits open communication at home.
While still in its early stages, the tool is not being pitched as a replacement for therapists. Instead, it’s designed as a clinical augmentation, a second set of ‘digital’ eyes that can pick up on emotional signals even trained professionals might miss in real time.
“Data like this is incredibly rare for AI systems,” Canavan added. “That’s what makes this so promising. We now have an ethically sound, objective way to support mental health assessments.”
If validated on a larger scale, the system could transform mental health diagnostics for children—especially for pre-verbal or very young patients—by turning non-verbal cues into actionable insights.
-
Funding & Business6 days ago
Kayak and Expedia race to build AI travel agents that turn social posts into itineraries
-
Jobs & Careers6 days ago
Mumbai-based Perplexity Alternative Has 60k+ Users Without Funding
-
Mergers & Acquisitions6 days ago
Donald Trump suggests US government review subsidies to Elon Musk’s companies
-
Funding & Business6 days ago
Rethinking Venture Capital’s Talent Pipeline
-
Jobs & Careers6 days ago
Why Agentic AI Isn’t Pure Hype (And What Skeptics Aren’t Seeing Yet)
-
Funding & Business4 days ago
Sakana AI’s TreeQuest: Deploy multi-model teams that outperform individual LLMs by 30%
-
Funding & Business7 days ago
From chatbots to collaborators: How AI agents are reshaping enterprise work
-
Jobs & Careers6 days ago
Astrophel Aerospace Raises ₹6.84 Crore to Build Reusable Launch Vehicle
-
Funding & Business4 days ago
HOLY SMOKES! A new, 200% faster DeepSeek R1-0528 variant appears from German lab TNG Technology Consulting GmbH
-
Funding & Business6 days ago
Europe’s Most Ambitious Startups Aren’t Becoming Global; They’re Starting That Way