Structured 45-Day Roadmap + Resume-Based Q&A + Study Tracker
Inspired by Coding Interview University, tailored to my resume (see)
📄 Resume: View Image
🔗 LinkedIn | Github | HackerRank | Gmail
| Week | Topics |
|---|---|
| 1 | Data Analysis Basics, Excel, Data Cleaning |
| 2 | SQL (Medium/Advanced: Joins, Window, Subqueries, CTEs) |
| 3 | Python (Pandas, Numpy, Data Processing & Analysis) |
| 4 | Power BI/DAX, Data Visualization, Storytelling, KPI Design |
| 5 | Statistics, Case Studies, Resume-based Q&A, Mock Interviews |
-
Q: Tell me about yourself as a data analyst.
A: I am a data analyst skilled in Power BI, SQL, and Python, with hands-on experience in transforming raw data into actionable business insights. My focus has been on developing data solutions for customer segmentation, sales forecasting, and employee attrition analysis. I am proficient in data cleaning, KPI tracking, and building interactive dashboards that support informed decision-making. -
Q: Describe a time you turned raw data into actionable insights.
A: In my internship at CloudyML, I worked with HR attrition and call center datasets. By cleaning and transforming the data, and building Power BI dashboards, I identified key drivers of employee attrition, such as department and role. My analysis led to data-driven recommendations that helped HR teams target retention strategies more effectively. -
Q: How do you approach KPI tracking and dashboard creation?
A: I start by understanding business objectives and identifying relevant KPIs. I use Power BI to build interactive dashboards, leveraging DAX for dynamic calculations and filters. My approach emphasizes clarity, user-friendly visuals, and actionable storytelling, ensuring stakeholders can monitor performance and trends easily.
-
Q: Why did you choose Production Engineering and how does it help in Data Analytics?
A: Production Engineering provided me with a strong foundation in process optimization, problem-solving, and data-driven decision-making. These skills are directly applicable to analytics, where understanding processes and using data to improve them is crucial. -
Q: Describe a project/coursework where you used analytics/statistics.
A: During my coursework, I frequently used statistical methods for process improvement case studies. For example, I used regression analysis and hypothesis testing to analyze production data and recommend process changes that increased efficiency.
-
Q: What is your process for Data Cleaning and Transformation?
A: I start by identifying and handling missing values, outliers, and inconsistencies using Python (Pandas) or Power Query in Power BI. I standardize and normalize data, ensuring it’s well-structured for analysis. I document every cleaning step for reproducibility. -
Q: How do you use Power BI for dashboards and KPI analysis?
A: In Power BI, I design dashboards that visualize key metrics using DAX for calculated fields. I use slicers and filters to enable dynamic data exploration and ensure the dashboards provide clear, actionable insights for different stakeholders. -
Q: Explain how you’ve used DAX or Excel in real projects.
A: I used DAX to create measures for dynamic filtering and custom KPIs in Power BI dashboards. In Excel, I used pivot tables, VLOOKUP, and conditional formatting to explore and visualize data during my initial analysis phases. -
Q: Describe your experience with SQL joins and aggregations.
A: I regularly use SQL joins to combine data from multiple tables, such as inner joins for customer orders and left joins for including all records in a primary table. Aggregations like GROUP BY and window functions help me summarize and analyze trends effectively. -
Q: How comfortable are you with Python for data analysis?
A: I am very comfortable using Python for data analysis, especially with Pandas and Numpy for data manipulation, cleaning, and exploratory analysis. I also use Matplotlib and Seaborn for visualizations. -
Q: Difference between Power Query and DAX?
A: Power Query is used for data extraction, transformation, and loading (ETL), while DAX is used within Power BI for creating calculated columns, measures, and custom aggregations once data is loaded.
-
Q: Describe a real-world data analytics project you completed as an intern.
A: As a Data Analyst Intern at CloudyML, I built Power BI dashboards for HR attrition and call center performance. I handled all stages from data cleaning to visualization, uncovering actionable insights for management. -
Q: How did you approach employee attrition or call center data?
A: I started by cleaning and transforming the data, followed by exploratory analysis to identify trends. I then designed dashboards to highlight KPIs such as attrition rates by department and agent performance metrics. -
Q: What challenges did you face in transforming or visualizing data?
A: One challenge was handling incomplete and inconsistent data. I overcame this by using robust cleaning techniques and validating data sources before building visualizations. -
Q: How did you use Power BI and DAX for insights?
A: I used Power BI for interactive dashboards and DAX for advanced calculations, like dynamic attrition rates and agent performance scores, enabling stakeholders to explore insights through customized filters.
-
Q: What KPIs did you choose and why?
A: I focused on attrition rate by department and role, average tenure, and high-risk employee segments. These KPIs were directly linked to business objectives for improving retention. -
Q: How did you use DAX for custom metrics?
A: I created DAX measures for dynamic attrition calculations and custom filtering by department and role, enhancing the dashboard’s interactivity and value. -
Q: How did you enable quick trend analysis in Power BI?
A: By using slicers and interactive visuals, stakeholders could instantly filter and view trends by different employee segments and periods.
-
Q: How did you segment customers and forecast revenue?
A: I used sales data to segment customers by purchase behavior and DAX for predictive revenue modeling, using historical sales for trend projections. -
Q: Explain the use of DAX calculations and predictive modeling in your report.
A: I implemented DAX measures to calculate customer lifetime value and forecast revenue, enabling scenario analysis for business planning. -
Q: How did you design visuals to explore business KPIs?
A: I used clustered bar charts, line graphs, and interactive slicers to present KPIs, making it easy for users to analyze trends and segment data.
-
Q: What agent-wise KPIs did you track?
A: I tracked average handling time, resolution rate, and customer satisfaction scores for individual agents. -
Q: How did you use slicers/dynamic filters for analysis?
A: Slicers allowed users to filter results by time period, department, and agent, facilitating targeted performance analysis. -
Q: Describe your process for building and sharing dashboards.
A: After building the dashboards in Power BI, I shared them via the Power BI service and ensured stakeholders could interact with and export the reports as needed.
-
Q: How does your sports experience help you in teamwork and communication?
A: Being part of the NIT Agartala handball team taught me discipline, leadership, and teamwork—skills that translate directly into effective collaboration and communication in professional settings. -
Q: Share an example where you demonstrated leadership or discipline.
A: As a team captain, I organized training sessions and motivated the team during tournaments, demonstrating leadership and resilience under pressure.
-
Q: Write a query to find top N employees by attrition risk.
A:SELECT employee_id, attrition_score FROM hr_data ORDER BY attrition_score DESC LIMIT N;
-
Q: How do you use window functions for running totals/averages?
A:
I use window functions likeSUM() OVER (ORDER BY date)to calculate running totals andAVG() OVER (PARTITION BY department)for moving averages, which helps uncover trends across time or categories. -
Q: Explain the difference between INNER, LEFT, and CROSS JOIN.
A:- INNER JOIN returns only matching rows between tables.
- LEFT JOIN returns all rows from the left table and matches from the right, filling in NULLs when there’s no match.
- CROSS JOIN returns all possible row combinations (Cartesian product).
-
Q: How do you optimize slow SQL queries?
A:
By adding appropriate indexes, avoiding SELECT *, analyzing query plans, and minimizing subqueries or redundant computations. -
Q: Solve a case: “Find department with the highest average handling time.”
A:SELECT department, AVG(handling_time) as avg_handling FROM call_data GROUP BY department ORDER BY avg_handling DESC LIMIT 1;
-
Q: How to handle missing data in large datasets?
A:
I use Pandas methods likefillna()to impute missing values ordropna()to remove incomplete rows, choosing the method based on the impact on analysis. -
Q: Write code to group by and aggregate metrics.
A:df.groupby('department')['attrition_score'].mean()
-
Q: Explain vectorized operations vs. loops.
A:
Vectorized operations in Pandas/Numpy execute computations on entire arrays at once, making them much faster and more efficient than explicit Python loops. -
Q: How do you merge multiple DataFrames with different keys?
A:
Usingpd.merge()with theleft_onandright_onparameters to specify the joining columns.
-
Q: Write a DAX formula for dynamic filtering.
A:CALCULATE([Total Sales], FILTER(Sales, Sales[Region] = "West")) -
Q: How to build a calculated column vs. measure?
A:
Calculated columns are computed row-by-row at data refresh, while measures are calculated dynamically based on filters or aggregations in reports. -
Q: How to implement row-level security?
A:
Define security roles in Power BI Desktop with DAX filters (e.g.,[Department] = USERPRINCIPALNAME()) and assign users to those roles when publishing.
-
Q: When would you use a t-test vs. chi-square?
A:
Use a t-test to compare means of two groups (continuous data), and chi-square to test associations between categorical variables. -
Q: Explain p-value and confidence interval in layman terms.
A:
A p-value tells us the probability that our results happened by chance. A confidence interval gives a range where we believe the true value lies, with a certain level of certainty (like 95%). -
Q: What is regression analysis and when is it used?
A:
Regression analysis models relationships between variables, often to predict outcomes (e.g., predicting sales based on marketing spend).
-
Q: How would you improve employee retention using analytics?
A:
By analyzing attrition data to identify high-risk groups, understanding key drivers through surveys and KPIs, and recommending targeted retention strategies. -
Q: How would you present insights to non-technical stakeholders?
A:
I simplify findings using clear visuals and focus on actionable recommendations, avoiding technical jargon to ensure clarity. -
Q: Walk me through your process in tackling a new analytics project.
A:
I start by understanding business goals, gathering and cleaning data, performing exploratory analysis, building models or dashboards, and communicating insights with clear visuals and recommendations.
- Sales Dashboard (Power BI/Tableau): Template
- Churn Prediction (Python, SQL): Template
- E-commerce Analytics (SQL): Template
- Covid-19 Data Analysis (Dashboards): Template
- SQL: Window functions, CTEs, PIVOT/UNPIVOT, query optimization
- Stats: Hypothesis testing, ANOVA, regression, time series
- Data Viz: Interactive dashboards, geo-visualization, storytelling
- StatQuest (Statistics)
- Alex The Analyst (SQL, Career)
- Ken Jee (Project/Case Studies)
- Data Skeptic (Podcast)
- Analytics Power Hour
| Date | Topic/Project/Mock | Status | Notes |
|---|---|---|---|
- Practice resume-based Q&A and STAR stories.
- Build & document your projects (GitHub, dashboards).
- Share your journey and projects online.
- Schedule mock interviews—track improvements here.
- Update this README as you progress!
Image reference:
– My resume for structured Q&A and preparation.