i
Wipro
Filter interviews by
Design an ETL pipeline to handle irregularly timed JSON file uploads for data transformation and schema matching.
Use a cloud storage service (e.g., AWS S3) to store incoming JSON files.
Implement a file watcher or event-driven architecture (e.g., AWS Lambda) to trigger processing when new files arrive.
Utilize a data processing framework (e.g., Apache Spark or Apache Beam) to read and transform the JSON data.
Define ...
Use SQL window functions to identify the highest sale amount for each store per day.
Use the ROW_NUMBER() function to rank sales within each day and store.
Partition the data by store and date to isolate daily sales.
Order the sales in descending order to get the highest sale at the top.
Example SQL query: SELECT store_id, sale_date, sale_amount, ROW_NUMBER() OVER (PARTITION BY store_id, sale_date ORDER BY sale_amount...
Kafka is used as a message broker to ingest data into Spark Streaming for real-time processing.
Kafka acts as a buffer between data producers and Spark Streaming to handle high throughput of data
Spark Streaming can consume data from Kafka topics in micro-batches for real-time processing
Kafka provides fault-tolerance and scalability for streaming data processing in Spark
SQL query to find users who purchased 3 consecutive months in a year
Use a self join on the table to compare purchase months for each user
Group by user and year, then filter for counts of 3 consecutive months
Example: SELECT user_id FROM purchases p1 JOIN purchases p2 ON p1.user_id = p2.user_id WHERE p1.month = p2.month - 1 AND p2.month = p1.month + 1 GROUP BY p1.user_id, YEAR(p1.purchase_date) HAVING COUNT(DISTINCT...
What people are saying about Wipro
Optimizing Spark jobs involves tuning configurations, partitioning data, caching, and using efficient transformations.
Tune Spark configurations for memory, cores, and parallelism
Partition data to distribute workload evenly
Cache intermediate results to avoid recomputation
Use efficient transformations like map, filter, and reduce
Avoid shuffling data unnecessarily
SQL query to find the second highest salary of employees in each department
Use a subquery to rank the salaries within each department
Filter the results to only include the second highest salary for each department
Join the result with the employee table to get additional information if needed
Spark is a distributed computing framework that provides an interface for programming entire clusters with implicit data parallelism and fault tolerance.
Spark is built around the concept of Resilient Distributed Datasets (RDDs) which are immutable distributed collections of objects.
It supports various programming languages like Java, Scala, Python, and R.
Spark provides high-level APIs like Spark SQL for structured...
I applied via Approached by Company and was interviewed in Nov 2024.Β There was 1 interview round.
Use SQL window functions to identify the highest sale amount for each store per day.
Use the ROW_NUMBER() function to rank sales within each day and store.
Partition the data by store and date to isolate daily sales.
Order the sales in descending order to get the highest sale at the top.
Example SQL query: SELECT store_id, sale_date, sale_amount, ROW_NUMBER() OVER (PARTITION BY store_id, sale_date ORDER BY sale_amount DESC...
Design an ETL pipeline to handle irregularly timed JSON file uploads for data transformation and schema matching.
Use a cloud storage service (e.g., AWS S3) to store incoming JSON files.
Implement a file watcher or event-driven architecture (e.g., AWS Lambda) to trigger processing when new files arrive.
Utilize a data processing framework (e.g., Apache Spark or Apache Beam) to read and transform the JSON data.
Define a sch...
I applied via Approached by Company and was interviewed in May 2024.Β There was 1 interview round.
Spark is a distributed computing framework that provides an interface for programming entire clusters with implicit data parallelism and fault tolerance.
Spark is built around the concept of Resilient Distributed Datasets (RDDs) which are immutable distributed collections of objects.
It supports various programming languages like Java, Scala, Python, and R.
Spark provides high-level APIs like Spark SQL for structured data...
Optimizing Spark jobs involves tuning configurations, partitioning data, caching, and using efficient transformations.
Tune Spark configurations for memory, cores, and parallelism
Partition data to distribute workload evenly
Cache intermediate results to avoid recomputation
Use efficient transformations like map, filter, and reduce
Avoid shuffling data unnecessarily
SQL query to find the second highest salary of employees in each department
Use a subquery to rank the salaries within each department
Filter the results to only include the second highest salary for each department
Join the result with the employee table to get additional information if needed
SQL query to find users who purchased 3 consecutive months in a year
Use a self join on the table to compare purchase months for each user
Group by user and year, then filter for counts of 3 consecutive months
Example: SELECT user_id FROM purchases p1 JOIN purchases p2 ON p1.user_id = p2.user_id WHERE p1.month = p2.month - 1 AND p2.month = p1.month + 1 GROUP BY p1.user_id, YEAR(p1.purchase_date) HAVING COUNT(DISTINCT MONT...
The number of records retrieved depends on the default join type, which is usually INNER JOIN, returning matching records only.
If INNER JOIN is assumed, only records with matching keys in both tables are returned.
For example, if Table A has 5 records and Table B has 3 matching records, the result will be 3.
If LEFT JOIN is assumed, all records from the left table and matching records from the right table are returned.
Fo...
Dynamic break in report allows for automatic page breaks based on data.
Define a break field in the report design
Set the break field to trigger a page break when its value changes
Use dynamic break to automatically adjust page breaks based on data
Example: A report on sales by region can use dynamic break to create a new page for each region
Foreach function is used to iterate through an array of strings in a report.
Foreach function is used to loop through each element in an array of strings.
It can be used to display each element in a report.
Example: foreach(string item in stringArray){ Console.WriteLine(item); }
Aggregate awareness is a feature in Business Objects that allows reports to be optimized for performance by using aggregate tables.
Aggregate awareness is used when creating reports in Business Objects
It allows the report to use aggregate tables instead of detailed tables for improved performance
Syntax: AGGR_AWARE( Example: AGGR_AWARE(Product, Sum(Sales), Sales_Fact_Table) The @prompt function syntax is used to display a prompt message to the user and receive input. The syntax is @prompt('message', 'default value') The message is the prompt message to be displayed to the user The default value is optional and will be displayed in the input field The function returns the value entered by the user
I applied via
Campus Placement
and was interviewed before Mar 2020.Β There
were
5 interview rounds.
I applied via
Campus Placement
and was interviewed in Oct 2020.Β There
were
4 interview rounds.
I applied via
Recruitment Consulltant
and was interviewed before Apr 2021.Β There
were
3 interview rounds.
Standard Aptitude questions Based on strings and array
I applied via
Company Website
and was interviewed in Feb 2021.Β There
was
1 interview round.
posted on
26 Jan 2022
I applied via
Naukri.com
and was interviewed before Jan 2021.Β There
was
1 interview round.
Some of the top questions asked at the Wipro Lead Data Engineer interview -
based on 3 interview experiences
Difficulty level
Duration
based on 2
reviews
Rating in categories
βΉ3.2
L/yr - βΉ7.5
L/yr
βΉ6.7
L/yr - βΉ19.5
L/yr
βΉ1.8
L/yr - βΉ5.5
L/yr
βΉ16.6
L/yr - βΉ30
L/yr
βΉ6.5
L/yr - βΉ18.4
L/yr
TCS Infosys Tesla Amazon)
Skills evaluated in this interview
Interview QuestionnaireΒ
1 Question
Interview Preparation Tips
Interview QuestionnaireΒ
3 Questions
Interview Preparation Tips
(1 Question)
Interview Preparation Tips
It's more on your self development and your carrier.
Interview QuestionnaireΒ
2 Questions
Interview Preparation Tips
Senior Software Engineer Interview Questions & Answers
Tech Mahindra
Suraj B R
Interview QuestionnaireΒ
1 Question
Interview Preparation Tips
Interview QuestionnaireΒ
1 Question
Interview QuestionnaireΒ
1 Question
Skills evaluated in this interview
Tell us how to improve this page.
Wipro Interviews By Designations
Interview Questions for Popular Designations
Overall Interview Experience Rating
Interview Questions from Similar Companies
Wipro Lead Data Engineer Reviews and Ratings
Explore more salaries
Project Engineer
Senior Software Engineer
Senior Associate
Technical Lead
Senior Project Engineer