i
Ernst &
Young
Filter interviews by
Find the highest salary for each employee in their respective departments using SQL queries.
Use the 'employees' table to get employee details including salary.
Join the 'employees' table with the 'departments' table on department ID.
Use the SQL 'GROUP BY' clause to group results by department.
Utilize the 'MAX()' function to find the highest salary within each department.
Example SQL query: SELECT department_id, MAX(...
Use SQL query to find the maximum salary for each department
Use GROUP BY clause to group the data by department
Use MAX() function to find the maximum salary within each group
Join the tables if necessary to get department information
Star schema has a single fact table connected to multiple dimension tables, while Snowflake schema has normalized dimension tables.
Star schema denormalizes data for faster query performance
Snowflake schema normalizes data to reduce redundancy
Star schema is easier to understand and query
Snowflake schema is more flexible and scalable
Example: A star schema for a sales database would have a fact table for sales transa...
Transient tables are temporary tables, fact tables contain quantitative data, and dimension tables contain descriptive data.
Transient tables are temporary and used for intermediate processing or storing temporary data.
Fact tables contain quantitative data such as sales, revenue, or quantities.
Dimension tables contain descriptive data like customer names, product categories, or dates.
What people are saying about Ernst & Young
List and tuple are both sequence data types in Python, but the main difference is that lists are mutable while tuples are immutable.
Lists are enclosed in square brackets [], while tuples are enclosed in parentheses ().
Lists can be modified by adding, removing, or changing elements, while tuples cannot be modified once created.
Lists are typically used for collections of similar items, while tuples are used for hete...
SQL order of execution determines the sequence in which different clauses are processed in a query.
SQL query is parsed and validated first
Next, the query optimizer creates an execution plan
Execution plan includes steps like table scans, index scans, joins, etc.
Finally, the query is executed and results are returned
Spark optimization techniques improve performance and efficiency of Spark jobs.
Use partitioning to distribute data evenly across nodes
Cache intermediate results to avoid recomputation
Optimize shuffle operations by reducing data shuffling
Use broadcast variables for small lookup tables
Tune memory and executor settings for optimal performance
I applied via Recruitment Consulltant and was interviewed in Dec 2024. There was 1 interview round.
SQL order of execution determines the sequence in which different clauses are processed in a query.
SQL query is parsed and validated first
Next, the query optimizer creates an execution plan
Execution plan includes steps like table scans, index scans, joins, etc.
Finally, the query is executed and results are returned
Find the highest salary for each employee in their respective departments using SQL queries.
Use the 'employees' table to get employee details including salary.
Join the 'employees' table with the 'departments' table on department ID.
Use the SQL 'GROUP BY' clause to group results by department.
Utilize the 'MAX()' function to find the highest salary within each department.
Example SQL query: SELECT department_id, MAX(salar...
I applied via Recruitment Consulltant and was interviewed in Dec 2024. There was 1 interview round.
Use SQL query to find the maximum salary for each department
Use GROUP BY clause to group the data by department
Use MAX() function to find the maximum salary within each group
Join the tables if necessary to get department information
Amazon Redshift is a data warehouse service, while DynamoDB is a NoSQL database service for key-value and document data.
Redshift is optimized for complex queries and analytics, while DynamoDB is designed for high-speed transactions.
Redshift uses a columnar storage format, making it efficient for analytical queries; DynamoDB uses key-value pairs for quick lookups.
Redshift is suitable for large-scale data warehousing, wh...
Query to find the second-highest average salary in each department using SQL.
Use the AVG() function to calculate average salaries per department.
Utilize a subquery to rank the average salaries within each department.
Use the ROW_NUMBER() or RANK() window function to assign ranks to average salaries.
Filter the results to get only the second-highest average salary.
AWS S3 offers various storage classes to optimize cost and performance based on data access patterns.
S3 Standard: General-purpose storage for frequently accessed data.
S3 Intelligent-Tiering: Automatically moves data between two access tiers when access patterns change.
S3 Standard-IA (Infrequent Access): Lower-cost storage for infrequently accessed data, with retrieval fees.
S3 One Zone-IA: Lower-cost option for infreque...
Star schema has a single fact table connected to multiple dimension tables, while Snowflake schema has normalized dimension tables.
Star schema denormalizes data for faster query performance
Snowflake schema normalizes data to reduce redundancy
Star schema is easier to understand and query
Snowflake schema is more flexible and scalable
Example: A star schema for a sales database would have a fact table for sales transaction...
Transient tables are temporary tables, fact tables contain quantitative data, and dimension tables contain descriptive data.
Transient tables are temporary and used for intermediate processing or storing temporary data.
Fact tables contain quantitative data such as sales, revenue, or quantities.
Dimension tables contain descriptive data like customer names, product categories, or dates.
I applied via Company Website and was interviewed in Sep 2024. There were 2 interview rounds.
Spark and SQL question can expect
Focusing on the past project
Spark optimization techniques improve performance and efficiency of Spark jobs.
Use partitioning to distribute data evenly across nodes
Cache intermediate results to avoid recomputation
Optimize shuffle operations by reducing data shuffling
Use broadcast variables for small lookup tables
Tune memory and executor settings for optimal performance
I applied via Referral and was interviewed in May 2024. There was 1 interview round.
I applied via LinkedIn and was interviewed in Jan 2024. There were 4 interview rounds.
List and tuple are both sequence data types in Python, but the main difference is that lists are mutable while tuples are immutable.
Lists are enclosed in square brackets [], while tuples are enclosed in parentheses ().
Lists can be modified by adding, removing, or changing elements, while tuples cannot be modified once created.
Lists are typically used for collections of similar items, while tuples are used for heterogen...
I appeared for an interview in Jan 2024.
Some of the top questions asked at the Ernst & Young Data Engineer interview -
based on 17 interview experiences
Difficulty level
Duration
based on 28 reviews
Rating in categories
Bangalore / Bengaluru
8-13 Yrs
Not Disclosed
Senior Consultant
19.2k
salaries
| ₹9.1 L/yr - ₹30 L/yr |
Consultant
13k
salaries
| ₹6.6 L/yr - ₹21 L/yr |
Manager
8k
salaries
| ₹16.5 L/yr - ₹50 L/yr |
Assistant Manager
6.7k
salaries
| ₹9.3 L/yr - ₹28.3 L/yr |
Associate Consultant
4.2k
salaries
| ₹4.7 L/yr - ₹12 L/yr |
Deloitte
PwC
EY Global Delivery Services ( EY GDS)
Accenture