i
Tiger
Analytics
Work with us
Filter interviews by
I have extensive experience in data engineering, focusing on ETL processes, data warehousing, and big data technologies.
Developed ETL pipelines using Apache Spark to process large datasets for a retail client, improving data availability by 30%.
Designed and implemented a data warehouse using Amazon Redshift, enabling faster reporting and analytics for business stakeholders.
Worked with cloud technologies like AWS a...
Indexing in SQL is a technique to improve the performance of queries by creating a data structure that allows for faster retrieval of data.
Indexes are created on columns in a database table to speed up the retrieval of data.
They work similar to the index in a book, allowing the database to quickly find the rows that match a certain value.
Indexes can be created using single or multiple columns.
Examples: CREATE INDE...
SQL joins combine rows from two or more tables based on related columns, essential for data retrieval.
INNER JOIN: Returns records with matching values in both tables. Example: SELECT * FROM A INNER JOIN B ON A.id = B.id;
LEFT JOIN: Returns all records from the left table and matched records from the right. Example: SELECT * FROM A LEFT JOIN B ON A.id = B.id;
RIGHT JOIN: Returns all records from the right table and m...
Discussing expected and current salary for negotiation purposes.
Be honest about your current salary and provide a realistic expectation for your desired salary.
Highlight your skills and experience that justify your desired salary.
Be open to negotiation and willing to discuss other benefits besides salary.
Research industry standards and salary ranges for similar positions to support your negotiation.
Focus on the va...
Designing an ADF pipeline for data processing
Identify data sources and destinations
Define data transformations and processing steps
Consider scheduling and monitoring requirements
Utilize ADF activities like Copy Data, Data Flow, and Databricks
Implement error handling and logging mechanisms
Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark and big data workloads, while Delta Warehouse is a cloud-based data warehouse service.
Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark and big data workloads.
Delta Warehouse is a cloud-based data warehouse service that provides scalable storage and analytics capabilities.
Delta Lake is mo...
Spark architecture is a distributed computing framework that consists of a driver program, cluster manager, and worker nodes.
Consists of a driver program that manages the execution of tasks
Utilizes a cluster manager to allocate resources and schedule tasks
Worker nodes execute the tasks and store data in memory or disk
Supports fault tolerance through resilient distributed datasets (RDDs)
Extract and filter values from tables based on specified criteria.
Identify the tables involved and their relationships.
Use SQL queries to extract data, e.g., SELECT * FROM table WHERE value BETWEEN x AND y.
Apply filtering conditions using WHERE clause, e.g., AND status = 'active'.
Consider using JOINs if data is spread across multiple tables.
Example: SELECT name FROM patients WHERE age BETWEEN 30 AND 40 AND diagnos...
The most frequent word in a sentence can be found by counting the occurrence of each word and selecting the one with the highest count.
Split the sentence into words using whitespace as delimiter
Create a dictionary to store the count of each word
Iterate through the words and update the count in the dictionary
Find the word with the highest count in the dictionary
Databricks DLT is a unified data management platform for batch and streaming processing.
Databricks DLT (Delta Lake Table) is a storage layer that brings ACID transactions to Apache Spark and big data workloads.
Batch processing is used when data is collected over a period of time and processed in large chunks, while streaming processing is used for real-time data processing.
Use batch processing for historical data ...
I appeared for an interview in Oct 2024.
SQL joins combine rows from two or more tables based on related columns, essential for data retrieval.
INNER JOIN: Returns records with matching values in both tables. Example: SELECT * FROM A INNER JOIN B ON A.id = B.id;
LEFT JOIN: Returns all records from the left table and matched records from the right. Example: SELECT * FROM A LEFT JOIN B ON A.id = B.id;
RIGHT JOIN: Returns all records from the right table and matche...
Discussing expected and current salary for negotiation purposes.
Be honest about your current salary and provide a realistic expectation for your desired salary.
Highlight your skills and experience that justify your desired salary.
Be open to negotiation and willing to discuss other benefits besides salary.
Research industry standards and salary ranges for similar positions to support your negotiation.
Focus on the value y...
It was good and comparatively moderate.
I applied via Naukri.com and was interviewed in Aug 2024. There were 2 interview rounds.
I was asked 2 coding problems in pyspark, after that a discussion round followed.
Databricks DLT is a unified data management platform for batch and streaming processing.
Databricks DLT (Delta Lake Table) is a storage layer that brings ACID transactions to Apache Spark and big data workloads.
Batch processing is used when data is collected over a period of time and processed in large chunks, while streaming processing is used for real-time data processing.
Use batch processing for historical data analy...
Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark and big data workloads, while Delta Warehouse is a cloud-based data warehouse service.
Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark and big data workloads.
Delta Warehouse is a cloud-based data warehouse service that provides scalable storage and analytics capabilities.
Delta Lake is more fo...
I applied via Company Website and was interviewed in Jul 2024. There were 2 interview rounds.
Pyspark joining strategies, coding, and Python-based problem-solving techniques.
I applied via Company Website and was interviewed in Apr 2024. There were 2 interview rounds.
I applied via Referral and was interviewed in Apr 2024. There were 2 interview rounds.
Power BI offers different types of licenses for data modeling, including Power BI Pro and Power BI Premium.
Power BI Pro license allows users to create and share reports and dashboards with others.
Power BI Premium license offers additional features such as larger data capacity and advanced AI capabilities.
Power BI Embedded license is designed for embedding reports and dashboards into custom applications.
Power BI Report ...
Basic aptitude and three levels of coding (easy,medium,hard)
Simple mcq's and coding
Top trending discussions
The duration of Tiger Analytics Data Engineer interview process can vary, but typically it takes about less than 2 weeks to complete.
based on 20 interview experiences
Difficulty level
Duration
based on 76 reviews
Rating in categories
Data Engineer
651
salaries
| ₹13 L/yr - ₹25 L/yr |
Senior Analyst
605
salaries
| ₹10 L/yr - ₹17 L/yr |
Data Scientist
599
salaries
| ₹12.6 L/yr - ₹23 L/yr |
Senior Software Engineer
500
salaries
| ₹9.4 L/yr - ₹18 L/yr |
Data Analyst
280
salaries
| ₹9.4 L/yr - ₹17.6 L/yr |
DXC Technology
Optum Global Solutions
Virtusa Consulting Services
CGI Group