Your AI Wingman for your next interview

The most comprehensive bank Interview Answer Review tooling available online.

2 months ago
Behavioral
In your opinion, what has been the most innovative idea you've ever had? How did you create it and how did you implement it?
Data ScientistData Engineering ManagerEngineering ManagerUX Designer

Expedia

NetApp Logo

NetApp

Bill.com Logo

Bill.com

+7

When I joined my new team , I took lead on a new project which was complex in nature, had many unknowns and our team had to work on new framework, techstack, data models. The team was following agile methodology and will start with the work in first sprint. there was no concept of zero sprint or story grooming prior to the start of first sprint.

with the existing approach I envisioned some issues:

1. Because the number of unknown were high, the chances of doing wrong estimations for development were high.

2. There were high chances that the prioritisation is not right and we might have to revisit the features again and again.

3.  We might find out about unknowns while we are in the development process and that might cause delays.

To overcome these issues, I suggested some changes in our approach:

1. Increase the planning for one more release cycle and group the features as now, next and later.

2. For the features planned in the now group , prioritise them for the next release and complete the research and PRDs for them one sprint early. That means dev and QA will have 2 weeks extra to go through the stories, find out the dependencies and share the feedback. This helped in identifying if we need POCs for any feature or we can right away start developing them. Or if further modularisation is required.

3. I suggested to track the features related to new project in a different group, because to begin with the pace of the delivery will be slow.

When we begin, we took one complete release only in the RnD and understanding the new framework and doing small POCs to validate our understanding.

From second release, when we as a team were ready with good understanding on the expectations and other tech dependencies we started delivering and we could complete 2 full flows from end to end.

We had very less amount of spillover and could complete almost everything in the backlog, which was a big issue in earlier releases.

5 months ago
ML Knowledge
What is the principle behind bootstrapping, and would you recommend using it to increase sample sizes?
Data ScientistMachine Learning Engineer

Expedia

Atlassian Logo

Atlassian

Dell Logo

Dell

Bootstrapping is a statistical resampling method used to estimate the distribution of a sample statistic (like the mean, median, variance) by repeatedly sampling from the original dataset with replacement. The primary idea is to generate "new" samples (called bootstrap samples) by randomly selecting data points from the original sample, allowing some points to be selected multiple times while others may not be selected at all.

Steps of Bootstrapping:

  1. Original sample: Start with a dataset of size n.
  2. Resampling: Generate multiple new datasets (bootstrap samples) of the same size n by sampling with replacement from the original dataset.
  3. Statistic calculation: For each bootstrap sample, calculate the statistic of interest (e.g., the mean).
  4. Aggregation: After many resamplings (typically thousands), aggregate these statistics to estimate properties like confidence intervals, standard errors, or the distribution of the statistic.

Efficacy in Augmenting Sample Size:

While bootstrapping doesn’t actually create new, independent data, it is effective at enhancing statistical insights from small samples by simulating variability and giving a better approximation of the underlying population’s distribution. Its efficacy is most pronounced when:

  • Small samples: Bootstrapping is especially useful for datasets where traditional parametric methods may not be applicable due to the small sample size or assumptions (like normality).
  • Non-parametric nature: It does not require assumptions about the distribution of the data, making it versatile.
  • Uncertainty Estimation: It helps estimate confidence intervals, standard errors, and biases for small samples when direct analytical solutions are difficult.

However, since bootstrapping is based on the assumption that the original sample is representative of the population, its effectiveness can be limited when the original sample is biased or unrepresentative. It’s not a substitute for truly increasing the sample size but is a powerful technique for making the most of available data.

6 months ago

Showing 1 to 10 of 178 results

*All interview questions are submitted by recent Expedia Data Scientist candidates, labelled and categorized by Prepfully, and then published after being verified by Data Scientists at Expedia.