Interview Experience for Product Developer Engineer I | Phenom People
General Questions
Round 1
- Introduce yourself
- Why do you want to switch?
- What are different RAG systems?
- How does LightGBM work?
- Boosting vs Bagging
- Why do we need activation functions in neural networks? Different types of activation functions? Why is ReLU more effective than sigmoid for intermediate layers?
- How to mitigate overfitting?
- Does batch size affect accuracy?
- What is the vanishing gradients problem and how do we handle it?
- How does attention work in transformers?
- Drawbacks of transformers
- Positional encoding in transformers
- If a model is developed and tested, but fails in production, what could be the possible reasons?
- If two models perform similarly, how do you decide which one to deploy (beyond inference time)?
Coding Questions:
-
First non-repeating character in a stream
def solution(x): non_repeating_chars = [x[0]] solution = "a" for _ in x[1:]: if _ not in non_repeating_chars: non_repeating_chars.append(_) solution += non_repeating_chars[0] continue else: non_repeating_chars = [character for character in non_repeating_chars if character != _] if len(non_repeating_chars) == 0: solution += "#" continue solution += non_repeating_chars[0] return solution
-
Minimum number of coins to make a sum
def solution(n, sum_): if sum_ == 0: return 0 if sum_ < 0: return None sols = [solution(n, sum_ - _) for _ in n] sols = [_ for _ in sols if _ is not None] return min(sols) + 1 solution([5, 1], 0)
Round 2
- Introduce yourself
- Why do you want to switch?
- Your best work in the past 1.5 years
- How does LoRA work, particularly for large models like Jais-75B?
- Azure Machine Learning
- What is attention?
- Architectural difference between T5 and GPT/Llama models
- Evolution from RNNs to LSTMs to Transformers
- How does LSTM solve the vanishing gradient problem?
- How does decoding happen in transformers?
- What is GRU, and which is better: GRU or LSTM?
- Difference between BERT and GPT
- How is BERT trained?
- How is GPT trained?
- Designing a neural network for a classification problem with text, categorical, and numerical features
- Handling imbalanced datasets
- Metrics for performance evaluation on imbalanced datasets
- Interpretation of train/validation/test accuracies (80/70/60)
- How to fix overfitting?
- What is an agent?
Round 3
- Introduce yourself
- Benefits of ChromaDB and comparison with its peers
- How to censor data retrieval based on user access
- Where do you see yourself in the next 5 years?
- One mistake you have made in a project and how you fixed it
- Why do you want to switch?
- How does RAG work?
Project-Specific Questions (Gov AI POC, Confluence Chatbot, etc.)
- Explain your current project (Gov AI - RAG Bot)
- Specifics about RAG:
- What chunking strategies did you use?
- What chunking strategies do you know?
- How can you improve chunking?
- Which framework did you use? (LangChain) and what class for chunking?
- Deployment Specifics:
- Chainlit framework
- Azure App Service
- What other components would you use if Azure infra isn’t available?
- Confluence chatbot:
- How to restrict document retrieval based on user access