1. How do supervised and unsupervised learning differ in approach?
Ans:
Supervised learning relies on labeled datasets where both inputs and outputs are known, enabling models to make accurate predictions. Unsupervised learning, however, works with unlabeled data to discover patterns, structures, or clusters, such as segmenting customers by purchase behavior. While supervised learning teaches models to predict known results, unsupervised learning uncovers insights without explicit guidance.
2. How does transfer learning enhance model development?
Ans:
Transfer learning uses knowledge gained from training a model on a large dataset and applies it to a smaller, specific dataset. This technique reduces training time, improves accuracy, and is particularly valuable for generative AI projects with limited data. It allows models to leverage pre-existing knowledge instead of learning entirely from scratch.
3. What is overfitting, and how can it be prevented?
Ans:
Overfitting occurs when a model learns the training data too precisely, including noise, which reduces its ability to generalize to new data. Approaches such as dropout, regularization, pruning, and cross-validation help maintain model robustness. These techniques ensure that the model captures meaningful patterns without memorizing irrelevant details.
4. How do GANs generate realistic outputs?
Ans:
Generative Adversarial Networks (GANs) consist of a generator that produces synthetic data and a discriminator that evaluates its authenticity. The adversarial interaction between the two networks allows the generator to gradually improve, creating outputs like images, videos, or audio that closely resemble real-world examples.
5. How does attention improve performance in transformers?
Ans:
Attention mechanisms help models focus on the most important parts of input sequences. In transformers, self-attention evaluates the relationship between all tokens, capturing context and dependencies. This enables the model to understand sequences more effectively, improving tasks such as language translation, text generation, or summarization.
6. Why is feature engineering critical in AI models?
Ans:
Feature engineering involves transforming raw data into meaningful inputs for models. Creating, selecting, or modifying features helps the model identify relevant patterns, improving learning efficiency and enhancing prediction accuracy. Proper feature engineering ensures models perform well on both training and unseen data.
7. How should missing values in datasets be handled?
Ans:
Missing data can be managed using techniques like imputation with mean, median, or mode values, using models that handle gaps, or removing rows or columns with excessive missing information. Correct handling of missing values is essential for accurate analysis and maintaining model performance.
8. What is the difference between Random Forest and XGBoost?
Ans:
Random Forest builds multiple independent decision trees and averages their predictions to reduce variance and prevent overfitting. XGBoost constructs trees sequentially, learning from errors of previous trees, often achieving higher accuracy but requiring careful parameter tuning. Both methods are widely used for predictive modeling in AI.
9. How is model performance evaluated?
Ans:
The choice of evaluation metrics depends on the task. For regression problems, metrics like MSE, MAE, and R-squared are used, while classification tasks rely on accuracy, precision, recall, F1-score, and ROC-AUC. These metrics provide insight into the model’s reliability, accuracy, and ability to generalize to new data.
10. What ethical considerations are essential in AI development?
Ans:
Ethical AI development focuses on fairness, transparency, and accountability. Efforts should be made to minimize bias, protect user privacy, and provide interpretable outcomes. Following these principles ensures responsible AI deployment, builds user trust, and maintains compliance with societal and legal expectations.