0% found this document useful (0 votes)
39 views1 page

Deep Learning Laboratory Exam Tasks

The document outlines a model exam for a Deep Learning Laboratory course, detailing 18 practical tasks for students to implement various neural network architectures. Tasks include solving the XOR problem with a DNN, training CNNs for digit and alphabet recognition, and developing RNNs for text generation and sentiment analysis. Additional projects involve using GANs for image generation and applying transfer learning for image classification on different datasets.

Uploaded by

kannan.niran
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
39 views1 page

Deep Learning Laboratory Exam Tasks

The document outlines a model exam for a Deep Learning Laboratory course, detailing 18 practical tasks for students to implement various neural network architectures. Tasks include solving the XOR problem with a DNN, training CNNs for digit and alphabet recognition, and developing RNNs for text generation and sentiment analysis. Additional projects involve using GANs for image generation and applying transfer learning for image classification on different datasets.

Uploaded by

kannan.niran
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd

SUBJECT NAME: DEEP LEARNING LABORATORY

SUBJECT CODE : AD23521


SEMESTER/YEAR: V / III

MODEL EXAM

1. Implement a Deep Neural Network (DNN) to solve the XOR problem. Visualize the learned decision
boundary.
2. Train a Convolutional Neural Network (CNN) for handwritten digit recognition using MNIST
dataset and report accuracy.
3. Build a Recurrent Neural Network (RNN) for text generation on a small dataset of your choice.
4. Implement a Multi-Layer Perceptron (MLP) to classify Iris dataset. Compare training and testing
accuracy.
5. Train a CNN to recognize alphabets using EMNIST dataset. Display confusion matrix.
6. Implement a simple RNN to predict the next word in a given text sequence.
7. Develop a CNN-based face recognition system. Use embeddings to classify faces using cosine
similarity or k-NN.
8. Implement an LSTM-based model for sentiment analysis using the IMDB movie review dataset.
Report accuracy and confusion matrix.
9. Build a Sequence-to-Sequence model for Parts-of-Speech (POS) tagging. Demonstrate predictions on
at least 3 input sentences.
10. Implement an Encoder–Decoder model for machine translation (English → French). Demonstrate
with test sentences.
11. Build a CNN for object recognition on CIFAR-10 dataset. Show training/validation accuracy curves.
12. Implement a Bi-directional LSTM for text classification. Compare with a vanilla RNN.
13. Construct a Sequence-to-Sequence model for question answering (input: simple question, output:
answer). Demonstrate at least 3 queries.
14. Train a Variational Autoencoder (VAE) on MNIST dataset. Display original vs reconstructed
images.
15. Design and implement a Generative Adversarial Network (GAN) for image augmentation. Display
generated synthetic images alongside originals.
16. Apply Transfer Learning using a pre-trained CNN (VGG16/ResNet) for flower image classification.
Compare results with a simple CNN.
17. Design a GAN for handwritten digit generation. Display at least 5 synthetic digits generated.
18. Apply Transfer Learning using MobileNet/ResNet for classifying medical images (X-ray dataset).
Evaluate with precision, recall, and F1-score.

Common questions

Powered by AI

LSTM-based models offer significant advantages for sentiment analysis due to their capability to learn long-term dependencies, handling vanishing gradient problems that traditional RNNs suffer from. LSTMs leverage memory cells and gating mechanisms to retain relevant information over longer sequences, which is crucial for capturing context in sentiment analysis tasks where earlier words can significantly affect the interpretation of later ones. When applied to datasets like the IMDB movie review dataset, LSTMs can effectively identify sentiment shifts across a review, leading to more accurate sentiment classification outcomes than simpler RNN structures that may fail to maintain contextual information over extensive reviews.

The XOR problem is a classic example of a linearly inseparable problem that cannot be solved by a single-layer perceptron due to its non-linearity. Implementing a DNN to solve the XOR problem involves creating a neural network with multiple layers and non-linear activation functions like ReLU or sigmoid to handle the required complexity. Visualizing the learned decision boundary helps in understanding how well the DNN has managed to separate the input space into distinct regions corresponding to different outputs. This visualization can illustrate whether the model has adequately captured the underlying patterns and interactions in the data.

CNNs enhance the performance of handwritten digit recognition tasks because they automatically learn spatial hierarchies of features through convolutional layers, pooling layers, and multiple stages of feature extraction. Unlike traditional machine learning models, which may rely heavily on manual feature extraction, CNNs efficiently capture patterns such as edges, textures, and shapes, essential for recognizing handwritten digits. The built-in translational invariance and hierarchical feature learning allow CNNs to achieve higher accuracy rates by effectively leveraging the spatial structure of image data that characterizes renowned datasets like MNIST.

Encoder-Decoder models are instrumental in machine translation as they involve an encoder that processes the input sentence into a fixed-size context vector and a decoder that generates the translated sentence in the target language. This architecture captures complex relationships between source and target languages, allowing for non-linear transformations necessary for translation. However, limitations arise in handling long sentences where fixed-size context vectors may lose essential details due to compression, resulting in loss of meaning or nuances in translation from English to French. Further improvements can be achieved by implementing attention mechanisms that allow the model to focus on specific parts of the source sentence dynamically for better translation accuracy.

GANs for image augmentation are crucial as they generate new synthetic data, enhancing the dataset's diversity without manual data creation. Unlike traditional augmentation techniques (rotations, flips, scaling, etc.), which manipulate existing images, GANs can create entirely new images with variations in style, color, and features that aren't present in the original dataset. This capability helps models generalize better to unseen data by reducing overfitting and creating a more robust training set. However, GAN-generated images must maintain quality to avoid introducing noise in the dataset, marking a potential downside compared to deterministic traditional methods.

Bi-directional LSTMs differ from vanilla RNNs by processing the input data in both forward and backward directions, allowing the network to capture context from past and future states simultaneously. This bidirectional processing can uncover dependencies that a single directional RNN might miss, especially when the significance of words depends on their position in the text. In text classification tasks, this can translate to more accurate model predictions, as the model is equipped with comprehensive context awareness, leading to better performance and resilience against sequence length variations that can affect the outcome in vanilla RNNs.

MLPs are a type of neural network that learns representations by adjusting weights through backpropagation, focusing on reducing error over a series of layers through the input dataset. SVMs, on the other hand, find the optimal hyperplane that maximizes the margin between different classes in the training dataset. When comparing these models on datasets like the Iris dataset, MLPs excel in discovering underlying complex patterns through non-linear functions but may overfit if not regularized properly, leading to discrepancies between training and testing accuracy. SVMs often provide better generalization by focusing on decision boundaries but might struggle with larger, more complex datasets due to computational limitations. Thus, MLPs might show variability in accuracy due to their dependency on hyperparameter tuning, whereas SVMs might provide more consistent accuracy outcomes with effective kernel selection.

Transfer learning using pre-trained models like VGG16 or ResNet improves flower image classification by leveraging prior knowledge captured on large and diverse datasets such as ImageNet. These models have learned robust feature representations that can generalize to new, similar tasks. Transfer learning allows the network to adapt quickly with fewer training samples, reducing the computational effort and time needed compared to training a simple CNN from scratch. It helps circumvent the need for large labeled datasets and provides higher feature extraction accuracy, making it highly efficient for tasks with limited domain-specific data.

A VAE trained on the MNIST dataset showcases its image reconstruction capability by encoding input images into a latent space that captures essential characteristics and then reconstructs these images from this reduced representation. This process highlights the model’s ability to learn a compressed but meaningful representation of input data. However, limitations include potential blurriness or quality loss in reconstructed images, as the stochastic nature of VAEs introduces some degree of variability in output. Additionally, VAEs assume continuous latent spaces and may struggle with high-dimensional discrete datasets without significant adaptation or tuning.

CNNs play a pivotal role in object recognition tasks on datasets like CIFAR-10 by effectively extracting hierarchical features through convolutional and pooling layers. These layers detect local patterns like edges at lower levels and more complex features at higher levels, crucial for accurate object recognition. Evaluating training and validation accuracy curves provides insights into the model's learning process. If both curves converge to a high accuracy value, it indicates good generalization; however, if the training accuracy is high and validation accuracy is low, it suggests overfitting to the training data. Employing techniques like dropout, data augmentation, or fine-tuning can help mitigate overfitting.

You might also like