0% found this document useful (0 votes)
149 views5 pages

Trends in Machine Learning Techniques

Unit 6 discusses recent trends in machine learning, focusing on advanced classification techniques beyond traditional methods. Key topics include self-supervised learning, few-shot learning, transfer learning, neural architecture search, explainable AI, federated learning, multi-modal learning, and graph neural networks, each with definitions, procedures, algorithms, applications, and examples. The document provides a comprehensive overview of how these innovations are applied in various real-world scenarios.

Uploaded by

Saurabh Sarkar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
149 views5 pages

Trends in Machine Learning Techniques

Unit 6 discusses recent trends in machine learning, focusing on advanced classification techniques beyond traditional methods. Key topics include self-supervised learning, few-shot learning, transfer learning, neural architecture search, explainable AI, federated learning, multi-modal learning, and graph neural networks, each with definitions, procedures, algorithms, applications, and examples. The document provides a comprehensive overview of how these innovations are applied in various real-world scenarios.

Uploaded by

Saurabh Sarkar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd

Unit 6 Machine Learning

Unit 6: Recent Trends in Various Learning Techniques and Classification Methods


explores the latest advancements and innovations in machine learning, especially
focusing on classification techniques that go beyond traditional supervised learning.
Below is a detailed breakdown of recent trends, including:
 Definitions
 Algorithms/Procedures
 Applications
 Real-world Examples

🟩 1. Self-Supervised Learning
🔹 Definition:
A form of unsupervised learning where models learn representations by solving pretext
tasks (e.g., predicting missing parts of input data) without requiring human-labeled data.
🔹 Procedure:
1. Design a pretext task like masked language modeling or image inpainting.
2. Train model to predict masked or missing parts.
3. Use learned representations for downstream tasks like classification.
🔹 Algorithms:
 BERT (Bidirectional Encoder Representations from Transformers) – NLP
 SimCLR, MoCo – Computer Vision
 Contrastive Predictive Coding (CPC)
🔹 Applications:
 Language understanding (e.g., translation, summarization)
 Image classification with minimal labels
 Speech recognition
✅ Example:
 Training a model to reconstruct masked words in a sentence (like in BERT), then
fine-tuning it for sentiment analysis.

🟩 2. Few-Shot / One-Shot Learning


🔹 Definition:
Learning models that can generalize from very few examples per class (few-shot) or
even one example (one-shot).
🔹 Procedure:
1. Use meta-learning or similarity-based methods.
2. Learn a metric space where similar inputs are close to each other.
3. Classify new samples based on nearest neighbors or prototypes.
🔹 Algorithms:
 Siamese Networks
 Prototypical Networks
 Matching Networks
 Meta-learning (MAML – Model-Agnostic Meta-Learning)
🔹 Applications:
 Rare disease diagnosis (limited labeled cases)
 Personalized recommendation systems
 Object detection in robotics
✅ Example:
 Recognizing a new type of malware from just one or two known samples.
🟩 3. Transfer Learning and Domain Adaptation
🔹 Definition:
Transfer learning involves using knowledge from a source domain/task to improve
performance on a target domain/task. Domain adaptation focuses on adapting models
when training and test data come from different distributions.
🔹 Procedure:
1. Pre-train on large, general dataset (e.g., ImageNet).
2. Fine-tune on smaller, specific dataset.
3. For domain adaptation, align feature distributions across domains.
🔹 Algorithms:
 Fine-tuning CNNs , Transformer-based models (e.g., BERT)
 Domain-Adversarial Neural Networks (DANN)
 CORAL (Correlation Alignment)
🔹 Applications:
 Medical imaging (train on natural images, adapt to X-rays)
 Cross-lingual NLP tasks
 Autonomous vehicles trained in simulation before real-world use
✅ Example:
 Using a pre-trained ResNet on ImageNet and fine-tuning it for plant disease
classification.

🟩 4. Neural Architecture Search (NAS)


🔹 Definition:
Automated design of neural network architectures using search algorithms to find
optimal structures for a given task.
🔹 Procedure:
1. Define a search space of possible architectures.
2. Evaluate candidate networks on validation data.
3. Use reinforcement learning, evolutionary algorithms, or gradient-based
optimization to evolve better models.
🔹 Algorithms:
 Reinforcement Learning-based NAS
 Differentiable Architecture Search (DARTS)
 Evolutionary NAS
🔹 Applications:
 Mobile-friendly models (e.g., MobileNet)
 Edge AI applications
 Custom hardware-aware model design
✅ Example:
 Google's AutoML uses NAS to design efficient vision models for mobile devices.

🟩 5. Explainable AI (XAI)
🔹 Definition:
Techniques to make ML models more interpretable and transparent, especially important
in high-stakes domains like healthcare and finance.
🔹 Procedure:
1. Apply post-hoc explanation tools to understand model decisions.
2. Visualize feature importance or decision paths.
3. Use inherently interpretable models if needed.
🔹 Algorithms:
 LIME (Local Interpretable Model-agnostic Explanations)
 SHAP (SHapley Additive exPlanations)
 Grad-CAM (for CNNs)
🔹 Applications:
 Explaining why a loan was denied
 Understanding which features drive medical diagnoses
 Debugging model bias
✅ Example:
 Using SHAP values to explain why a patient was predicted at high risk for heart
disease.

🟩 6. Federated Learning
🔹 Definition:
A distributed learning approach where models are trained across decentralized devices
holding local data, preserving privacy and reducing data transfer.
🔹 Procedure:
1. Devices train local models using their own data.
2. Models are sent to a central server.
3. Server aggregates updates (e.g., via Federated Averaging).
4. Updated global model is sent back to clients.
🔹 Algorithms:
 Federated Averaging (FedAvg)
 Secure Aggregation Protocols
 Differential Privacy in Federated Learning
🔹 Applications:
 Mobile keyboard prediction (Gboard)
 Healthcare data analysis across hospitals
 Smart home/IoT device learning
✅ Example:
 Training a voice assistant on user devices without uploading private voice
recordings.

🟩 7. Multi-modal Learning
🔹 Definition:
Combines multiple modalities (e.g., text, image, audio) into a single model to improve
understanding and classification.
🔹 Procedure:
1. Process each modality separately using specialized encoders.
2. Fuse modalities at intermediate or final layers.
3. Train end-to-end using multi-task loss or contrastive learning.
🔹 Algorithms:
 CLIP (Contrastive Language–Image Pre-training)
 Flamingo (multi-modal transformer)
 Late Fusion vs Early Fusion models
🔹 Applications:
 Caption generation for images
 Video question answering
 Assistive technologies for visually impaired users
✅ Example:
 A model that understands both visual content and spoken commands to assist in
navigation.

🟩 8. Graph Neural Networks (GNNs)


🔹 Definition:
Deep learning methods designed to operate on graph-structured data, where nodes
represent entities and edges represent relationships.
🔹 Procedure:
1. Aggregate information from neighboring nodes.
2. Update node embeddings iteratively.
3. Use final embeddings for classification or regression.
🔹 Algorithms:
 Graph Convolutional Network (GCN)
 Graph Attention Network (GAT)
 GraphSAGE
🔹 Applications:
 Social network analysis
 Drug discovery (molecular graphs)
 Recommendation systems
✅ Example:
 Detecting fraudulent transactions in a financial transaction graph.

✅ Summary Table
TREND TYPE CORE IDEA ALGORITHM(S) APPLICATION
Self- Representation Learn from BERT, SimCLR Language & Vision
Supervised Learning unlabeled data Tasks
Learning using pretext tasks
Few-Shot Generalization Learn from very ProtoNet, Siamese Rare disease diagnosis
Learning few examples Nets
Transfer Optimization Reuse knowledge Fine-tuning, DANN Medical imaging
Learning from source to
target domain
NAS Automation Automatically DARTS, Efficient model design
search best Reinforcement
architecture Learning
Explainabl Transparency Make model LIME, SHAP Financial and medical
e AI decisions decisions
interpretable
Federated Privacy- Train models FedAvg IoT, mobile health
Learning preserving ML without sharing
raw data
Multi- Integration Combine text, CLIP, Flamingo Image captioning, video
modal image, and audio QA
Learning for richer insights
GNNs Graph Modeling Learn from GCN, GAT Social networks, drug
relational data discovery
structures

Common questions

Powered by AI

Neural Architecture Search (NAS) automates the design of neural network architectures, leveraging algorithms like reinforcement learning, evolutionary algorithms, or gradient-based methods to explore a search space of potential architectures. For mobile devices, NAS identifies architectures that balance performance with resource constraints, leading to the creation of models such as MobileNet, which are optimized for mobile environments, ensuring efficient computation without sacrificing accuracy .

In high-stakes domains like healthcare, Explainable AI (XAI) is crucial for understanding model decisions to ensure transparency and trust. Tools like SHAP (SHapley Additive exPlanations) provide interpretable insights by illustrating which features contribute to a model's predictions, thus allowing medical professionals to validate and understand automated diagnoses. This contributes to improved accountability and can help detect biases in model predictions .

Transfer learning in medical imaging involves pre-training models on a large, general dataset like ImageNet and then fine-tuning them on smaller, specific medical datasets (e.g., X-rays). This process allows the model to leverage previously learned features, enhancing performance in the target domain. Domain adaptation further refines this by aligning feature distributions between training and test data from different distributions, making the model more robust to domain variations. These techniques can significantly improve diagnostic accuracy and efficiency in medical applications .

Few-shot learning addresses the challenge of learning from a very limited number of examples, which is common in rare disease diagnosis where available labeled cases are scarce. The advantage of this approach is that it can generalize well from few samples using techniques like meta-learning or similarity-based learning, such as Siamese or Prototypical Networks. However, the challenge lies in designing robust models that can handle noisy or low-quality examples and ensuring the model's ability to accurately classify new, unseen instances with minimal prior knowledge .

Federated learning enhances data privacy by training models locally on devices without sharing raw data with a central server, thus preserving user privacy and complying with regulations like GDPR. Additionally, it facilitates collaborative learning, as updates from decentralized devices are aggregated to improve the global model. This approach is advantageous for tasks like mobile keyboard prediction or healthcare data analysis where privacy is paramount, but poses challenges in terms of managing communication overhead and ensuring model convergence .

Contrastive Learning methods like SimCLR advance self-supervised learning by leveraging instance discrimination tasks where the model learns to distinguish between augmented views of the same image and different images. This approach allows the network to learn meaningful representations without requiring labeled data, effectively improving performance on computer vision tasks such as image classification. By focusing on the similarity between different views of the same data, SimCLR enhances the model's ability to generalize and learn robust features useful for downstream tasks .

Multi-modal learning enhances assistive technologies by integrating information from various modalities, such as visual, text, and audio data. For visually impaired users, this can improve accessibility through technologies that process visual content alongside spoken commands, thus providing richer, more contextual support. For example, a model could assist in navigation by understanding both environmental visuals (via cameras) and user instructions (via voice recognition), enhancing interaction with the environment .

Graph Neural Networks (GNNs) are highly suitable for social network analysis as they are designed to handle graph-structured data where nodes represent entities and edges represent relationships. GNNs, through algorithms like Graph Convolutional Networks (GCN) and Graph Attention Networks (GAT), can effectively capture and aggregate information from neighboring nodes, enabling them to model complex interactions and relationships within social networks. This capability is ideal for tasks such as community detection, influence maximization, and link prediction .

Applying Domain-Adversarial Neural Networks (DANN) requires careful consideration of the domain shift between training and testing data. DANN works by aligning feature distributions between source and target domains, using adversarial learning to make domain-invariant representations. Key considerations include selecting representative samples from both domains to train the adversarial component effectively, ensuring that the model learns features that are truly invariant to domain differences, and evaluating the risk of negative transfer where performance might degrade when domain alignment fails .

Self-supervised learning differs from supervised learning by not requiring human-labeled data for training. Instead, it uses pretext tasks, like masked language modeling or image inpainting, to learn data representations. This approach allows models such as BERT to effectively learn from large amounts of unlabeled data, which can then be fine-tuned for specific tasks like language understanding, where they have been shown to perform well in translation and summarization tasks without extensive labeled datasets .

You might also like