일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | 6 | 7 |
8 | 9 | 10 | 11 | 12 | 13 | 14 |
15 | 16 | 17 | 18 | 19 | 20 | 21 |
22 | 23 | 24 | 25 | 26 | 27 | 28 |
29 | 30 |
- 우분투db
- 프레임 구조
- 플로이드워셜
- 데이터 전송
- tcp 프로토콜
- well known 포트
- 오류검출
- 토큰 버스
- 스레드
- til
- 코딩테스트준비
- 주기신호
- 개발자취업
- 오블완
- reducible
- xv6
- 오류제어
- leetcode
- 항해99
- git merge
- 그리디 알고리즘
- 서비스 프리미티브
- 티스토리챌린지
- IEEE 802
- tcp 세그먼트
- i-type
- 비주기신호
- mariadb
- 99클럽
- 순서번호
- Today
- Total
Unfazed❗️🎯
Multi-Layered Perceptron (MLP) : Structure 본문
1. Input Layer:
Purpose: It receives the input data and acts as a gateway to funnel the data into the network.
Nodes: The number of nodes in the input layer typically matches the number of input features in the dataset. For instance, for an image that's 28x28 pixels (like those in the MNIST dataset), you'd have 784 input nodes.
2. Output Layer:
Purpose: This is where the network provides its prediction or classification based on the information it has learned during training.
Nodes: The number of nodes in the output layer depends on the type of task:
Regression: Usually a single node that gives a continuous value.
Binary Classification: Often one node (with a sigmoid activation function to provide an output between 0 and 1).
Multi-class Classification: One node for each class. For example, for a 10 class problem, there'd be 10 nodes, often with a softmax activation function to convert node values to probabilities.
3. Hidden Layers:
Purpose: Hidden layers enable the network to learn and represent more complex, higher-level features from the input data. They capture the intricate patterns by transforming the input data.
Nodes: The number of hidden layers and the number of nodes in each layer can vary widely. There's no strict rule for this, and it often requires experimentation. However, adding more hidden layers and nodes increases the capacity of the network, allowing it to learn more complex functions. But it also increases the risk of overfitting, especially if you have limited training data.
Activation Functions: It's important to note that nodes in the hidden layers typically use activation functions like ReLU (Rectified Linear Unit) to introduce non-linearity into the network. This non-linearity is crucial for the MLP's ability to learn from errors and adjust its weights, making it capable of modeling non-linear relationships in the data.
The arrangement and size of the layers determine the network's architecture. While deeper networks (more layers) can model more complex relationships, they also require more data and time to train and are more prone to overfitting if not handled properly (though techniques like dropout, regularization, etc., can mitigate this).
'AI > 머신러닝' 카테고리의 다른 글
ch2_learning2 편향-분산 Bias-variance Trade-off (0) | 2024.04.16 |
---|---|
Comparative Analysis of Classification Models: Logistic Regression, Naive Bayes, LDA, and QDA (0) | 2024.04.03 |
ch2_learning2 (Balancing Flexibility to Optimize Model Accuracy) 모델의 정확도 평가 (0) | 2024.03.13 |
ch2_learning (0) | 2024.03.12 |
ch1_overview (0) | 2024.03.06 |