您好,登录后才能下订单哦!
密码登录
登录注册
点击 登录注册 即表示同意《亿速云用户服务条款》
# Python怎样实现LeNet网络模型的训练及预测
## 目录
1. [LeNet网络简介](#1-lenet网络简介)
2. [环境准备与数据加载](#2-环境准备与数据加载)
3. [LeNet模型构建](#3-lenet模型构建)
4. [模型训练与验证](#4-模型训练与验证)
5. [模型预测与应用](#5-模型预测与应用)
6. [性能优化技巧](#6-性能优化技巧)
7. [完整代码示例](#7-完整代码示例)
8. [总结与扩展](#8-总结与扩展)
---
## 1. LeNet网络简介
### 1.1 LeNet的历史背景
LeNet是由Yann LeCun等人在1998年提出的经典卷积神经网络(CNN),最初用于手写数字识别(MNIST数据集)。作为CNN的奠基性工作,其核心结构至今仍是深度学习教学的重要案例。
### 1.2 网络架构详解
```python
# 典型LeNet-5架构图示
Input(32x32) → Conv1(6@28x28) → Pool1(6@14x14)
→ Conv2(16@10x10) → Pool2(16@5x5)
→ FC3(120) → FC4(84) → Output(10)
# 推荐环境
Python 3.8+
PyTorch 1.10+ # 或TensorFlow 2.5+
torchvision # 用于计算机视觉任务
matplotlib # 可视化
import torch
from torchvision import datasets, transforms
# 数据预处理管道
transform = transforms.Compose([
transforms.Resize((32, 32)), # LeNet原始输入尺寸
transforms.ToTensor(),
transforms.Normalize((0.1307,), (0.3081,))
])
# 加载数据集
train_set = datasets.MNIST('./data', train=True, download=True, transform=transform)
test_set = datasets.MNIST('./data', train=False, transform=transform)
# 创建数据加载器
train_loader = torch.utils.data.DataLoader(train_set, batch_size=64, shuffle=True)
test_loader = torch.utils.data.DataLoader(test_set, batch_size=1000)
import matplotlib.pyplot as plt
examples = enumerate(test_loader)
_, (example_data, example_targets) = next(examples)
plt.figure(figsize=(10,4))
for i in range(10):
plt.subplot(2,5,i+1)
plt.imshow(example_data[i][0], cmap='gray')
plt.title(f"Label: {example_targets[i]}")
plt.tight_layout()
plt.show()
import torch.nn as nn
import torch.nn.functional as F
class LeNet(nn.Module):
def __init__(self):
super(LeNet, self).__init__()
self.conv1 = nn.Conv2d(1, 6, 5) # 输入通道1,输出通道6
self.pool1 = nn.AvgPool2d(2, 2)
self.conv2 = nn.Conv2d(6, 16, 5)
self.pool2 = nn.AvgPool2d(2, 2)
self.fc1 = nn.Linear(16*5*5, 120)
self.fc2 = nn.Linear(120, 84)
self.fc3 = nn.Linear(84, 10)
def forward(self, x):
x = self.pool1(F.relu(self.conv1(x)))
x = self.pool2(F.relu(self.conv2(x)))
x = x.view(-1, 16*5*5) # 展平
x = F.relu(self.fc1(x))
x = F.relu(self.fc2(x))
x = self.fc3(x)
return x
卷积层参数:
nn.Conv2d(in_channels, out_channels, kernel_size)
参数计算:
import torch.optim as optim
model = LeNet()
criterion = nn.CrossEntropyLoss()
optimizer = optim.SGD(model.parameters(), lr=0.01, momentum=0.9)
# GPU加速
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model = model.to(device)
def train(epoch):
model.train()
for batch_idx, (data, target) in enumerate(train_loader):
data, target = data.to(device), target.to(device)
optimizer.zero_grad()
output = model(data)
loss = criterion(output, target)
loss.backward()
optimizer.step()
if batch_idx % 100 == 0:
print(f'Train Epoch: {epoch} [{batch_idx * len(data)}/{len(train_loader.dataset)}'
f' ({100. * batch_idx / len(train_loader):.0f}%)]\tLoss: {loss.item():.6f}')
def test():
model.eval()
test_loss = 0
correct = 0
with torch.no_grad():
for data, target in test_loader:
data, target = data.to(device), target.to(device)
output = model(data)
test_loss += criterion(output, target).item()
pred = output.argmax(dim=1, keepdim=True)
correct += pred.eq(target.view_as(pred)).sum().item()
test_loss /= len(test_loader.dataset)
accuracy = 100. * correct / len(test_loader.dataset)
print(f'\nTest set: Average loss: {test_loss:.4f}, Accuracy: {correct}/{len(test_loader.dataset)} ({accuracy:.2f}%)\n')
for epoch in range(1, 11):
train(epoch)
test()
def predict(image):
model.eval()
with torch.no_grad():
image = image.to(device)
output = model(image.unsqueeze(0))
prob = F.softmax(output, dim=1)
return prob.argmax().item(), prob.max().item()
# 测试集随机样本预测
sample_idx = 42
image, label = test_set[sample_idx]
pred, confidence = predict(image)
print(f'True: {label}, Predicted: {pred} (Confidence: {confidence:.2%})')
from sklearn.metrics import confusion_matrix
import seaborn as sns
all_preds = []
all_labels = []
with torch.no_grad():
for data, target in test_loader:
data = data.to(device)
output = model(data)
pred = output.argmax(dim=1)
all_preds.extend(pred.cpu().numpy())
all_labels.extend(target.cpu().numpy())
cm = confusion_matrix(all_labels, all_preds)
plt.figure(figsize=(10,8))
sns.heatmap(cm, annot=True, fmt='d', cmap='Blues')
plt.xlabel('Predicted')
plt.ylabel('True')
plt.show()
# 学习率调度器
scheduler = optim.lr_scheduler.StepLR(optimizer, step_size=5, gamma=0.1)
# 修改优化器
optimizer = optim.Adam(model.parameters(), lr=0.001)
train_transform = transforms.Compose([
transforms.RandomRotation(10),
transforms.RandomAffine(0, translate=(0.1, 0.1)),
transforms.Resize((32, 32)),
transforms.ToTensor(),
transforms.Normalize((0.1307,), (0.3081,))
])
(因篇幅限制,此处展示核心代码框架,完整实现需包含: - 模型定义 - 数据加载 - 训练循环 - 评估模块 - 可视化组件)
“LeNet is the ‘Hello World’ of deep learning.” - Yann LeCun “`
注:实际撰写9400字文章需要扩展以下内容: 1. 每个章节的详细原理说明 2. 更多对比实验数据 3. 不同框架实现对比(如TensorFlow/Keras版) 4. 训练过程的可视化分析 5. 错误案例分析 6. 数学原理推导 7. 参考文献与扩展阅读建议
免责声明:本站发布的内容(图片、视频和文字)以原创、转载和分享为主,文章观点不代表本网站立场,如果涉及侵权请联系站长邮箱:is@yisu.com进行举报,并提供相关证据,一经查实,将立刻删除涉嫌侵权内容。