加一个三分类

This commit is contained in:
琉璃月光
2025-08-14 18:27:52 +08:00
parent c1b5481a29
commit eae16a7a66
23 changed files with 541 additions and 9 deletions

144
class2/train_best_pt.py Normal file
View File

@ -0,0 +1,144 @@
import argparse
import torch
import torch.nn as nn
import torch.optim as optim
from torchvision import datasets, transforms
from torch.utils.data import DataLoader
from class2.test.mobilenetv3 import MobileNetV3_Large
import os
def train(args):
# 创建输出目录
os.makedirs("checkpoints", exist_ok=True)
# 数据增强 & 预处理
transform_train = transforms.Compose([
transforms.Resize((224, 224)),
transforms.RandomHorizontalFlip(),
transforms.ToTensor(),
transforms.Normalize(mean=[0.485, 0.456, 0.406],
std=[0.229, 0.224, 0.225])
])
transform_val = transforms.Compose([
transforms.Resize((224, 224)),
transforms.ToTensor(),
transforms.Normalize(mean=[0.485, 0.456, 0.406],
std=[0.229, 0.224, 0.225])
])
# 加载数据集
train_dataset = datasets.ImageFolder(root=f"{args.data}/train", transform=transform_train)
val_dataset = datasets.ImageFolder(root=f"{args.data}/val", transform=transform_val)
train_loader = DataLoader(train_dataset, batch_size=args.batch_size, shuffle=True, num_workers=4, pin_memory=True)
val_loader = DataLoader(val_dataset, batch_size=args.batch_size, shuffle=False, num_workers=4, pin_memory=True)
# 模型
model = MobileNetV3_Large(num_classes=2)
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model = model.to(device)
# 损失函数 & 优化器
criterion = nn.CrossEntropyLoss()
optimizer = optim.Adam(model.parameters(), lr=args.lr)
# 学习率调度器
scheduler = optim.lr_scheduler.StepLR(optimizer, step_size=30, gamma=0.1)
# ========== 改进的早停 + 保存机制 ==========
best_val_acc = 0.0
best_epoch = 0
patience = args.patience # 默认 5
wait = 0 # 连续未超越最佳的次数
print(f"开始训练,共 {args.epochs} 轮,早停耐心值: {patience}")
for epoch in range(args.epochs):
# 训练阶段
model.train()
running_loss = 0.0
correct, total = 0, 0
for images, labels in train_loader:
images, labels = images.to(device), labels.to(device)
optimizer.zero_grad()
outputs = model(images)
loss = criterion(outputs, labels)
loss.backward()
optimizer.step()
running_loss += loss.item()
_, predicted = outputs.max(1)
total += labels.size(0)
correct += predicted.eq(labels).sum().item()
train_acc = 100. * correct / total
avg_loss = running_loss / len(train_loader)
scheduler.step()
# 验证阶段
model.eval()
correct, total = 0, 0
with torch.no_grad():
for images, labels in val_loader:
images, labels = images.to(device), labels.to(device)
outputs = model(images)
_, predicted = outputs.max(1)
total += labels.size(0)
correct += predicted.eq(labels).sum().item()
val_acc = 100. * correct / total
print(f"[Epoch {epoch+1:2d}/{args.epochs}] "
f"Loss: {avg_loss:.4f} | "
f"Train Acc: {train_acc:.2f}% | "
f"Val Acc: {val_acc:.2f}% | "
f"LR: {scheduler.get_last_lr()[0]:.6f}")
# ========== 核心逻辑:保存最后一个最高权重 + 改进早停 ==========
if val_acc >= best_val_acc: # ✅ 只要不低于最佳,就视为“保持高水平”
if val_acc > best_val_acc:
print(f"新的最高验证准确率: {val_acc:.2f}% (原: {best_val_acc:.2f}%)")
else:
print(f"验证准确率持平历史最佳: {val_acc:.2f}%")
# 更新最佳
best_val_acc = val_acc
best_epoch = epoch + 1
wait = 0 # ✅ 重置等待计数
# 保存完整检查点(我们想要“最后一个”最高权重)
checkpoint = {
'epoch': epoch,
'model_state_dict': model.state_dict(),
'optimizer_state_dict': optimizer.state_dict(),
'scheduler_state_dict': scheduler.state_dict(),
'val_acc': val_acc,
'best_val_acc': best_val_acc,
'args': args
}
torch.save(checkpoint, "checkpoints/mobilenetv3_last_best.pth")
print(f"✅ 已保存当前最佳权重epoch {best_epoch}")
else:
# ✅ 只有严格低于最佳才增加 wait
wait += 1
print(f"验证准确率下降 ({val_acc:.2f}% < {best_val_acc:.2f}%),已连续 {wait}/{patience} 次未超越")
if wait >= patience:
print(f"早停触发!连续 {patience} 次验证准确率未达到或超越历史最佳。")
break
# 训练结束
print(f"\n✅ 训练完成!")
print(f"最终最佳验证准确率: {best_val_acc:.2f}% (第 {best_epoch} 轮)")
print(f"最后一个最高权重已保存至: 'checkpoints/mobilenetv3_last_best.pth'")
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("--data", type=str, default="/media/hx/04e879fa-d697-4b02-ac7e-a4148876ebb0/dataset/classdata", help="数据集路径")
parser.add_argument("--epochs", type=int, default=100, help="训练轮数")
parser.add_argument("--batch_size", type=int, default=32, help="批大小")
parser.add_argument("--lr", type=float, default=1e-3, help="学习率")
parser.add_argument("--patience", type=int, default=5, help="早停耐心值(连续多少次未达到或超越历史最佳才停止)")
args = parser.parse_args()
train(args)