时间:2025-07-25 作者:游乐小编
应该是全网第一个用paddle2.0+高层api实现小车车道线预测回归模型的搭建、训练和部署全流程的项目啦,希望能帮助做小车和机器人的小伙伴快速实现小车的无人驾驶功能~
车道线打角回归预测模型也是每年的全国大学生智能汽车竞赛都会包含的基础任务之一,希望能作为一个baseline让大家快速学习起来。
智能小车上使用标准的广角摄像头作为视觉传感器,其基本参数如下表所示:
采集的数据已上传至AI Studio:车道线检测回归数据集
import osimport cv2import iofrom tqdm import tqdmimport numpy as npimport matplotlib.pyplot as pltfrom PIL import Image as PilImageimport paddlefrom paddle.nn import functional as Fpaddle.__version__登录后复制
'2.2.0'登录后复制In [5]
!unzip -oq /home/aistudio/data/data46903/0728_carline.zip登录后复制
通过手柄遥控小车,小车实时记录当前图片和手柄的打角数据,从而得到数据集。
本数据集共13448张图像和对应的打角数据。
每一张图像都会对应一个打角数据。 图像和打角数据分别保存在img和data.txt下,在dataset的建立过程中,我们只需要将图像和打角数据一一对应进行封装即可。
数据对应形式如图所示:
img_folder_path = "data/img"dataset_path = "data"txt_path = "data/data.txt"IMAGE_SIZE = (224, 224)登录后复制In [6]
# 检查数据集中图像和打角数据个数是否对应import os# 1、读取图像数据和角度数据angle_list = []f = open(txt_path)for txt in f.readlines(): txt.split("\n") angle_num = int(txt) angle_list.append(angle_num)img_list = []for img_path in tqdm(os.listdir(img_folder_path)): img_path = os.path.join(img_folder_path, img_path) img_list.append(img_path)# 2、数据归一化(如果过于发散,模型不容易收敛)# 这里我的打角数据大小为:900-2100,中间值为1500for i, angle in enumerate(angle_list): angle = (angle-1500)/600 angle_list[i] = angletry: os.remove(os.path.join(dataset_path, "train.txt")) os.remove(os.path.join(dataset_path, "eval.txt"))except: passtrain_txt = open(os.path.join(dataset_path, "train.txt"), "w")eval_txt = open(os.path.join(dataset_path, "eval.txt"), "w")n = 0for img_path, label in tqdm(zip(img_list, angle_list)): n += 1 if n % 10 != 0: train_txt.write(img_path+" "+str(label)) train_txt.write("\n") else: eval_txt.write(img_path+" "+str(label)) eval_txt.write("\n")train_txt.close()eval_txt.close()登录后复制
创建PaddlePaddle高层API—Dataloder可直接读入的数据集格式
In [7]import randomimport iofrom paddle.io import Datasetfrom paddle.vision.transforms import transforms as Tfrom PIL import Image as PilImageimport numpy as npclass MyDataset(Dataset): """ 数据集定义 """ def __init__(self, mode, dataset_path): """ 构造函数 """ self.image_size = IMAGE_SIZE self.mode = mode.lower() self.dataset_path = dataset_path assert self.mode in ['train', 'test', 'eval'], \ "mode should be 'train' or 'test' or 'eval', but got {}".format(self.mode) self.train_images = [] self.label_list = [] with open(os.path.join(self.dataset_path, ('{}.txt'.format(self.mode))), 'r') as f: for line in tqdm(f.readlines()): image, label = line.strip().split(' ') img = PilImage.open(image) self.train_images.append(image) self.label_list.append(label) def _load_img(self, path, color_mode='rgb', transforms=[]): """ 统一的图像处理接口封装,用于规整图像大小和通道 """ with open(path, 'rb') as f: img = PilImage.open(io.BytesIO(f.read())) if color_mode == 'grayscale': # if image is not already an 8-bit, 16-bit or 32-bit grayscale image # convert it to an 8-bit grayscale image. if img.mode not in ('L', 'I;16', 'I'): img = img.convert('L') elif color_mode == 'rgba': if img.mode != 'RGBA': img = img.convert('RGBA') elif color_mode == 'rgb': if img.mode != 'RGB': img = img.convert('RGB') else: raise ValueError('color_mode must be "grayscale", "rgb", or "rgba"') return T.Compose([ T.Resize(self.image_size) ] + transforms)(img) def __getitem__(self, idx): """ 返回 image, label """ train_image = self._load_img(self.train_images[idx], transforms=[ T.Transpose(), T.Normalize(mean=[127.5], std=[127.5]) # 归一化处理,将0~255归一化到-1~1 ]) # 加载原始图像 label = self.label_list[idx] # 加载Label # 返回image, label train_image = np.array(train_image, dtype='float32') label = np.array(label, dtype='float32') return train_image, label def __len__(self): """ 返回数据集总数 """ return len(self.train_images)# 定义训练集和验证集train_dataset = MyDataset(mode='train', dataset_path=dataset_path) # 训练数据集val_dataset = MyDataset(mode='eval', dataset_path=dataset_path) # 验证数据集登录后复制
100%|██████████| 13448/13448 [00:01<00:00, 12012.05it/s]100%|██████████| 1494/1494 [00:00<00:00, 12004.41it/s]登录后复制
x, y = val_dataset.__getitem__(0)print(x, y)print(x.shape)print(train_dataset.__len__())with open(os.path.join(dataset_path, 'train.txt'), 'r') as f: i = 0 for line in f.readlines(): image_path, label = line.strip().split(' ') image = np.array(PilImage.open(image_path)) if i > 2: break # 进行图片的展示 plt.figure() plt.title(label) plt.imshow(image.astype('uint8')) plt.axis('off') plt.show() i = i + 1登录后复制
模型结构可自行修改以获得更好的效果。
In [86]import paddleimport paddle.nn as nnimport paddle.nn.functional as Fclass My_Model(nn.Layer): def __init__(self, num_classes): super(My_Model, self).__init__() self.conv1 = paddle.nn.Conv2D(in_channels=3, out_channels=32, kernel_size=(3, 3)) self.pool1 = paddle.nn.MaxPool2D(kernel_size=2, stride=2) self.conv2 = paddle.nn.Conv2D(in_channels=32, out_channels=64, kernel_size=(3,3)) self.pool2 = paddle.nn.MaxPool2D(kernel_size=2, stride=2) self.conv3 = paddle.nn.Conv2D(in_channels=64, out_channels=32, kernel_size=(3,3)) self.pool3 = paddle.nn.MaxPool2D(kernel_size=2, stride=2) self.conv4 = paddle.nn.Conv2D(in_channels=32, out_channels=16, kernel_size=(3,3)) self.flatten = paddle.nn.Flatten() self.linear1 = paddle.nn.Linear(in_features=9216, out_features=16) self.linear2 = paddle.nn.Linear(in_features=16, out_features=1) def forward(self, x): x = self.conv1(x) x = F.relu(x) x = self.pool1(x) x = self.conv2(x) x = F.relu(x) x = self.pool2(x) x = self.conv3(x) x = F.relu(x) x = self.pool3(x) x= self.conv4(x) x = F.relu(x) x = self.flatten(x) x = self.linear1(x) x = F.relu(x) x = self.linear2(x) return xpaddle.summary(My_Model(num_classes=1), input_size=(1,3, 224,224))登录后复制
import paddleclass SaveBestModel(paddle.callbacks.Callback): def __init__(self, target=0.5, path='./best_model', verbose=0): self.target = target self.epoch = None self.path = path def on_epoch_end(self, epoch, logs=None): self.epoch = epoch def on_eval_end(self, logs=None): if logs.get('loss')[0] < self.target: self.target = logs.get('loss')[0] self.model.save(self.path) print('best model is loss {} at epoch {}'.format(self.target, self.epoch))callback_visualdl = paddle.callbacks.VisualDL(log_dir='./')callback_savebestmodel = SaveBestModel(target=1, path='./model')callbacks = [callback_visualdl, callback_savebestmodel]# 模型初始化model = paddle.Model(My_Model(num_classes=1)) # 线性回归模型# 优化器optim = paddle.optimizer.Momentum(learning_rate=0.0001, momentum=0.9, parameters=model.parameters())# 损失函数loss = paddle.nn.MSELoss()model.prepare(optimizer=optim, loss=loss) # 线性回归模型一定不能用交叉熵,因为这个自带softmax,需要指定输出类别# 用 DataLoader 实现数据加载train_loader = paddle.io.DataLoader(train_dataset, places=paddle.CUDAPlace(0), batch_size=64)eval_loader = paddle.io.DataLoader(val_dataset, places=paddle.CUDAPlace(0), batch_size= 64)登录后复制
model.fit(train_loader, eval_loader, epochs=5, callbacks=callbacks, verbose=1)登录后复制
这里直接保存模型为可本地部署的形式。
In [10]# training=True时,只会保存优化器参数和模型参数# training=False时,会保存模型结构、参数和优化器结构model.save("./model_dir/model", training=False)登录后复制
这里的预测部分和本地预测代码相同,需要保证输入图像的预处理和训练时对齐。
In [90]import cv2import numpy as npfrom paddle.inference import Configfrom paddle.inference import create_predictor# ————————————————图像预处理函数————————————————def resize(img, target_size): """ resize """ percent = float(target_size) / min(img.shape[0], img.shape[1]) resized_width = int(round(img.shape[1] * percent)) resized_height = int(round(img.shape[0] * percent)) resized_short = cv2.resize(img, (resized_width, resized_height)) resized = cv2.resize(resized_short, (target_size, target_size)) return resizeddef preprocess(img, target_size): mean = [127.5, 127.5, 127.5] std = [127.5, 127.5, 127.5] # resize img = resize(img, target_size) # bgr-> rgb && hwc->chw img = img[:, :, ::-1].astype('float32').transpose((2, 0, 1)) img_mean = np.array(mean).reshape((3, 1, 1)) img_std = np.array(std).reshape((3, 1, 1)) img -= img_mean img /= img_std return img[np.newaxis, :]#——————————————————————模型配置、预测相关函数——————————————————————————def predict_config(model_file, params_file): # 根据预测部署的实际情况,设置Config config = Config() # 读取模型文件 config.set_prog_file(model_file) config.set_params_file(params_file) # Config默认是使用CPU预测,若要使用GPU预测,需要手动开启,设置运行的GPU卡号和分配的初始显存。 config.enable_use_gpu(500, 0) # 可以设置开启IR优化、开启内存优化。 config.switch_ir_optim() config.enable_memory_optim() predictor = create_predictor(config) return predictordef predict(image, predictor, target_size): img = preprocess(image, target_size) input_names = predictor.get_input_names() input_tensor = predictor.get_input_handle(input_names[0]) input_tensor.reshape(img.shape) input_tensor.copy_from_cpu(img.copy()) # 执行Predictor predictor.run() # 获取输出 output_names = predictor.get_output_names() output_tensor = predictor.get_output_handle(output_names[0]) output_data = output_tensor.copy_to_cpu() print("output_names", output_names) print("output_tensor", output_tensor) print("output_data", output_data) return output_dataif __name__ == '__main__': model_file = "model_dir/model.pdmodel" params_file = "model_dir/model.pdiparams" import random # image = cv2.imread("data/img/7419.jpg") # image = cv2.imread("data/img/8891.jpg") image = cv2.imread("data/img/{}.jpg".format(random.randint(0,5000))) predictor = predict_config(model_file, params_file) res = predict(image, predictor, target_size=224) # 进行图片的展示 plt.figure() plt.title(res) image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB) plt.imshow(image.astype('uint8')) plt.axis('on') plt.show()登录后复制
output_names ['relu_0.tmp_0']output_tensor登录后复制output_data [[0.03796072]]
--- Running analysis [ir_graph_build_pass]--- Running analysis [ir_graph_clean_pass]--- Running analysis [ir_analysis_pass]--- Running IR pass [is_test_pass]--- Running IR pass [simplify_with_basic_ops_pass]--- Running IR pass [conv_affine_channel_fuse_pass]--- Running IR pass [conv_eltwiseadd_affine_channel_fuse_pass]--- Running IR pass [conv_bn_fuse_pass]--- Running IR pass [conv_eltwiseadd_bn_fuse_pass]--- Running IR pass [embedding_eltwise_layernorm_fuse_pass]--- Running IR pass [multihead_matmul_fuse_pass_v2]--- Running IR pass [squeeze2_matmul_fuse_pass]--- Running IR pass [reshape2_matmul_fuse_pass]--- Running IR pass [flatten2_matmul_fuse_pass]--- Running IR pass [map_matmul_v2_to_mul_pass]I1117 16:08:14.154837 128 fuse_pass_base.cc:57] --- detected 2 subgraphs--- Running IR pass [map_matmul_v2_to_matmul_pass]--- Running IR pass [map_matmul_to_mul_pass]--- Running IR pass [fc_fuse_pass]I1117 16:08:14.155347 128 fuse_pass_base.cc:57] --- detected 2 subgraphs--- Running IR pass [fc_elementwise_layernorm_fuse_pass]--- Running IR pass [conv_elementwise_add_act_fuse_pass]--- Running IR pass [conv_elementwise_add2_act_fuse_pass]--- Running IR pass [conv_elementwise_add_fuse_pass]--- Running IR pass [transpose_flatten_concat_fuse_pass]--- Running IR pass [runtime_context_cache_pass]--- Running analysis [ir_params_sync_among_devices_pass]I1117 16:08:14.158038 128 ir_params_sync_among_devices_pass.cc:45] Sync params from CPU to GPU--- Running analysis [adjust_cudnn_workspace_size_pass]--- Running analysis [inference_op_replace_pass]--- Running analysis [memory_optimize_pass]I1117 16:08:14.159536 128 memory_optimize_pass.cc:214] Cluster name : x size: 38535168I1117 16:08:14.159554 128 memory_optimize_pass.cc:214] Cluster name : relu_0.tmp_0 size: 403734528I1117 16:08:14.159556 128 memory_optimize_pass.cc:214] Cluster name : relu_3.tmp_0 size: 2359296I1117 16:08:14.159559 128 memory_optimize_pass.cc:214] Cluster name : pool2d_0.tmp_0 size: 100933632--- Running analysis [ir_graph_to_program_pass]I1117 16:08:14.166565 128 analysis_predictor.cc:717] ======= optimize end =======I1117 16:08:14.166777 128 naive_executor.cc:98] --- skip [feed], feed -> xI1117 16:08:14.167583 128 naive_executor.cc:98] --- skip [relu_0.tmp_0], fetch -> fetch登录后复制
登录后复制
2021-11-05 11:52
手游攻略2021-11-19 18:38
手游攻略2021-10-31 23:18
手游攻略2022-06-03 14:46
游戏资讯2025-06-28 12:37
单机攻略