当前位置: 首页 > news >正文

建设公司网站的原则wordpress转换成dede

建设公司网站的原则,wordpress转换成dede,禁止wordpress升级,做网站销售门窗怎么做Vins-Moon运行 求助#xff01;#xff01;#xff01;源码地址电脑配置环境配置编译Kitti数据集制作IMU时间戳问题 适配Kitti数据集运行结果Euroc数据集kitti数据集 evo评估#xff08;KITTI数据#xff09;输出轨迹(tum格式)结果 求助#xff01;#xff01;#xff… Vins-Moon运行 求助源码地址电脑配置环境配置编译Kitti数据集制作IMU时间戳问题 适配Kitti数据集运行结果Euroc数据集kitti数据集 evo评估KITTI数据输出轨迹(tum格式)结果 求助 求助 跑了vinsmoon在kitti_2011_10_03_0034上出现big translation,然后重启的问题。想请问下大家怎么理解这个问题然后从rviz上看重启之后是因为重启代码的问题导致轨迹再次从初始点开始了这个是不是需要修改代码解决有想讨论的朋友或者大佬可以私信我也可以私信加联系方式感谢 源码地址 源码链接https://github.com/HKUST-Aerial-Robotics/VINS-Mono.git 电脑配置 Ubuntu 18.04 ROS Melodic GTSAM 4.0.2 CERES 1.14.0 pcl1.8vtk8.2.0opencv3.2.0 环境配置 之前已经配置过LVI-SAM的环境所以没有什么额外需要配置的可参考之前的博客 编译 cd ~/catkin_ws/srcgit clone https://github.com/HKUST-Aerial-Robotics/VINS-Mono.gitcd ..catkin_make -j2注直接catkin_make会死机 Kitti数据集制作 参考链接 1.https://blog.csdn.net/GuanLingde/article/details/133938758?spm1001.2101.3001.6650.1utm_mediumdistribute.pc_relevant.none-task-blog-2%7Edefault%7ECTRLIST%7ERate-1-133938758-blog-127442772.235%5Ev43%5Econtroldepth_1-utm_sourcedistribute.pc_relevant.none-task-blog-2%7Edefault%7ECTRLIST%7ERate-1-133938758-blog-127442772.235%5Ev43%5Econtrolutm_relevant_index2 2.https://zhuanlan.zhihu.com/p/115562083 IMU时间戳问题 首先用*_extract/oxts文件夹把*_sync/oxt的文件夹替换掉 KITTI提供的的原始的IMU数据的时间戳存在断续和逆序的情况只能解决逆序情况断续问题无法解决, 通过下面的程序查看断续的和逆序的IMU时间戳并对逆序的IMU数据的时间戳进行手动修改 import datetime as dt import glob import os import matplotlib.pyplot as plt import numpy as npdata_path /home/nssc/sbk/code/slam/datasets/kitti2bag/modify/2011_10_03_drive_0034_sync def load_timestamps(dataoxts):Load timestamps from file.timestamp_file os.path.join(data_path, data, timestamps.txt)# Read and parse the timestampstimestamps []with open(timestamp_file, r) as f:for line in f.readlines():# NB: datetime only supports microseconds, but KITTI timestamps# give nanoseconds, so need to truncate last 4 characters to# get rid of \n (counts as 1) and extra 3 digitst dt.datetime.strptime(line[:-4], %Y-%m-%d %H:%M:%S.%f)t dt.datetime.timestamp(t)timestamps.append(t)# Subselect the chosen range of frames, if anyreturn timestamps timestamps np.array(load_timestamps()) x np.arange(0, len(timestamps))last_timestamp timestamps[:-1] curr_timestamp timestamps[1:] dt np.array(curr_timestamp - last_timestamp) #计算前后帧时间差print(dt 0.015: \n{}.format(dt[dt 0.015])) # 打印前后帧时间差大于0.015的IMU index dt dt.tolist() dt.append(0.01) dt np.array(dt) print(dt 0.015: \n{}.format(x[dt 0.015])) # 打印时间差大于0.015的具体时间差 plt.plot(x, timestamps, r, labelimu)# 可视化的时间戳 plt.show()打开timestamps.txt文件分别找到 380 4890 7833 32773 33734 34035 34539 38553行手动修改时间戳 如380行 依次按顺序把时间加0.01改完后 运行kitti2bag.py LIO-SAM中有config/doc/kitti2bag.py的工具文件 注如果手动替换了文件夹则注释掉这行代码 unsynced_path synced_path.replace(‘sync’, ‘extract’) python kitti2bag.py -t 2011_10_03 -r 0034 raw_synced .#!env python # -*- coding: utf-8 -*-import systry:import pykitti except ImportError as e:print(Could not load module \pykitti\. Please run pip install pykitti)sys.exit(1)import tf import os import cv2 import rospy import rosbag from tqdm import tqdm from tf2_msgs.msg import TFMessage from datetime import datetime from std_msgs.msg import Header from sensor_msgs.msg import CameraInfo, Imu, PointField, NavSatFix import sensor_msgs.point_cloud2 as pcl2 from geometry_msgs.msg import TransformStamped, TwistStamped, Transform from cv_bridge import CvBridge import numpy as np import argparsedef save_imu_data(bag, kitti, imu_frame_id, topic):print(Exporting IMU)for timestamp, oxts in zip(kitti.timestamps, kitti.oxts):q tf.transformations.quaternion_from_euler(oxts.packet.roll, oxts.packet.pitch, oxts.packet.yaw)imu Imu()imu.header.frame_id imu_frame_idimu.header.stamp rospy.Time.from_sec(float(timestamp.strftime(%s.%f)))imu.orientation.x q[0]imu.orientation.y q[1]imu.orientation.z q[2]imu.orientation.w q[3]imu.linear_acceleration.x oxts.packet.afimu.linear_acceleration.y oxts.packet.alimu.linear_acceleration.z oxts.packet.auimu.angular_velocity.x oxts.packet.wfimu.angular_velocity.y oxts.packet.wlimu.angular_velocity.z oxts.packet.wubag.write(topic, imu, timu.header.stamp)def save_imu_data_raw(bag, kitti, imu_frame_id, topic):print(Exporting IMU Raw)synced_path kitti.data_path# unsynced_path synced_path.replace(sync, extract)imu_path os.path.join(synced_path, oxts)# read time stamp (convert to ros seconds format)with open(os.path.join(imu_path, timestamps.txt)) as f:lines f.readlines()imu_datetimes []for line in lines:if len(line) 1:continuetimestamp datetime.strptime(line[:-4], %Y-%m-%d %H:%M:%S.%f)imu_datetimes.append(float(timestamp.strftime(%s.%f)))# fix imu time using a linear model (may not be ideal, ^_^)imu_index np.asarray(range(len(imu_datetimes)), dtypenp.float64)z np.polyfit(imu_index, imu_datetimes, 1)imu_datetimes_new z[0] * imu_index z[1]imu_datetimes imu_datetimes_new.tolist()# get all imu dataimu_data_dir os.path.join(imu_path, data)imu_filenames sorted(os.listdir(imu_data_dir))imu_data [None] * len(imu_filenames)for i, imu_file in enumerate(imu_filenames):imu_data_file open(os.path.join(imu_data_dir, imu_file), r)for line in imu_data_file:if len(line) 1:continuestripped_line line.strip()line_list stripped_line.split()imu_data[i] line_listassert len(imu_datetimes) len(imu_data)for timestamp, data in zip(imu_datetimes, imu_data):roll, pitch, yaw float(data[3]), float(data[4]), float(data[5]), q tf.transformations.quaternion_from_euler(roll, pitch, yaw)imu Imu()imu.header.frame_id imu_frame_idimu.header.stamp rospy.Time.from_sec(timestamp)imu.orientation.x q[0]imu.orientation.y q[1]imu.orientation.z q[2]imu.orientation.w q[3]imu.linear_acceleration.x float(data[11])imu.linear_acceleration.y float(data[12])imu.linear_acceleration.z float(data[13])imu.angular_velocity.x float(data[17])imu.angular_velocity.y float(data[18])imu.angular_velocity.z float(data[19])bag.write(topic, imu, timu.header.stamp)imu.header.frame_id imu_enu_linkbag.write(/imu_correct, imu, timu.header.stamp) # for LIO-SAM GPSdef save_dynamic_tf(bag, kitti, kitti_type, initial_time):print(Exporting time dependent transformations)if kitti_type.find(raw) ! -1:for timestamp, oxts in zip(kitti.timestamps, kitti.oxts):tf_oxts_msg TFMessage()tf_oxts_transform TransformStamped()tf_oxts_transform.header.stamp rospy.Time.from_sec(float(timestamp.strftime(%s.%f)))tf_oxts_transform.header.frame_id worldtf_oxts_transform.child_frame_id base_linktransform (oxts.T_w_imu)t transform[0:3, 3]q tf.transformations.quaternion_from_matrix(transform)oxts_tf Transform()oxts_tf.translation.x t[0]oxts_tf.translation.y t[1]oxts_tf.translation.z t[2]oxts_tf.rotation.x q[0]oxts_tf.rotation.y q[1]oxts_tf.rotation.z q[2]oxts_tf.rotation.w q[3]tf_oxts_transform.transform oxts_tftf_oxts_msg.transforms.append(tf_oxts_transform)bag.write(/tf, tf_oxts_msg, tf_oxts_msg.transforms[0].header.stamp)elif kitti_type.find(odom) ! -1:timestamps map(lambda x: initial_time x.total_seconds(), kitti.timestamps)for timestamp, tf_matrix in zip(timestamps, kitti.T_w_cam0):tf_msg TFMessage()tf_stamped TransformStamped()tf_stamped.header.stamp rospy.Time.from_sec(timestamp)tf_stamped.header.frame_id worldtf_stamped.child_frame_id camera_leftt tf_matrix[0:3, 3]q tf.transformations.quaternion_from_matrix(tf_matrix)transform Transform()transform.translation.x t[0]transform.translation.y t[1]transform.translation.z t[2]transform.rotation.x q[0]transform.rotation.y q[1]transform.rotation.z q[2]transform.rotation.w q[3]tf_stamped.transform transformtf_msg.transforms.append(tf_stamped)bag.write(/tf, tf_msg, tf_msg.transforms[0].header.stamp)def save_camera_data(bag, kitti_type, kitti, util, bridge, camera, camera_frame_id, topic, initial_time):print(Exporting camera {}.format(camera))if kitti_type.find(raw) ! -1:camera_pad {0:02d}.format(camera)image_dir os.path.join(kitti.data_path, image_{}.format(camera_pad))image_path os.path.join(image_dir, data)image_filenames sorted(os.listdir(image_path))with open(os.path.join(image_dir, timestamps.txt)) as f:image_datetimes map(lambda x: datetime.strptime(x[:-4], %Y-%m-%d %H:%M:%S.%f), f.readlines())calib CameraInfo()calib.header.frame_id camera_frame_idcalib.width, calib.height tuple(util[S_rect_{}.format(camera_pad)].tolist())calib.distortion_model plumb_bobcalib.K util[K_{}.format(camera_pad)]calib.R util[R_rect_{}.format(camera_pad)]calib.D util[D_{}.format(camera_pad)]calib.P util[P_rect_{}.format(camera_pad)]elif kitti_type.find(odom) ! -1:camera_pad {0:01d}.format(camera)image_path os.path.join(kitti.sequence_path, image_{}.format(camera_pad))image_filenames sorted(os.listdir(image_path))image_datetimes map(lambda x: initial_time x.total_seconds(), kitti.timestamps)calib CameraInfo()calib.header.frame_id camera_frame_idcalib.P util[P{}.format(camera_pad)]iterable zip(image_datetimes, image_filenames)for dt, filename in tqdm(iterable, totallen(image_filenames)):image_filename os.path.join(image_path, filename)cv_image cv2.imread(image_filename)calib.height, calib.width cv_image.shape[:2]if camera in (0, 1):cv_image cv2.cvtColor(cv_image, cv2.COLOR_BGR2GRAY)encoding mono8 if camera in (0, 1) else bgr8image_message bridge.cv2_to_imgmsg(cv_image, encodingencoding)image_message.header.frame_id camera_frame_idif kitti_type.find(raw) ! -1:image_message.header.stamp rospy.Time.from_sec(float(datetime.strftime(dt, %s.%f)))topic_ext /image_rawelif kitti_type.find(odom) ! -1:image_message.header.stamp rospy.Time.from_sec(dt)topic_ext /image_rectcalib.header.stamp image_message.header.stampbag.write(topic topic_ext, image_message, t image_message.header.stamp)bag.write(topic /camera_info, calib, t calib.header.stamp) def save_velo_data(bag, kitti, velo_frame_id, topic):print(Exporting velodyne data)velo_path os.path.join(kitti.data_path, velodyne_points)velo_data_dir os.path.join(velo_path, data)velo_filenames sorted(os.listdir(velo_data_dir))with open(os.path.join(velo_path, timestamps.txt)) as f:lines f.readlines()velo_datetimes []for line in lines:if len(line) 1:continuedt datetime.strptime(line[:-4], %Y-%m-%d %H:%M:%S.%f)velo_datetimes.append(dt)iterable zip(velo_datetimes, velo_filenames)count 0for dt, filename in tqdm(iterable, totallen(velo_filenames)):if dt is None:continuevelo_filename os.path.join(velo_data_dir, filename)# read binary datascan (np.fromfile(velo_filename, dtypenp.float32)).reshape(-1, 4)# get ring channeldepth np.linalg.norm(scan, 2, axis1)pitch np.arcsin(scan[:, 2] / depth) # arcsin(z, depth)fov_down -24.8 / 180.0 * np.pifov (abs(-24.8) abs(2.0)) / 180.0 * np.piproj_y (pitch abs(fov_down)) / fov # in [0.0, 1.0]proj_y * 64 # in [0.0, H]proj_y np.floor(proj_y)proj_y np.minimum(64 - 1, proj_y)proj_y np.maximum(0, proj_y).astype(np.int32) # in [0,H-1]proj_y proj_y.reshape(-1, 1)scan np.concatenate((scan,proj_y), axis1)scan scan.tolist()for i in range(len(scan)):scan[i][-1] int(scan[i][-1])# create headerheader Header()header.frame_id velo_frame_idheader.stamp rospy.Time.from_sec(float(datetime.strftime(dt, %s.%f)))# fill pcl msgfields [PointField(x, 0, PointField.FLOAT32, 1),PointField(y, 4, PointField.FLOAT32, 1),PointField(z, 8, PointField.FLOAT32, 1),PointField(intensity, 12, PointField.FLOAT32, 1),PointField(ring, 16, PointField.UINT16, 1)]pcl_msg pcl2.create_cloud(header, fields, scan)pcl_msg.is_dense True# print(pcl_msg)bag.write(topic, pcl_msg, tpcl_msg.header.stamp)# count 1# if count 200:# breakdef get_static_transform(from_frame_id, to_frame_id, transform):t transform[0:3, 3]q tf.transformations.quaternion_from_matrix(transform)tf_msg TransformStamped()tf_msg.header.frame_id from_frame_idtf_msg.child_frame_id to_frame_idtf_msg.transform.translation.x float(t[0])tf_msg.transform.translation.y float(t[1])tf_msg.transform.translation.z float(t[2])tf_msg.transform.rotation.x float(q[0])tf_msg.transform.rotation.y float(q[1])tf_msg.transform.rotation.z float(q[2])tf_msg.transform.rotation.w float(q[3])return tf_msgdef inv(transform):Invert rigid body transformation matrixR transform[0:3, 0:3]t transform[0:3, 3]t_inv -1 * R.T.dot(t)transform_inv np.eye(4)transform_inv[0:3, 0:3] R.Ttransform_inv[0:3, 3] t_invreturn transform_invdef save_static_transforms(bag, transforms, timestamps):print(Exporting static transformations)tfm TFMessage()for transform in transforms:t get_static_transform(from_frame_idtransform[0], to_frame_idtransform[1], transformtransform[2])tfm.transforms.append(t)for timestamp in timestamps:time rospy.Time.from_sec(float(timestamp.strftime(%s.%f)))for i in range(len(tfm.transforms)):tfm.transforms[i].header.stamp timebag.write(/tf_static, tfm, ttime)def save_gps_fix_data(bag, kitti, gps_frame_id, topic):for timestamp, oxts in zip(kitti.timestamps, kitti.oxts):navsatfix_msg NavSatFix()navsatfix_msg.header.frame_id gps_frame_idnavsatfix_msg.header.stamp rospy.Time.from_sec(float(timestamp.strftime(%s.%f)))navsatfix_msg.latitude oxts.packet.latnavsatfix_msg.longitude oxts.packet.lonnavsatfix_msg.altitude oxts.packet.altnavsatfix_msg.status.service 1bag.write(topic, navsatfix_msg, tnavsatfix_msg.header.stamp)def save_gps_vel_data(bag, kitti, gps_frame_id, topic):for timestamp, oxts in zip(kitti.timestamps, kitti.oxts):twist_msg TwistStamped()twist_msg.header.frame_id gps_frame_idtwist_msg.header.stamp rospy.Time.from_sec(float(timestamp.strftime(%s.%f)))twist_msg.twist.linear.x oxts.packet.vftwist_msg.twist.linear.y oxts.packet.vltwist_msg.twist.linear.z oxts.packet.vutwist_msg.twist.angular.x oxts.packet.wftwist_msg.twist.angular.y oxts.packet.wltwist_msg.twist.angular.z oxts.packet.wubag.write(topic, twist_msg, ttwist_msg.header.stamp)if __name__ __main__:parser argparse.ArgumentParser(description Convert KITTI dataset to ROS bag file the easy way!)# Accepted argument valueskitti_types [raw_synced, odom_color, odom_gray]odometry_sequences []for s in range(22):odometry_sequences.append(str(s).zfill(2))parser.add_argument(kitti_type, choices kitti_types, help KITTI dataset type)parser.add_argument(dir, nargs ?, default os.getcwd(), help base directory of the dataset, if no directory passed the deafult is current working directory)parser.add_argument(-t, --date, help date of the raw dataset (i.e. 2011_09_26), option is only for RAW datasets.)parser.add_argument(-r, --drive, help drive number of the raw dataset (i.e. 0001), option is only for RAW datasets.)parser.add_argument(-s, --sequence, choices odometry_sequences,help sequence of the odometry dataset (between 00 - 21), option is only for ODOMETRY datasets.)args parser.parse_args()bridge CvBridge()compression rosbag.Compression.NONE# compression rosbag.Compression.BZ2# compression rosbag.Compression.LZ4# CAMERAScameras [(0, camera_gray_left, /kitti/camera_gray_left),(1, camera_gray_right, /kitti/camera_gray_right),(2, camera_color_left, /kitti/camera_color_left),(3, camera_color_right, /kitti/camera_color_right)]if args.kitti_type.find(raw) ! -1:if args.date None:print(Date option is not given. It is mandatory for raw dataset.)print(Usage for raw dataset: kitti2bag raw_synced [dir] -t date -r drive)sys.exit(1)elif args.drive None:print(Drive option is not given. It is mandatory for raw dataset.)print(Usage for raw dataset: kitti2bag raw_synced [dir] -t date -r drive)sys.exit(1)bag rosbag.Bag(kitti_{}_drive_{}_{}.bag.format(args.date, args.drive, args.kitti_type[4:]), w, compressioncompression)kitti pykitti.raw(args.dir, args.date, args.drive)if not os.path.exists(kitti.data_path):print(Path {} does not exists. Exiting..format(kitti.data_path))sys.exit(1)if len(kitti.timestamps) 0:print(Dataset is empty? Exiting.)sys.exit(1)try:# IMUimu_frame_id imu_linkimu_topic /kitti/oxts/imuimu_raw_topic /imu_rawgps_fix_topic /gps/fixgps_vel_topic /gps/velvelo_frame_id velodynevelo_topic /points_rawT_base_link_to_imu np.eye(4, 4)T_base_link_to_imu[0:3, 3] [-2.71/2.0-0.05, 0.32, 0.93]# tf_statictransforms [(base_link, imu_frame_id, T_base_link_to_imu),(imu_frame_id, velo_frame_id, inv(kitti.calib.T_velo_imu)),(imu_frame_id, cameras[0][1], inv(kitti.calib.T_cam0_imu)),(imu_frame_id, cameras[1][1], inv(kitti.calib.T_cam1_imu)),(imu_frame_id, cameras[2][1], inv(kitti.calib.T_cam2_imu)),(imu_frame_id, cameras[3][1], inv(kitti.calib.T_cam3_imu))]util pykitti.utils.read_calib_file(os.path.join(kitti.calib_path, calib_cam_to_cam.txt))# Export# save_static_transforms(bag, transforms, kitti.timestamps)# save_dynamic_tf(bag, kitti, args.kitti_type, initial_timeNone)# save_imu_data(bag, kitti, imu_frame_id, imu_topic)save_imu_data_raw(bag, kitti, imu_frame_id, imu_raw_topic)save_gps_fix_data(bag, kitti, imu_frame_id, gps_fix_topic)save_gps_vel_data(bag, kitti, imu_frame_id, gps_vel_topic)for camera in cameras:save_camera_data(bag, args.kitti_type, kitti, util, bridge, cameracamera[0], camera_frame_idcamera[1], topiccamera[2], initial_timeNone)breaksave_velo_data(bag, kitti, velo_frame_id, velo_topic)finally:print(## OVERVIEW ##)print(bag)bag.close()elif args.kitti_type.find(odom) ! -1:if args.sequence None:print(Sequence option is not given. It is mandatory for odometry dataset.)print(Usage for odometry dataset: kitti2bag {odom_color, odom_gray} [dir] -s sequence)sys.exit(1)bag rosbag.Bag(kitti_data_odometry_{}_sequence_{}.bag.format(args.kitti_type[5:],args.sequence), w, compressioncompression)kitti pykitti.odometry(args.dir, args.sequence)if not os.path.exists(kitti.sequence_path):print(Path {} does not exists. Exiting..format(kitti.sequence_path))sys.exit(1)kitti.load_calib() kitti.load_timestamps() if len(kitti.timestamps) 0:print(Dataset is empty? Exiting.)sys.exit(1)if args.sequence in odometry_sequences[:11]:print(Odometry dataset sequence {} has ground truth information (poses)..format(args.sequence))kitti.load_poses()try:util pykitti.utils.read_calib_file(os.path.join(args.dir,sequences,args.sequence, calib.txt))current_epoch (datetime.utcnow() - datetime(1970, 1, 1)).total_seconds()# Exportif args.kitti_type.find(gray) ! -1:used_cameras cameras[:2]elif args.kitti_type.find(color) ! -1:used_cameras cameras[-2:]save_dynamic_tf(bag, kitti, args.kitti_type, initial_timecurrent_epoch)for camera in used_cameras:save_camera_data(bag, args.kitti_type, kitti, util, bridge, cameracamera[0], camera_frame_idcamera[1], topiccamera[2], initial_timecurrent_epoch)finally:print(## OVERVIEW ##)print(bag)bag.close() 适配Kitti数据集 在config文件夹下新建kitti文件夹 新建kitti_config.yaml文件 (具体参数设置的方式可以参考之前LVI-SAM博客) %YAML:1.0#common parameters imu_topic: /imu_raw #/kitti/oxts/imu image_topic: /kitti/camera_gray_left/image_raw output_path: /home/nssc/sbk/outputs/map/vinsmoon/#camera calibration model_type: PINHOLE camera_name: camera0 #10_03 # image_width: 1241 # image_height: 376 # 09_30 image_width: 1226 image_height: 370 distortion_parameters:k1: 0.0k2: 0.0p1: 0.0p2: 0.0 projection_parameters: # 10_03 # fx: 7.188560e02 # fy: 7.188560e02 # cx: 6.071928e02 # cy: 1.852157e02# 09_30fx: 7.070912e02fy: 7.070912e02cx: 6.018873e02cy: 1.831104e02# Extrinsic parameter between IMU and Camera. estimate_extrinsic: 0 # 0 Have an accurate extrinsic parameters. We will trust the following imu^R_cam, imu^T_cam, dont change it.# 1 Have an initial guess about extrinsic parameters. We will optimize around your initial guess.# 2 Dont know anything about extrinsic parameters. You dont need to give R,T. We will try to calibrate it. Do some rotation movement at beginning. #If you choose 0 or 1, you should write down the following matrix. #Rotation from camera frame to imu frame, imu^R_cam extrinsicRotation: !!opencv-matrixrows: 3cols: 3dt: d# 10_03# data: [0.00875116, -0.00479609, 0.99995027, -0.99986428, -0.01400249, 0.00868325, 0.01396015, -0.99989044, -0.00491798]# 09_30data: [0.00781298, -0.0042792, 0.99996033,-0.99985947, -0.01486805, 0.00774856, 0.0148343 , -0.99988023, -0.00439476] #Translation from camera frame to imu frame, imu^T_cam extrinsicTranslation: !!opencv-matrixrows: 3cols: 1dt: d# 10_03# data: [1.10224312,-0.31907194, 0.74606588] #09_30 data: [1.14389871,-0.31271847, 0.72654605]#feature traker paprameters max_cnt: 150 # max feature number in feature tracking min_dist: 30 # min distance between two features freq: 10 # frequence (Hz) of publish tracking result. At least 10Hz for good estimation. If set 0, the frequence will be same as raw image F_threshold: 1.0 # ransac threshold (pixel) show_track: 1 # publish tracking image as topic equalize: 1 # if image is too dark or light, trun on equalize to find enough features fisheye: 0 # if using fisheye, trun on it. A circle mask will be loaded to remove edge noisy points#optimization parameters max_solver_time: 0.04 # max solver itration time (ms), to guarantee real time max_num_iterations: 8 # max solver itrations, to guarantee real time keyframe_parallax: 10.0 # keyframe selection threshold (pixel)#imu parameters The more accurate parameters you provide, the better performance acc_n: 0.08 # accelerometer measurement noise standard deviation. #0.2 0.04 gyr_n: 0.004 # gyroscope measurement noise standard deviation. #0.05 0.004 acc_w: 0.00004 # accelerometer bias random work noise standard deviation. #0.02 gyr_w: 2.0e-6 # gyroscope bias random work noise standard deviation. #4.0e-5 g_norm: 9.81007 # gravity magnitude#loop closure parameters loop_closure: 1 # start loop closure load_previous_pose_graph: 0 # load and reuse previous pose graph; load from pose_graph_save_path fast_relocalization: 0 # useful in real-time and large project pose_graph_save_path: /home/nssc/sbk/outputs/map/vinsmoon/pose_graph/ # save and load path#unsynchronization parameters estimate_td: 0 # online estimate time offset between camera and imu td: 0.0 # initial value of time offset. unit: s. readed image clock td real image clock (IMU clock)#rolling shutter parameters rolling_shutter: 0 # 0: global shutter camera, 1: rolling shutter camera rolling_shutter_tr: 0 # unit: s. rolling shutter read out time per frame (from data sheet). #visualization parameters save_image: 1 # save image in pose graph for visualization prupose; you can close this function by setting 0 visualize_imu_forward: 0 # output imu forward propogation to achieve low latency and high frequence results visualize_camera_size: 0.4 # size of camera marker in RVIZ 在vins_estimator/launch/文件夹下新建文件kitti.launch 主要修改一下config_path的路径 launcharg nameconfig_path default $(find feature_tracker)/../config/kitti/kitti_config.yaml /arg namevins_path default $(find feature_tracker)/../config/../ /node namefeature_tracker pkgfeature_tracker typefeature_tracker outputlogparam nameconfig_file typestring value$(arg config_path) /param namevins_folder typestring value$(arg vins_path) //nodenode namevins_estimator pkgvins_estimator typevins_estimator outputscreenparam nameconfig_file typestring value$(arg config_path) /param namevins_folder typestring value$(arg vins_path) //nodenode namepose_graph pkgpose_graph typepose_graph outputscreenparam nameconfig_file typestring value$(arg config_path) /param namevisualization_shift_x typeint value0 /param namevisualization_shift_y typeint value0 /param nameskip_cnt typeint value0 /param nameskip_dis typedouble value0 //node/launch 运行结果 Euroc数据集 roslaunch vins_estimator euroc.launch roslaunch vins_estimator vins_rviz.launchrosbag play YOUR_PATH_TO_DATASET/MH_01_easy.bag 同时看到groundtrue: roslaunch benchmark_publisher publish.launch sequence_name:MH_01_easykitti数据集 有关kitti数据集生成bag包的方式可参考之前生成LVI-SAM适配数据的博客 roslaunch vins_estimator kitti.launch roslaunch vins_estimator vins_rviz.launchrosbag play kitti_2011_09_30_drive_0027_synced.bagevo评估KITTI数据 输出轨迹(tum格式) vins_estimator/src/utility/visualization.cpp pubOdometry()函数150行 // write result to file// ofstream foutC(VINS_RESULT_PATH, ios::app);// foutC.setf(ios::fixed, ios::floatfield);// foutC.precision(0);// foutC header.stamp.toSec() * 1e9 ,;// foutC.precision(5);// foutC estimator.Ps[WINDOW_SIZE].x() ,// estimator.Ps[WINDOW_SIZE].y() ,// estimator.Ps[WINDOW_SIZE].z() ,// tmp_Q.w() ,// tmp_Q.x() ,// tmp_Q.y() ,// tmp_Q.z() ,// estimator.Vs[WINDOW_SIZE].x() ,// estimator.Vs[WINDOW_SIZE].y() ,// estimator.Vs[WINDOW_SIZE].z() , endl;ofstream foutC(VINS_RESULT_PATH, ios::app);foutC.setf(ios::fixed, ios::floatfield);foutC.precision(9);foutC header.stamp.toSec() ;foutC.precision(5);foutC estimator.Ps[WINDOW_SIZE].x() estimator.Ps[WINDOW_SIZE].y() estimator.Ps[WINDOW_SIZE].z() tmp_Q.x() tmp_Q.y() tmp_Q.z() tmp_Q.w() endl;foutC.close();pose_graph/src/pose_graph.cpp addKeyFrame()函数150行 if (SAVE_LOOP_PATH){// ofstream loop_path_file(VINS_RESULT_PATH, ios::app);// loop_path_file.setf(ios::fixed, ios::floatfield);// loop_path_file.precision(0);// loop_path_file cur_kf-time_stamp * 1e9 ,;// loop_path_file.precision(5);// loop_path_file P.x() ,// P.y() ,// P.z() ,// Q.w() ,// Q.x() ,// Q.y() ,// Q.z() ,// endl;ofstream loop_path_file(VINS_RESULT_PATH, ios::app);loop_path_file.setf(ios::fixed, ios::floatfield);loop_path_file.precision(0);loop_path_file cur_kf-time_stamp ;loop_path_file.precision(5);loop_path_file P.x() P.y() P.z() Q.x() Q.y() Q.z() Q.w() endl; loop_path_file.close();}updatePath()函数600行 if (SAVE_LOOP_PATH){// ofstream loop_path_file(VINS_RESULT_PATH, ios::app);// loop_path_file.setf(ios::fixed, ios::floatfield);// loop_path_file.precision(0);// loop_path_file (*it)-time_stamp * 1e9 ,;// loop_path_file.precision(5);// loop_path_file P.x() ,// P.y() ,// P.z() ,// Q.w() ,// Q.x() ,// Q.y() ,// Q.z() ,// endl;ofstream loop_path_file(VINS_RESULT_PATH, ios::app);loop_path_file.setf(ios::fixed, ios::floatfield);loop_path_file.precision(0);loop_path_file (*it)-time_stamp ;loop_path_file.precision(5);loop_path_file P.x() P.y() P.z() Q.x() Q.y() Q.z() Q.w() endl;loop_path_file.close();}pose_graph_node.cpp中的main()函数 # VINS_RESULT_PATH VINS_RESULT_PATH /vins_result_loop.csv;VINS_RESULT_PATH VINS_RESULT_PATH /vins_result_loop.txt;对输出的vins_result_loop.txt文件修改时间戳 # 读取txt文件 with open(vins_result_loop.txt, r) as file:lines file.readlines()# 处理数据 first_line lines[0].strip().split() first_num int(first_line[0]) output_lines [] for line in lines[0:]:parts line.split()new_num float(parts[0]) - first_numnew_line str(new_num) .join(parts[1:]) \noutput_lines.append(new_line)# 写入txt文件 with open(output.txt, w) as file:for line in output_lines:file.write(.join(line)) 结果 evo_traj tum output.txt 07_gt_tum.txt --ref07_gt_tum.txt -a -p --plot_modexyz求助 跑了vinsmoon在kitti_2011_10_03_0034上出现big translation,然后重启的问题。想请问下大家怎么理解这个问题然后从rviz上看重启之后是因为重启代码的问题导致轨迹再次从初始点开始了这个是不是需要修改代码解决有想讨论的朋友或者大佬可以私信我也可以私信加联系方式感谢 参考链接 https://blog.csdn.net/m0_49066914/article/details/131814856 https://blog.csdn.net/Hanghang_/article/details/104535370 https://zhuanlan.zhihu.com/p/75672946
http://www.hkea.cn/news/14457058/

相关文章:

  • 网站建设 英文版qq群文件wordpress
  • 物流企业网站建设方案咖啡豆网站模板
  • 网站建设进度时间表湖州网站设计建设
  • 美橙互联网站模板猪场宣传网站怎么建设
  • 个人可以做自媒体网站吗成都住建局官网住建蓉e办
  • 域名和网站建设中医院网站模板
  • 郑州春蕾网站建设淄博公司网站建设
  • 企业网站优化服务公司网页设计模板图片四张
  • 邮箱注册网站查询合肥做网站需要多少钱
  • 用phpnow搭建网站的整个流程网易企业邮箱收费和免费什么区别
  • 一个网站怎么做软件好用吗wordpress 微信连接数据库文件
  • 做网站需要多少费用wordpress更改网页标题
  • 个人两字印章在线制作网站广州培训机构网页设计
  • 网站建设推广选哪家免费网站看完你会回来感谢我的
  • 广西学校网站建设建立个人博客wordpress
  • 用手机可以建设一个手机网站吗安庆网站建设电话
  • php网站源码大全制造业人才网
  • 九江网站建设排行榜为网站做IPhone客户端
  • ionic3 做网站wordpress制作页面模板
  • 新网站开发费用建设一个网站可以放视频的多少钱
  • 网站如何导流量网站建设与管理就业前景
  • 网站的结构与布局优化设计服务网站建设方案
  • 在华图做网站编辑seo网站优化方案案例
  • 唐山网站建设互众动力怎么创建网页链接文件
  • 大淘客网站logo怎么做优化网站建设哪家专业
  • 图书馆门户网站建设方案张家界seo优化首选
  • 怎么和其它网站做友情链接义乌网站建设电话
  • 网站制作新手太原seo网站管理
  • 电子商务网站网络推广方式下载百度导航最新版本
  • 高端网站建设公司成都河北住房和城乡建设厅官方网站