{"id":353,"date":"2021-07-05T22:17:55","date_gmt":"2021-07-06T02:17:55","guid":{"rendered":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/?p=353"},"modified":"2021-07-27T16:00:30","modified_gmt":"2021-07-27T20:00:30","slug":"autonomous-car-navigation-using-visual-inertial-sensor-fusion","status":"publish","type":"post","link":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/","title":{"rendered":"Autonomous Car Navigation using Visual\/Inertial Sensor Fusion"},"content":{"rendered":"<p>To enable autonomous car navigation in GPS-denied environments, this project developed a multi-state Kalman filter that fuses visual features and inertial sensors in a tightly-coupled integration mode. The fusion of visual and inertial modalities enjoys growing popularity due to miniaturization and affordability of the sensors which make them ideal backups to the Global Navigation Satellite Systems (GNSS) where the service is interrupted in urban areas.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>To enable autonomous car navigation in GPS-denied environments, this project developed a multi-state Kalman filter that fuses visual features and inertial sensors in a tightly-coupled integration mode. The fusion of visual and inertial modalities enjoys growing popularity due to miniaturization and affordability of the sensors which make them ideal backups to the Global Navigation Satellite [&hellip;]<\/p>\n","protected":false},"author":12,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_relevanssi_hide_post":"","_relevanssi_hide_content":"","_relevanssi_pin_for_all":"","_relevanssi_pin_keywords":"","_relevanssi_unpin_keywords":"","_relevanssi_related_keywords":"","_relevanssi_related_include_ids":"","_relevanssi_related_exclude_ids":"","_relevanssi_related_no_append":"","_relevanssi_related_not_related":"","_relevanssi_related_posts":"","_relevanssi_noindex_reason":"","_mi_skip_tracking":false,"_exactmetrics_sitenote_active":false,"_exactmetrics_sitenote_note":"","_exactmetrics_sitenote_category":0,"footnotes":"","_links_to":"","_links_to_target":""},"categories":[24],"tags":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v21.2 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Autonomous Car Navigation using Visual\/Inertial Sensor Fusion - Embedded and Multi-sensor Systems Lab (EMSLab)<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Autonomous Car Navigation using Visual\/Inertial Sensor Fusion - Embedded and Multi-sensor Systems Lab (EMSLab)\" \/>\n<meta property=\"og:description\" content=\"To enable autonomous car navigation in GPS-denied environments, this project developed a multi-state Kalman filter that fuses visual features and inertial sensors in a tightly-coupled integration mode. The fusion of visual and inertial modalities enjoys growing popularity due to miniaturization and affordability of the sensors which make them ideal backups to the Global Navigation Satellite [&hellip;]\" \/>\n<meta property=\"og:url\" content=\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/\" \/>\n<meta property=\"og:site_name\" content=\"Embedded and Multi-sensor Systems Lab (EMSLab)\" \/>\n<meta property=\"article:published_time\" content=\"2021-07-06T02:17:55+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2021-07-27T20:00:30+00:00\" \/>\n<meta name=\"author\" content=\"changqiu\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"changqiu\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/\",\"url\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/\",\"name\":\"Autonomous Car Navigation using Visual\/Inertial Sensor Fusion - Embedded and Multi-sensor Systems Lab (EMSLab)\",\"isPartOf\":{\"@id\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#website\"},\"datePublished\":\"2021-07-06T02:17:55+00:00\",\"dateModified\":\"2021-07-27T20:00:30+00:00\",\"author\":{\"@id\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#\/schema\/person\/ebc5f46cab0b5b353237723cab5111e4\"},\"breadcrumb\":{\"@id\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Autonomous Car Navigation using Visual\/Inertial Sensor Fusion\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#website\",\"url\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/\",\"name\":\"Embedded and Multi-sensor Systems Lab (EMSLab)\",\"description\":\"Carleton University\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#\/schema\/person\/ebc5f46cab0b5b353237723cab5111e4\",\"name\":\"changqiu\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/9211d856b3181fb6ef752ba4579e3b71?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/9211d856b3181fb6ef752ba4579e3b71?s=96&d=mm&r=g\",\"caption\":\"changqiu\"},\"url\":\"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/author\/changqiu\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Autonomous Car Navigation using Visual\/Inertial Sensor Fusion - Embedded and Multi-sensor Systems Lab (EMSLab)","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/","og_locale":"en_US","og_type":"article","og_title":"Autonomous Car Navigation using Visual\/Inertial Sensor Fusion - Embedded and Multi-sensor Systems Lab (EMSLab)","og_description":"To enable autonomous car navigation in GPS-denied environments, this project developed a multi-state Kalman filter that fuses visual features and inertial sensors in a tightly-coupled integration mode. The fusion of visual and inertial modalities enjoys growing popularity due to miniaturization and affordability of the sensors which make them ideal backups to the Global Navigation Satellite [&hellip;]","og_url":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/","og_site_name":"Embedded and Multi-sensor Systems Lab (EMSLab)","article_published_time":"2021-07-06T02:17:55+00:00","article_modified_time":"2021-07-27T20:00:30+00:00","author":"changqiu","twitter_card":"summary_large_image","twitter_misc":{"Written by":"changqiu"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/","url":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/","name":"Autonomous Car Navigation using Visual\/Inertial Sensor Fusion - Embedded and Multi-sensor Systems Lab (EMSLab)","isPartOf":{"@id":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#website"},"datePublished":"2021-07-06T02:17:55+00:00","dateModified":"2021-07-27T20:00:30+00:00","author":{"@id":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#\/schema\/person\/ebc5f46cab0b5b353237723cab5111e4"},"breadcrumb":{"@id":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/2021\/autonomous-car-navigation-using-visual-inertial-sensor-fusion\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/"},{"@type":"ListItem","position":2,"name":"Autonomous Car Navigation using Visual\/Inertial Sensor Fusion"}]},{"@type":"WebSite","@id":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#website","url":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/","name":"Embedded and Multi-sensor Systems Lab (EMSLab)","description":"Carleton University","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#\/schema\/person\/ebc5f46cab0b5b353237723cab5111e4","name":"changqiu","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/9211d856b3181fb6ef752ba4579e3b71?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/9211d856b3181fb6ef752ba4579e3b71?s=96&d=mm&r=g","caption":"changqiu"},"url":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/author\/changqiu\/"}]}},"acf":{"Post Thumbnail Icon":"","Post Thumbnail":{"ID":437,"id":437,"title":"Screen Shot 2021-07-15 at 12.43.03 AM","filename":"Screen-Shot-2021-07-15-at-12.43.03-AM.png","filesize":119413,"url":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM.png","link":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/cu-home\/homepage\/screen-shot-2021-07-15-at-12-43-03-am\/","alt":"","author":"12","description":"","caption":"","name":"screen-shot-2021-07-15-at-12-43-03-am","status":"inherit","uploaded_to":14,"date":"2021-07-15 04:43:18","modified":"2021-09-26 12:07:54","menu_order":0,"mime_type":"image\/png","type":"image","subtype":"png","icon":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp\/wp-includes\/images\/media\/default.png","width":519,"height":271,"sizes":{"thumbnail":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM-160x84.png","thumbnail-width":160,"thumbnail-height":84,"medium":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM-240x125.png","medium-width":240,"medium-height":125,"medium_large":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM.png","medium_large-width":519,"medium_large-height":271,"large":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM-400x209.png","large-width":400,"large-height":209,"gallery-thumb":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM-300x230.png","gallery-thumb-width":300,"gallery-thumb-height":230,"1536x1536":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM.png","1536x1536-width":519,"1536x1536-height":271,"2048x2048":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM.png","2048x2048-width":519,"2048x2048-height":271,"banner":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM.png","banner-width":519,"banner-height":271,"people":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM-200x200.png","people-width":200,"people-height":200,"post-thumb":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM-300x230.png","post-thumb-width":300,"post-thumb-height":230,"rotator-image":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM.png","rotator-image-width":519,"rotator-image-height":271,"video-thumb":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-content\/uploads\/Screen-Shot-2021-07-15-at-12.43.03-AM-360x188.png","video-thumb-width":360,"video-thumb-height":188}}},"_links":{"self":[{"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/posts\/353"}],"collection":[{"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/users\/12"}],"replies":[{"embeddable":true,"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/comments?post=353"}],"version-history":[{"count":3,"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/posts\/353\/revisions"}],"predecessor-version":[{"id":517,"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/posts\/353\/revisions\/517"}],"wp:attachment":[{"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/media?parent=353"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/categories?post=353"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/carleton.ca\/embedded-and-multisensor-systems\/wp-json\/wp\/v2\/tags?post=353"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}