Skip to content
  • Our Work
    • Fields
      • Cardiology
      • ENT
      • Gastro
      • Orthopedics
      • Ophthalmology
      • Pulmonology
      • Surgical
      • Urology
      • Other
    • Modalities
      • Endoscopy
      • Medical Segmentation
      • Microscopy
      • Ultrasound
  • Success Stories
  • Insights
    • Magazine
    • Upcoming Events
    • Webinars
    • Meetups
    • News
    • Blog
  • The company
    • About us
    • Careers
Menu
  • Our Work
    • Fields
      • Cardiology
      • ENT
      • Gastro
      • Orthopedics
      • Ophthalmology
      • Pulmonology
      • Surgical
      • Urology
      • Other
    • Modalities
      • Endoscopy
      • Medical Segmentation
      • Microscopy
      • Ultrasound
  • Success Stories
  • Insights
    • Magazine
    • Upcoming Events
    • Webinars
    • Meetups
    • News
    • Blog
  • The company
    • About us
    • Careers
Contact

Object Tracking at High fps

Object tracking in video sequences is a classical challenge in computer vision, which finds applications in nearly all domains of the industry: from assembly line automation, security, traffic control, automatic driving assistance systems and agriculture. Presently state of the art algorithms performs relatively well in control environments, where illumination and camera angle remain relatively stable throughout acquisition and when object occlusion is at minimum. Complications pile-up rapidly when multiple objects appear in the scene, or when objects undergo non-rigid transformation from one frame to the next.

object tracking in video frames

Classically, object tracking algorithms start with a known object to track (contained within a bounding box); the algorithm then computes features in that bounding box, which are associated with the object to track, and continues to find the most probable bounding box in the next frame within which features are as close as possible to those appearing in the previous frame. The search algorithm advances by looks for bounding box in frame t+1 after applying an affine transformation of that in the previous frame t. Such an approach is highlighted in the algorithm of Lucas, Kanade and Tomasi.

Several drawbacks with the above classical approach are encountered. Firstly, bounding boxes of non-convex object might contain a large portion of the background, which might be complex and the features extracted from it might confuse matching for the next frame. Secondly, feature matching algorithms can easily lose track of the main object when occluded (e.g., hand passing over face). Finally, feature computation and matching for object undergoing small continuous motion forms a bottleneck for the speed at which objects can be tracked in real-time and can be as slow as 0.8-10 fps.

To resolve some of the difficulties of classic tracking algorithms, deep convolutional networks have been offered as a possible solution. These off-line trained networks provide means of overcoming occlusion, based on the knowledge of continuity learned in the training phase. In addition, the network can handle a variety of object types (convex and non-convex), which improves their overall usability over classical solutions.

The output of such trained network are the coordinates of a bounding box containing the object to track in each frame of the video sequence. A notable network architecture to perform tracking tasks contains an input layer with two nodes corresponding to two subsequent frames: one annotated and centered on the object to track, and the second for the frame within which the bounding box is to be localized. Frame information is passed to convolutional (filter) layers, where feature values are concatenated and further passed through three fully connected layers; finally, information is passed to four node outputs representing the four corners of the bounding box containing the image in the next frame.

Due to the nature of offline training of the network, object tracking can be performed on a surprising 100 fps on a modern GPU, and about 20 fps on a CPU. The construction and training of the CNN for tracking requires careful planning and execution. Our experts at RSIP Vision have been dealing for many years with state-of-the-art algorithms for object tracking in the most severe conditions in natural videos, such as low dose X-ray, cloudy road conditions and security cameras. We at RSIP Vision are committed to provide the highest quality tailor-made algorithms for our clients’ needs, with unparalleled accuracy and reproducibility. To learn more about RSIP Vision’s activities in a large variety of domains, please visit our project page. To learn how RSIP Vision can contribute to advancing your project, please contact us and consult our experts.

Share

Share on linkedin
Share on twitter
Share on facebook

Related Content

Prostate Guidance

Intra-op Prostate Guidance by RSIP Vision

Automated IBD Scoring

IBD Scoring – Clario, GI Reviewers and RSIP Vision Team Up

Soft Tissues Tracking during Brain Surgery

Soft Tissue Tracking during Brain Surgeries

Next Generation Intra-op Navigation

Neph - Partial Nephrectomy surgery

RSIP Neph Announces a Revolutionary Intra-op Solution for Partial Nephrectomy Surgeries

Benign Prostatic Hyperplasia BPH

AI for Benign Prostatic Hyperplasia BPH

Prostate Guidance

Intra-op Prostate Guidance by RSIP Vision

Automated IBD Scoring

IBD Scoring – Clario, GI Reviewers and RSIP Vision Team Up

Soft Tissues Tracking during Brain Surgery

Soft Tissue Tracking during Brain Surgeries

Next Generation Intra-op Navigation

Neph - Partial Nephrectomy surgery

RSIP Neph Announces a Revolutionary Intra-op Solution for Partial Nephrectomy Surgeries

Benign Prostatic Hyperplasia BPH

AI for Benign Prostatic Hyperplasia BPH

Show all

RSIP Vision

Field-tested software solutions and custom R&D, to power your next medical products with innovative AI and image analysis capabilities.

Read more about us

Get in touch

Please fill the following form and our experts will be happy to reply to you soon

Recent News

IBD Scoring – Clario, GI Reviewers and RSIP Vision Team Up

RSIP Neph Announces a Revolutionary Intra-op Solution for Partial Nephrectomy Surgeries

Announcement – RSIP Vision Presents Successful Preliminary Results from Clinical Study of 2D-to-3D Knee Bones Reconstruction

Announcement – New Urological AI Tool for 3D Reconstruction of the Ureter

All news
Upcoming Events
Stay informed for our next events
Subscribe to Our Magazines

Subscribe now and receive the Computer Vision News Magazine every month to your mailbox

 
Subscribe for free
Follow us
Linkedin Twitter Facebook Youtube

contact@rsipvision.com

Terms of Use

Privacy Policy

© All rights reserved to RSIP Vision 2023

Created by Shmulik

  • Our Work
    • title-1
      • Ophthalmology
      • Uncategorized
      • Ophthalmology
      • Pulmonology
      • Cardiology
      • Orthopedics
    • Title-2
      • Orthopedics
  • Success Stories
  • Insights
  • The company