What are the factors in 3D laser triangulation line rates?

When designing an application, one likes to read the specifications to determine whether a candidate solution will satisfy the applications requirements. Let’s say you want to design an application to do laser profiling of your continuously moving target(s). You know Teledyne DALSA is well-regarded for their Z-Trak 3D Laser Profiler. In the specifications you may see that up to 3.3K second are achievable, but what factors could influence the rate?

What factors affect the line rate?

When choosing a pickup truck or SUV, cubic displacement and horsepower matter. But so do whether you plan to tow a trailer of a certain weight. And whether the terrain is hilly or flat.

With an area scan camera, maximum framerate is expressed for reading out all pixels when operating at full resolution. Faster rates can be achieved by reading out partial rows with a reduced area of interest. One must match camera and interface capabilities to application requirements.

Laser triangulation is an effective 3D technique

Here too one must read the specifications – and think about application requirements.

Figure 1: Key laser profiler terms and concepts in relation to each other – Courtesy Teledyne DALSA

What considerations affect 3D triangulation laser profilers?

Data volume: With reference to Figure 2 below, the number of pixels per row (X) and the frequency of scans in the Y dimension, together with the number of Bytes expressed per pixel, determine the data volume. Ultimately you need what you need, and may purchase a line scanner with a wider or smaller field of view, or a faster or slower interface, or a more intense laser light, accordingly. Required resolution has a bearing on data volumes, too, and that’s the key consideration we’ll go into further below.

Figure 2: Each laser profile scan delivers X pixels’ Z values to build Y essentially continuous slices – Courtesy Teledyne DALSA

Resolution has a bearing on data volumes and application performance

Presumably it’s clear that application performance will require certain precision in resolution. In the Y dimension, how frequently do you need each successive data slice in order to track feature changes over time? In the Z dimension, how fine grained do you need to know of changes in object height? And in the X dimension, how many points must be captured at what resolution?

While you might be prepared to negotiate resolution tolerances as an engineering tradeoff on performance or cost or risk, generally speaking you’ve got certain resolutions you are aiming for if the technology and budget can achieve it.

We’re warming up to the key point of this article – how line rate varies according to application features. Consider Figure 3 below, noting the trapezoidal shape for 3 respective fields of view, in correlation with working distance.

Figure 3: Working distance in which Z dimension may vary also impacts resolution achievable for each value in the X dimension – Courtesy Teledyne DALSA.

Trapezoid bottom width and required X dimension resolution

To drive this final point home, consider both Figure 2 and Figure 3. Figure 2, among other things, reminds us that we need to capture each successive scan from the Y dimension at precisely timed intervals. Otherwise how would we usefully track the changes in height in the Z dimension as the target moves down the conveyance?

That means that regardless of target height, each scan must always take exactly the same time as each other scan – it cannot vary. But per Figure 3, regardless of whether using a short, medium, or longer working distance, X pixels correlating to target values found high up in the trapezoidal FoV will yield a de facto higher resolution than the same X pixels lower down.

Suppose the top of the trapezoid is 50cm wide, and the bottom of the trapezoid is 100cm wide. For any given short span along a line in the X dimension, the real-space mapped into a sensor pixel will be 2x and long for targets sampled at the bottom of the FoV.

Since the required minimum resolution and precision is an applications requirement, the whole system must be configured for sufficient resolution when sampling at the bottom of the trapezoid. So one must purchase a system the covers the required resolution, and deploy it in such a way that the “worst case” sampling at the limits of the system are within the requirements. One must sample as many points as needed at the bottom of the FoV, and that impacts line scan rate.

Height of object matters too

Not only the position of the object in the FoV matters – but also the maximum height of any object whose Z dimension you need to detect. Let’s illustrate the point:

Figure 4. The maximum height anticipated matters too – Courtesy Teledyne DALSA

Consider item labeled Object in Figure 4. Your application’s object(s) may of course be shaped differently, but this generic object serves discussion purposes just fine. In this conceptual application, there’s a continuous conveyor belt (the dark grey surface) moving at continous speed in the Y dimension. Whenever no Object is present, i.e. the gaps between Object_N and Object_N+1, we expect the profiler to deliver a Z value of 0 for each pixel. But when an Object is present, we anticipate positive values corresponding to the height of the object. That’s the whole point of the 3D application.

Important note re. camera sensor in 2D

While the laser emits a flat line as it exits the projector, the reflection sensed inside the camera is two-dimensional. The camera sensor is a rectangular grid or array of pixels, typically in a CMOS chip, similar to that used in an area-scan camera. If one needs all the data from the sensor, the higher data volume takes longer to transfer than if one only needs a subset. If you know your application’s design well, you may be able to achieve optimized performance by avoiding the transfer of “empty” data.

Now let’s do a thought experiment where we re-imagine the Object towards two different extremes:

Extreme 1: Imagine the Object flattened down to a few sheets of paper in a tight stack, or perhaps the flap of a cardboard box.

Extreme 2: Imagine the Object is stretched up to the height of a full box, as high in the Z dimension as in the X dimension shown.

If the Object would never be higher than Extreme 1, the number of pixel rows in the camera sensor registering non-zero values will be just a few rows. Which can be read out quickly, not bothering to read out the unused rows. Yielding a relatively faster line rate.

But if the Object(s) will sometimes be at Extreme 2, many/most of the pixel rows in the camera sensor will register non-zero values, per the reflected laser line ranging up to the full height of the Object. Consequently more lines must be read-out from the camera sensor in order to build the laser profile.

1. The application must be designed to perform for the tallest anticipated Object, as well as the width of the Object in the X dimension and the speed of motion in the Y dimension.

2. All other things being equal, shorter objects, utilizing less camera sensor real estate, will support faster line rates, than taller object.

Summary points regarding object height

By careful planning for your FoV, knowing your timing constraints, and selecting your laser profiler model within it’s performance range, you can optimize your outcomes.

Click to contact
Give us some brief idea of your application and we will contact you to discuss camera options.

Also consider – interface capacity; exposure time

Just as with area scan cameras, output rates may be limited by any of interface limits, exposure duration, or data volumes.

Interface limits: Whether using GigE Vision, USB3 Vision, Camera Link HS – whatever – the interface standard, camera settings, cable, and PC adapter card together determine a maximum frame rate or line rate expressed in Gigabits per second (Gbps), typically. Your intended data volume is a function of exposure time and line rate or frame rate. Be sure to understand maximum practical throughput, choosing components accordingly.

Exposure duration: Even without readout timing considerations (overlapped readout together with start of next exposure – or completion of readout n before start of exposure n+1), if there are, say, 100 exposures per second, one cannot receive more than 100 datasets per second. Even if the camera is capable of faster rates.

That may seem obvious to experienced machine vision applications designers, but it needs mentioning for any new to this. Every application needs to achieve good contrast between the imaging subject and its background field. And if lighting and lensing are optimized, exposure time is the last variable to control. Ideally, lighting and lensing, together with the camera sensor, permit exposures brief enough so that exposure time meets application objectives.

But whether manually parameterized or under auto-exposure control, one has to do the math and/or the empirical testing to insure your achievable line rates aren’t exposure-limited.

Planning for your laser profiler application

Some months ago we wrote a blog which summarizes Teledyne DALSA’s Z-Trak line scan product families. Besides highlighting the characteristics of three distinct product families, we provided a worksheet to help users identify key applications requirements for line scanning. It’s worth offering that same worksheet again below. Consider printing the page or creating a copy of it in a spreadsheet, and fill in the values for your known or evolving application.

3D application key attributes

The moral of the story…

The takeaway is that the scan rate you’ll achieve for your application is more complex to determine than just reading a spec sheet about a laser profiler’s maximum performance. Your application configuration and constraints factor into overall performance.

1st Vision’s sales engineers have over 100 years of combined experience to assist in your camera and components selection.  With a large portfolio of cameraslensescablesNIC cards and industrial computers, we can provide a full vision solution!

About you: We want to hear from you!  We’ve built our brand on our know-how and like to educate the marketplace on imaging technology topics…  What would you like to hear about?… Drop a line to info@1stvision.com with what topics you’d like to know more about.

Plug’n Stream IDS Imaging μEye SCP / μEye SLE industrial dashcams

Unpacking the μEye SCP/ μEyeSLE product names to preview what’s on offer, the “S” stands for Streaming. The rest of the product names come from IDS Imaging’s popular GigE Vision camera families μEye CP and μEye LE, respectively. So the value proposition is a bundled streaming solution piggybacked on top of another product platform. This creates economies of scale for the manufacturer and the customer alike.

uEye SCP / SLE single-device solution for process monitoring – Courtesy IDS Imaging

Continuous monitoring with event-triggered video recording

Anybody running complex systems has to monitor them for performance and quality control – and/or to recover from breakdowns or detected concerns. Traditionally one had two options:

  • Wait for a breakdown and try to deduce what went wrong, or
  • Construct a video monitoring system from the constituent components… and program as needed

The “wait and see” options is attractively inexpensive on the face of it. But it risks expensive losses from the period prior to detecting the failures. Worse, it may not be possible to determine what went wrong if one missed the event that triggered the failure.

Constructing a video monitoring system from scratch is possible – and many have done it. Until now it generally required sourcing camera, lens, and PC, and writing complex software capable of episodic streaming and recording, and event-detection and logging.

IDS μEye SCP / μEye SLE provide Plug’n Stream no-PC-needed solution

Systems evolution in many fields, including machine vision, periodically takes what once had to be programmed to something that need only be configured. The system provider helpfully packages the algorithms into parameterized controls that are user-friendly to the deployer. That way one can focus on the application domain, event management, and process control.

IDS Imaging has done exactly that to create μEye SCP / μEye SLE – think “industrial dashcam” – with both housed and board-level options. The comprehensive 9 minute video below provides a great introduction to the product, its capabilities, and some applications examples.

The comprehensive 9 minute video below provides a great introduction to the product, its capabilities, and some applications examples.


9 minute introduction and overview – Courtesy IDS Imaging

Event Recording

The system streams continuously to internal persistent memory, periodically overwriting previous streams that were not part of any events deemed worth saving. This creates a recorded stream for a defined period from x seconds prior to an event, through the event, and to y seconds afterwards, where x and y are user-definable.

That documents machine malfunctions or failures. Which makes it easier to analyze process errors – and address them for system improvement.

No PC needed – System on a Chip (SoC)

With System on a Chip (SoC) from Ambarella, the camera has the onboard smarts to directly process and evaluate image data.

The user need only configure the parameters that define an “event”, the duration to capture before and after the event, which of several formats to record, and whether to operate standalone or integrated into other systems.

Use cases

Just to get the juices flowing, consider use cases like the following:

Industrial process monitoring – the human operator has the overview but μEye SCP / SLE can monitor automatically at the detail level – recording events and raising alerts if needed – Courtesy IDS Imaging
Video analysis – for scouting or officiating for example – Courtesy IDS Imaging
Smart city applications – let the system identify pedestrians within a specific field of view – Courtesy IDS Imaging

WebCockpit configuration

With no PC required to operate the system in standalone mode, configuration may be done through a frontend in the browser. The frontend settings control streaming, recording, and video modes.

Optionally on can use the web service and a REST API to seamlessly integrate into existing systems, for those who prefer or require integrate over standalone deployments.

1st Vision’s sales engineers have over 100 years of combined experience to assist in your camera and components selection.  With a large portfolio of cameraslensescablesNIC cards and industrial computers, we can provide a full vision solution!

About you: We want to hear from you!  We’ve built our brand on our know-how and like to educate the marketplace on imaging technology topics…  What would you like to hear about?… Drop a line to info@1stvision.com with what topics you’d like to know more about.

Teledyne DALSA 16k TDI line scan camera 1 MHz line rate

Product innovation continues to serve machine vision customers well. Clever designs are built for evolving customer demands and new markets, supported by electronics miniaturization and speed. Long a market leader in line scan imaging, Teledyne DALSA now offers the Linea HS2 TDI line scan camera family.

Linea HS2 16k TDI line scan camera with 1 MHz line rate – courtesy Teledyne DALSA

Video overview

The video below is just over one minute in duration, and provides a nice overview:

Contact us for a quote

Backside illumination enhances quantum efficiency

Early sensors were all used frontside illumination, and everybody lived with that until about 10 years ago when backside illumination was innovated and refined. The key insight was to let the photons hit the light-sensitive surface first, with the sensor’s wiring layer on the other side. This greatly improves quantum efficiency, as seen in the graph below:

QE substantially enhanced using backside illumination (BSI – Courtesy Teledyne DALSA

Applications

This camera series is designed for high-speed imaging in light staved conditions. Applications include but are not limited to inspecting flat panel displays, semiconductor wafers, high density interconnects, and diverse life science uses.

Courtesy Teledyne DALSA

Line scan cameras

You may already be a user of line scan cameras. If you are new to that branch of machine vision, compare and contrast line scan vs. area scan imaging. If you want the concept in a phrase or two, think “slice” or line of pixels obtained as the continuous wide target is passed beneath the camera. Repeat indefinitely. Can be used to monitor quality, detect defects, and/or tune controls.

Time Delay Integration (TDI)

Perhaps you even use Time Delay Integration (TDI) technology already. TDI builds on top of “simple” line scan by tracking how a pixel appears across several successive time slices, turning motion blur into an asset through hardware or software averaging and analysis.

Maybe you already have one or more of Teledyne DALSA’s prior-generation Linea HS line scan cameras. They feature the same pixel size, optics, and cables as the new Linea HS2 series. With a 2.5x speed increase the Linea HS2 provides a seamless upgrade. The Linea HS2 offers an optional cooling accessory to enhance thermal stability.

Frame grabber

The Linea HS2 utilizes Camera Link High Speed (CLHS) to match the camera’s data output rate with an interface that can keep up. Teledyne DALSA manufactures not just the camera, but also the Xtium2-CL MX4 Camera Link Frame Grabber.

Xtium2-CL MX4 Camera Link Frame Grabber – Courtesy Teledyne DALSA

The Xtium2-CL MX4 is built on next generation CLHS technology and features:

  • 16 Gigapixels per second
  • dual CLHS CX4 connectors
  • drives active optical cables
  • supports parallel data processing in up to 12 PCs
  • allows cable lengths over 100 meters with complete EMI immunity

Which camera to choose?

As this blog is released, the Linea HS2, with 16k/5μm resolution provides an industry leading maximum line rate of 1 MHz, or 16 Gigapixels per second data throughput. Do you need the speed and sensitivity of this camera? Or is one of the “kid brother” models enough – they are already highly performant before the new kid came along. We can help you sort out the specifications according to your application requirements.

1st Vision’s sales engineers have over 100 years of combined experience to assist in your camera and components selection.  With a large portfolio of cameraslensescablesNIC cards and industrial computers, we can provide a full vision solution!

About you: We want to hear from you!  We’ve built our brand on our know-how and like to educate the marketplace on imaging technology topics…  What would you like to hear about?… Drop a line to info@1stvision.com with what topics you’d like to know more about

IDS Ensenso B-Series: 3D Vision at Close Range

IDS has developed and released the Ensenso B-Series, ideal for short object distances as close as 21cm. Even that close it achieves a Field of View (FOV) of 30 x 26 cm and depth values accurate to 0.1 mm. While 3D machine vision isn’t new, this camera series is.

Ensenso B compact 3D camera – Courtesy IDS Imaging

Ensenso family of cameras

We introduced IDS’ Ensenso 3D cameras in 2023, bringing new stereo and structured light solutions to the portfolio. Then later in 2023 we announced IDS Ensenso C Series, which added color capabilities. That rounded out the lineup with differentiated offerings under each of the following identifiers: C, N, S, X, and XR. See all Ensenso models.

The new Enenso B-Series

This blog focuses on the new Ensenso B-Series. The cameras are ultra-compact, and can work at close range, still delivering a large FOV.

Ensenso B mounted on robotic arm – Courtesy IDS Imaging

The compact unit contains the stereo cameras as well as the bright pattern projector used to support stereo 3D imaging. The durable housing is rate for IP65/67 protection, and is ideal for harsh industrial environments.

Maybe you need Ensenso B

Or perhaps your application would be best served by Series C, N, S, X, or XR?

IDS Imaging Ensenso 3D cameras and camera systems are built for industrial 3D imaging with a GigE interface for ease of setup. Ensenso 3D cameras are suitable for numerous 3D imaging applications including robotics, bin picking, warehouse automation and 3D measurement tasks. They are widely used for many industrial applications such as factory automation, logistics, and quality assurance.

Ensenso 3D cameras have numerous features, benefits, and options.

Please contact us for more information. We can provide you with additional technical information and help you choose the right 3D camera system for your 3D imaging application.

1st Vision’s sales engineers have over 100 years of combined experience to assist in your camera and components selection.  With a large portfolio of cameraslensescablesNIC cards and industrial computers, we can provide a full vision solution!

About you: We want to hear from you!  We’ve built our brand on our know-how and like to educate the marketplace on imaging technology topics…  What would you like to hear about?… Drop a line to info@1stvision.com with what topics you’d like to know more about.