All information provided in this manual is believed to be accurate and reliable. No
responsibility is assumed by Photonfocus AG for its use. Photonfocus AG reserves the right to
make changes to this information without notice.
Reproduction of this manual in whole or in part, by any means, is prohibited without prior
permission having been obtained from Photonfocus AG.
The Swiss company Photonfocus is one of the leading specialists in the development of CMOS
image sensors and corresponding industrial cameras for machine vision.
Photonfocus is dedicated to making the latest generation of CMOS technology commercially
available. Active Pixel Sensor (APS) and global shutter technologies enable high speed and
high dynamic range (120 dB) applications, while avoiding disadvantages like image lag,
blooming and smear.
Photonfocus’ product range is complemented by custom design solutions in the area of camera
electronics and CMOS image sensors.
Photonfocus is ISO 9001 certified. All products are produced with the latest techniques in order
to ensure the highest degree of quality.
Photonfocus products are available through an extensive international distribution network
and through our key account managers. Contacts to our key account managers can be found
at www.photonfocus.com.
1.5Further information
Photonfocus reserves the right to make changes to its products and documentation without notice. Photonfocus products are neither intended nor certified for
use in life support systems or in other critical systems. The use of Photonfocus
products in such applications is prohibited.
Photonfocus is a trademark and LinLog®is a registered trademark of Photonfocus AG. CameraLink®and GigE Vision®are a registered mark of the Automated
Imaging Association. Product and company names mentioned herein are trademarks or trade names of their respective companies.
Reproduction of this manual in whole or in part, by any means, is prohibited
without prior permission having been obtained from Photonfocus AG.
8 of 117MAN073 12/2016 V1.0
Photonfocus can not be held responsible for any technical or typographical errors.
1.6Legend
In this documentation the reader’s attention is drawn to the following icons:
Important note, additional information
Important instructions
General warning, possible component damage hazard
1.6 Legend
Warning, electric shock hazard
Warning, fire hazard
MAN073 12/2016 V1.09 of 117
1 Preface
10 of 117MAN073 12/2016 V1.0
2
MV1 - D12 80 - L0 1- 3 D05 - 12 80 - G2 - 8
P r e f i x 1
P r e f i x 2
S e n s o r w i d t h
C a m e r a
s p e e d
I n t e r f a c e t y p e
I n t e r f a c e
r e s o l u t i o n
3 D d a t a f o r m a t
S e n s o r
M a n u f a c t u r e r
S e n s o r
F a m i l y
Introduction
This manual describes the Photonfocus 3D camera series that have a Gigabit Ethernet (GigE)
interface and are based on the image sensor LUX1310 sensors from Luxima Technology.
A list of all cameras covered in this manual is shown in Table 4.2. The term MV1-D1280-L01-3D05 is
used in this manual to denote all available cameras of this series.
2.1Camera Naming convention
The naming convention of the MV1-D1280-L01-3D05 camera series is summarized in Fig. 2.1.
Figure 2.1: Camera naming convention
Prefix1 Specifies the form factor of the camera. MV1 equates to the form factor of 55 x 55 mm
(width x height of the camera housing).
Prefix2 All cameras covered in this manual have D as Prefix2 which denotes area scan cameras
with digital camera interace.
Sensor width All cameras covered in this manual use sensors with a width of 1280 pixels.
Sensor Family Sensor Family of the prior indicated manufacturer. "01": LUX series
3D coordinate data format The available cameras have the 3D data format 3D05 which is
described in Section 5.2.7.
Camera speed The camera speed is specified as the product of the camera data clock in MHz
and the number of parallel data channels (taps) in the camera data path..
Interface type Available interface type options: "G2": Gigabit Ethernet with RS-422 interface
for a shaft (rotary) encoder; "H2": Gigabit Ethernet with HTL (High Threshold Logic)
interface (instead of RS-422) for a shaft (rotary) encoder
Interface resolution Maximal resolution (bit width) of the camera interface.
2
MAN073 12/2016 V1.011 of 117
2 Introduction
12 of 117MAN073 12/2016 V1.0
3
How to get started (3D GigE G2)
3.1Introduction
This guide shows you:
•How to install the required hardware (see Section 3.2)
•How to install the required software (see Section 3.3) and configure the Network Adapter
Card (see Section 3.4 and Section 3.5)
•How to acquire your first images and how to modify camera settings (see Section 3.6)
A GigE Starter Guide [MAN051] can be downloaded from the Photonfocus support page. It
describes how to access Photonfocus GigE cameras from various third-party tools.
To start with the laser detection it is recommended to use the PF 3D Suite which can be
downloaded from the software section of the Photonfocus web page. The PF 3D Suite is a free
GUI for an easy system set up and visualisation of 3D scan. To get started, please read the
manual which is available in a sub-folder of the PF3DSuite installation.
Prior to running the PF 3D Suite, the GigE system should be configured as indicated in this chapter.
3.2Hardware Installation
The hardware installation that is required for this guide is described in this section.
The following hardware is required:
•PC with Microsoft Windows OS (XP, Vista, Windows 7)
•A Gigabit Ethernet network interface card (NIC) must be installed in the PC. The NIC
should support jumbo frames of at least 9014 bytes. In this guide the Intel PRO/1000 GT
desktop adapter is used. The descriptions in the following chapters assume that such a
network interface card (NIC) is installed. The latest drivers for this NIC must be installed.
•Photonfocus GigE camera.
•Suitable power supply for the camera (see in the camera manual for specification) which
can be ordered from your Photonfocus dealership.
•GigE cable of at least Cat 5E or 6 with shielding.
Photonfocus GigE cameras can also be used under Linux.
Photonfocus GigE cameras work also with network adapters other than the Intel
PRO/1000 GT. The GigE network adapter should support Jumbo frames.
MAN073 12/2016 V1.013 of 117
3 How to get started (3D GigE G2)
E t h e r n e t J a c k ( R J 4 5 )
P o w e r S u p p l y
a n d I / O C o n n e c t o r
S t a t u s L E D
Do not bend GigE cables too much. Excess stress on the cable results in transmission errors. In robots applications, the stress that is applied to the GigE cable is
especially high due to the fast movement of the robot arm. For such applications,
special drag chain capable cables are available.
The following list describes the connection of the camera to the PC (see in the camera manual
for more information):
1.Remove the Photonfocus GigE camera from its packaging. Please make sure the following
items are included with your camera:
•Power supply connector
•Camera body cap
If any items are missing or damaged, please contact your dealership.
2.Connect the camera to the GigE interface of your PC with a shielded GigE cable of at least
Cat 5E or 6.
Figure 3.1: Rear view of the Photonfocus GigE camera series with power supply and I/O connector, Ethernet
jack (RJ45) and status LED
3.Connect a suitable power supply to the power plug. The pin out of the connector is
shown in the camera manual.
Check the correct supply voltage and polarity! Do not exceed the operating
voltage range of the camera.
14 of 117MAN073 12/2016 V1.0
3.3 Software Installation
A suitable power supply can be ordered from your Photonfocus dealership.
4.Connect the power supply to the camera (see Fig. 3.1).
3.3Software Installation
This section describes the installation of the required software to accomplish the tasks
described in this chapter.
1.Install the latest drivers for your GigE network interface card.
2.Download the latest eBUS SDK installation file from the Photonfocus server.
You can find the latest version of the eBUS SDK on the support (Software Download) page at www.photonfocus.com.
3.Install the eBUS SDK software by double-clicking on the installation file. Please follow the
instructions of the installation wizard. A window might be displayed warning that the
software has not passed Windows Logo testing. You can safely ignore this warning and
click on Continue Anyway. If at the end of the installation you are asked to restart the
computer, please click on Yes to restart the computer before proceeding.
4.After the computer has been restarted, open the eBUS Driver Installation tool (Start ->All Programs -> eBUS SDK -> Tools -> Driver Installation Tool) (see Fig. 3.2). If there is
more than one Ethernet network card installed then select the network card where your
Photonfocus GigE camera is connected. In the Action drop-down list select Install eBUSUniversal Pro Driver and start the installation by clicking on the Install button. Close the
eBUS Driver Installation Tool after the installation has been completed. Please restart the
computer if the program asks you to do so.
Figure 3.2: eBUS Driver Installation Tool
5.Download the latest PFInstaller from the Photonfocus server.
6.Install the PFInstaller by double-clicking on the file. In the Select Components (see Fig. 3.3)
dialog check PF_GEVPlayer and doc for GigE cameras. For DR1 cameras select additionally
DR1 support and 3rd Party Tools. For 3D cameras additionally select PF3DSuite2 and SDK.
.
MAN073 12/2016 V1.015 of 117
3 How to get started (3D GigE G2)
Figure 3.3: PFInstaller components choice
16 of 117MAN073 12/2016 V1.0
3.4 Network Adapter Configuration
3.4Network Adapter Configuration
This section describes recommended network adapter card (NIC) settings that enhance the
performance for GigEVision. Additional tool-specific settings are described in the tool chapter.
1.Open the Network Connections window (Control Panel -> Network and InternetConnections -> Network Connections), right click on the name of the network adapter
where the Photonfocus camera is connected and select Properties from the drop down
menu that appears.
Figure 3.4: Local Area Connection Properties
.
MAN073 12/2016 V1.017 of 117
3 How to get started (3D GigE G2)
2.By default, Photonfocus GigE Vision cameras are configured to obtain an IP address
automatically. For this quick start guide it is recommended to configure the network
adapter to obtain an IP address automatically. To do this, select Internet Protocol (TCP/IP)
(see Fig. 3.4), click the Properties button and select Obtain an IP address automatically
(see Fig. 3.5).
Figure 3.5: TCP/IP Properties
.
18 of 117MAN073 12/2016 V1.0
3.4 Network Adapter Configuration
3.Open again the Local Area Connection Properties window (see Fig. 3.4) and click on the
Configure button. In the window that appears click on the Advanced tab and click on JumboFrames in the Settings list (see Fig. 3.6). The highest number gives the best performance.
Some tools however don’t support the value 16128. For this guide it is recommended to
select 9014 Bytes in the Value list.
Figure 3.6: Advanced Network Adapter Properties
.
MAN073 12/2016 V1.019 of 117
3 How to get started (3D GigE G2)
4.No firewall should be active on the network adapter where the Photonfocus GigE camera
is connected. If the Windows Firewall is used then it can be switched off like this: Open
the Windows Firewall configuration (Start -> Control Panel -> Network and InternetConnections -> Windows Firewall) and click on the Advanced tab. Uncheck the network
where your camera is connected in the Network Connection Settings (see Fig. 3.7).
Figure 3.7: Windows Firewall Configuration
.
20 of 117MAN073 12/2016 V1.0
3.5 Network Adapter Configuration for Pleora eBUS SDK
3.5Network Adapter Configuration for Pleora eBUS SDK
Open the Network Connections window (Control Panel -> Network and Internet Connections ->
Network Connections), right click on the name of the network adapter where the Photonfocus
camera is connected and select Properties from the drop down menu that appears. A
Properties window will open. Check the eBUS Universal Pro Driver (see Fig. 3.8) for maximal
performance. Recommended settings for the Network Adapter Card are described in Section
3.4.
Figure 3.8: Local Area Connection Properties
3.6Getting started
This section describes how to acquire images from the camera and how to modify camera
settings.
1.Open the PF_GEVPlayer software (Start -> All Programs -> Photonfocus -> GigE_Tools ->
PF_GEVPlayer) which is a GUI to set camera parameters and to see the grabbed images
(see Fig. 3.9).
.
MAN073 12/2016 V1.021 of 117
3 How to get started (3D GigE G2)
Figure 3.9: PF_GEVPlayer start screen
2.Click on the Select / Connect button in the PF_GEVPlayer . A window with all detected
devices appears (see Fig. 3.10). If your camera is not listed then select the box Showunreachable GigE Vision Devices.
Figure 3.10: GEV Device Selection Procedure displaying the selected camera
3.Select camera model to configure and click on Set IP Address....
4.Select a valid IP address for selected camera (see Fig. 3.12). There should be no
exclamation mark on the right side of the IP address. Click on Ok in the Set IP Address
dialog. Select the camera in the GEV Device Selection dialog and click on Ok.
Figure 3.12: Setting IP address
5.Finish the configuration process and connect the camera to PF_GEVPlayer .
6.The camera is now connected to the PF_GEVPlayer. Click on the Play button to grab
images.
An additional check box DR1 resp. QR1 appears for DR1/QR1 cameras. The camera
is in double rate mode if this check box is checked. The demodulation is done
in the PF_GEVPlayer software. If the check box is not checked, then the camera
outputs an unmodulated image and the frame rate will be lower than in double
rate mode.
MAN073 12/2016 V1.023 of 117
3 How to get started (3D GigE G2)
Figure 3.13: PF_GEVPlayer is readily configured
If no images can be grabbed, close the PF_GEVPlayer and adjust the Jumbo
Frame parameter (see Section 3.3) to a lower value and try again.
Figure 3.14: PF_GEVPlayer displaying live image stream
7.Check the status LED on the rear of the camera.
The status LED light is green when an image is being acquired, and it is red when
serial communication is active.
8.Camera parameters can be modified by clicking on GEV Device control (see Fig. 3.15). The
visibility option Beginner shows most the basic parameters and hides the more advanced
parameters. If you don’t have previous experience with Photonfocus GigE cameras, it is
recommended to use Beginner level.
24 of 117MAN073 12/2016 V1.0
3.6 Getting started
Figure 3.15: Control settings on the camera
9.To modify the exposure time scroll down to the AcquisitionControl control category (bold
title) and modify the value of the ExposureTime property.
MAN073 12/2016 V1.025 of 117
3 How to get started (3D GigE G2)
26 of 117MAN073 12/2016 V1.0
4
Product Specification
4.1Introduction
This manual describes the Photonfocus MV1-D1280-L01-3D05 camera series. The cameras have
a Gigabit Ethernet interface and are optimized for very high speed laser triangulation
applications with 43700 profiles/s (at 1280 x 16 pixels resolution). The MV1-D1280-L01-3D05
camera contains a 1.3 megapixel CMOS image sensor. The laser line detection algorithm
(Center of Gravitiy) is able to compute the peak position of a laser line with sub-pixel accuracy.
Thus, the height profile of an object gets computed in real-time within the camera, making
additional calculations in the PC needless.
The cameras are built around the monochrome CMOS image sensor LUX1310 (1.3 MP),
developed by Luxima. The principal advantages are:
•Up to 43700 profiles/s (@ 1280 x 16 pixels resolution)
•Maximal scan area 1280 x 1024 pixels
•Center of Gravity (COG) based laser line detection with up to 1/16 sub pixel accuracy
•Combined 2D/3D applications can be realized in the 2D/3D mode of the camera (at a
reduced frame rate)
•Gigabit Ethernet interface with GigE Vision and GenICam compliance
•Global shutter
•Image sensor with high sensitivity
•Region of interest (ROI) freely selectable in x and y direction
•Column Fixed Pattern Noise Correction for improved image quality.
•Advanced I/O capabilities: 2 isolated trigger inputs, 2 differential isolated inputs (RS-422
or HTL) and 2 isolated outputs
•A/B shaft encoder interface: RS-422 (G2 models) or HTL (H2 models). HTL is recommended
for noisy environments.
•Programmable Logic Controller (PLC) for powerful operations on input and output signals
•Wide power input range from 12 V (-10 %) to 24 V (+10 %)
•The compact size of only 55 x 55 x 51.5 mm3makes the Photonfocus MV1-D1280-L01-3D05
cameras the perfect solution for applications in which space is at a premium
•Free GUI available (PF 3D Suite) for an easy system set up and visualisation of 3D scans
The basic components for 3D imaging consist of a laser line and a high speed CMOS camera in
a triangular arrangement to capture images (profiles) from objects that are moved on a
conveyor belt or in a similar setup (see Fig. 4.1 and Section 5.2.2).
You can find more information on the basics of laser triangulation and on the
principles of 3D image acquisition technique in the user manual "PF 3D Suite"
available in the support area at www.photonfocus.com.
MAN073 12/2016 V1.027 of 117
4 Product Specification
C o n v e y o r b e l t w i t h o b j e c t s
L a s e r
C a m e r a
Figure 4.1: Triangulation principle with objects moved on a conveyor belt
Figure 4.2: Photonfocus MV1-D1280-L01-3D05 camera series
28 of 117MAN073 12/2016 V1.0
4.2 Feature Overview
4.2Feature Overview
The general specification and features of the camera are listed in the following sections. The
detailed description of the camera features is given in Chapter 5.
Camera power supply+12 V DC (- 10 %) ... +24 V DC (+ 10 %)
1)
Trigger signal input range+5 .. +30 V DC
Typical power consumption< 8.7 W
Lens mountC-Mount
Dimensions55 x 55 x 51.5 mm
3
Mass258 g
ConformityRoHS, WEEE
Table 4.4: Physical characteristics and operating ranges (Footnotes:1)for minimal camera power consumption work with a power supply at +12V DC)
4.4.1Heat Dissipation
You must provide sufficient heat dissipation for the camera to maintain the temperature of the
camera housing at 50°C or less. The camera housing design ensures a good heat transfer to the
camera mounting. Since each system installation is unique, Photonfocus can give only hints for
proper heat dissipation:
•If your camera is mounted on a substantial metal component in your system and this part
is well below 40 °C, this may provide sufficient heat dissipation capability for the camera.
•If cooling over system parts is not possible, additional heat sinks may increase the heat
dissipation of the camera.
•The use of a fan to provide air flow over the camera is an extremely efficient method of
heat dissipation. The use of a fan in connection with additional heat sinks provides the
best heat dissipation. Fan operation may cause issues with dust in the optical path.
In all cases, you should monitor the temperature of the camera housing and make sure that
the temperature does not exceed 50°C. You can check the internal temperature of the camera
PCBs and the sensor with the help of temperature sensors inside the Photonfocus cameras. The
temperature of these sensors can be read out via software. Ensure that the PCB temperatures
not exceed 85°C.
To ensure good image quality, we recommend not to operate the camera at elevated
temperatures.
Note that elevated temperatures will worsen image quality and shorten the camera’s lifetime.
.
MAN073 12/2016 V1.031 of 117
4 Product Specification
4.4.2Absolute Maximum Ratings
ParameterValue
Power Supply Voltage26.4 V
ESD Contact Discharge Power Supply4 kV
ESD Air Discharge Power Supply8 kV
Fast Transients/Bursts Power Supply2 kV
Camera Control Input Signal Voltage Single Ended-30 V ... +30 V
Camera Control Input Signal Voltage RS422-25 V ... +25 V
Camera Control Input Signal Voltage HTL10 V ... 30 V
Common Mode Range Voltage RS422-10 V ... +13 V
Camera Control Output Signal Voltage Single Ended0 V ... +30 V
Camera Control Output Signal Output Current Single Ended0.5 A
Camera Control Output Signal Output Power Single Ended0.5 W
ESD Contact Discharge Camera Control Signals4 kV
ESD Air Discharge Camera Control Signals8 kV
Fast Transients/Bursts Data and Camera Control Signals1 kV
Surge immunity Data and Camera Control Signals1 kV
Table 4.5: Absolute Maximum Ratings
4.4.3Electrical Characteristics
ParameterValue
Camera Power Supply+12 V (-10%) ... +24 V (+10%)
Camera Control Input Single Ended+5 V ... +30 V
Camera Control Input RS422 Receiver Sensitivity+/- 200 mV
Camera Control Input RS422 Maximum Common Mode Range-7 V ... +7 V
Camera Control Input RS422 Maximum Differential Input Level10 V
Camera Control Input RS422 Minimum Differential Input Level2 V
ISO power RS-4225 V (-10%) ... 24 V (+10%)
ISO power HTL10 V (-10%) ... 30V (+10%)
Table 4.6: Electrical Characteristics
.
32 of 117MAN073 12/2016 V1.0
4.4 Technical Specification
0%
10%
20%
30%
40%
50%
60%
350400450500550600650700750800850900
Wavelength [nm]
LUX1310
4.4.4Spectral Response
Fig. 4.3 shows the quantum efficiency curve of the monochrome LUX1310 sensor from Luxima
Technology measured in the wavelength range from 400 nm to 1000 nm.
Figure 4.3: Spectral response of the LUX1310 CMOS monochrome image sensors (with micro lenses)
MAN073 12/2016 V1.033 of 117
4 Product Specification
34 of 117MAN073 12/2016 V1.0
5
Functionality
5.1Introduction
This chapter serves as an overview of the camera configuration modes and explains camera
features. The goal is to describe what can be done with the camera. The setup of the
MV1-D1280-L01-3D05 cameras is explained in later chapters.
5.23D Features
5.2.1Overview
The MV1-D1280-L01-3D05 cameras contain an accurate laser line detector for laser
triangulation (measurement of 3D profiles) that extracts 3D information in real time. For more
details see Section 5.2.4.
The camera should be placed so that the laser line is located in horizontal direction. The
outputs of the laser detector (COG) are the location coordinate of the laser line, the width of
the laser line and the grey value of the highest grey value inside the laser line (see Section
5.2.3).
The camera has a special mode (see 2D&3D mode in Section 5.2.5) for setup and debugging
purposes that allows to view the image and the detected laser line in the same image.
5.2.2Measuring Principle
For a triangulation setup a laser line generator and a camera is used. There are several
configurations which are used in the laser triangulation applications. Which setup is used in an
application is determined by the scattering of the material to be inspected. There are setups
for highly scattering materials and others for nearly reflecting surfaces.
In addition the penetration depth of light depends on the wavelength of light. The longer the
wavelength the deeper is the penetration of the light. Historically red line lasers with a
wavelength around 630 nm were used. With the modern high power semiconductor line laser
in blue (405 nm), green and also in the near infrared there is the possibility to adapt the
wavelengths due to the inspection needs.
But not only the penetration depth affects the choice of the wavelength of the line laser. For
an accurate measurement other disturbing effects as radiation or fluorescence of the object or
strong light from neighbourhood processes have to be suppressed by optical filtering and an
appropriate selection of the laser wavelength. Hot steel slabs for instance are best inspected
with blue line laser because of the possibility to separate the laser line with optical filters from
temperature radiation (Planck radiation) which occurs in red and NIR.
The accuracy of the triangulation system is determinate by the line extracting algorithm, the
optical setup, the quality parameters of the laser line generator and the parameters of the lens
which makes optical engineering necessary.
MAN073 12/2016 V1.035 of 117
5 Functionality
Camera
Line Laser
C a m e r a
L i n e L a s e r
Triangulation Setup 1
In this setup the camera looks with the viewing angle α on the laser line projected from the
top. A larger angle leads to a higher resolution. With larger angles the range of height is
reduced. Small angles have the benefit of little occlusions.
Figure 5.1: Triangulation setup 1
Triangulation Setup 2
This setup shows an opposite configuration of the laser line and the camera. The resolution at
same triangulation angle is slightly higher but artifacts which occur during the measurement at
borders of the object have to be suppressed by software.
Figure 5.2: Triangulation setup 2
36 of 117MAN073 12/2016 V1.0
5.2 3D Features
C a m e r a
L i n e L a s e r
C a m e r a
L i n e L a s e r
Triangulation Setup 3
In this setup the laser line generator and the camera are placed in a more reflecting
configuration. This gives more signal and could be used for dark or matte surfaces. In case of
reflecting surfaces there is only a little amount of scattering which can be used as signal for
triangulation. Also in this case this triangulation setup helps to get results.
Figure 5.3: Triangulation setup 3
Triangulation Setup 4
In contrast to the setup before this setup is used for high scattering material or for application
where strong reflections of the object have to be suppressed. The resolution is reduced due to
the relations of the angles α and β.
Figure 5.4: Triangulation setup 4
MAN073 12/2016 V1.037 of 117
5 Functionality
W i d t h
I n t e n s i t y
y - d i r e c t i o n
H e i g h t
G a u s s i a n s h a p e d
l a s e r l i n e
T h r e s h o l d
W i d t h T h r e s h o l d
5.2.3Laser Line Detection
The laser line detector takes a threshold value as its input. All pixels with grey value below the
threshold value will be ignored. This filters out the image background.
A second threshold value (WidthThreshold) is used in the calculation of the laser line width
(see also Fig. 5.5).
The output values are calculated column-wise. The camera takes the following measurement
data:
Laser line coordinate Vertical coordinate of the laser line peak
Laser line width The laser line width is the number of pixels that have a grey value above
WidthThreshold inside the laser line. If there are no pixels inside the laser line that have a
grey level above WidthThreshold, then the laser line width is 0. In this case the
WidthThreshold value should be changed.
Laser line height The laser line height is the highest grey value of the detected laser line.
The value of the threshold should be set slightly above the grey level of the image background. The threshold for the width calculation (WidthThreshold) should
not be smaller than the threshold for the laser line detection.
Figure 5.5: Schematic of the cross section of a laser line
.
38 of 117MAN073 12/2016 V1.0
5.2 3D Features
5.2.4Laser Line Detection Algorithm
Structured light based systems crucially rely on an accurate determination of the position of
the laser line. A Center Of Gravity (COG) algorithm is implemented in the
MV1-D1280-L01-3D05 camera series. A user-selectable threshold is used to separate the image
background from the laser line.
The orientation of the laser line should be horizontal. The position of the laser line is
determined seperately for every image column. The algorithm starts at the top of the image.
When the intensity of a pixel exceeds the user-settable threshold, the COG calculation starts.
The next pixel with intensity smaller than the threshold ends the calculation. If more than one
laser line is detected then the one with the highest peak intensity is chosen.
The line position (PEAK) is split into a coarse position and a fine position (sub-pixel). The coarse
position is based on the pixel pitch and is transferred in PEAK [15:4]. The sub-pixel position
that was calculated from the COG algorithm is mapped to PEAK [3:0] (see also Section 5.2.7).
5.2.5Camera Operating Modes
The camera has three modes that determine which data is transmitted to the user:
2Donly Laser detection is turned off and camera behaves as a normal area scan camera. This
mode serves as a preview mode in the setup and debugging phase.
2D&3D Laser line detection is turned on. The sensor image (2D image) is transmitted together
with the 3D data. In the PF 3D Suite, the detected laser line is shown as a coloured line in
the 2D image. This mode serves as a preview mode in the setup and debugging phase of
the triangulation system.
3Donly Laser line detection is turned on and only 3D data plus an additional image row is
transmitted. The scan rate of this mode is considerably faster than the 2D&3D mode. The
user can select if only the laser line positions (DataFormat3D=2) or also additional
information should be transmitted.
The 3Donly mode must be used to achieve the highest scan rate. DataFormat3D=2
gives the fastest scan rate but only laser line coordinates (pixel and sub-pixel
information) are transmitted.
5.2.6Peak Mirror
The property Peak0_Mirror flips the peak coordinates vertically by applying the formula
Peak0_3DH-p-1, where Peak0_3DH=height of scan area and p the detected laser position.
5.2.73D05 Data Format
For every image there are 4 lines that contain the 3D data. Every pixel contains 8 bits of 3D
data which are always placed in the 8 LSB. A table with the bit assignment of the 3D data for
DataFormat3D=3 is shown in Fig. 5.6, with DataFormat3D=4 in Fig. 5.7 and with DataFormat3D=2 in
Fig. 5.8. Note that every value described in this table (with exception of STAT data) is the value
for the corresponding image column. The laser line position coordinate (PEAK) is relative to
the scan area of the peak. To get the absolute position on the image sensor, the value
Peak0_3DY must be added for peak.
LL_HEIGHT and LL_WIDTH values are explained in Fig. 5.6 and in Fig. 5.7.
MAN073 12/2016 V1.039 of 117
5 Functionality
3 D r o w
76543210
B i t s
D e s c r i p t i o n
0
P E A K [ 1 5 : 8 ]
D e t e c t e d l a s e r l i n e c o o r d i n a t e .
P E A K [ 1 5 : 4 ] : i n t e g e r p a r t , P E A K [ 3 : 0 ] : f r a c t i o n a l p a r t .
1
2
3
P E A K [ 7 : 0 ]
' 0 '' 0 'L L _ W I D T H [ 5 : 0 ]
L L _ H E I G H T [ 3 : 0 ]S T A T
L L _ W I D T H : l a s e r l i n e w i d t h
L L _ H E I G H T : h i g h e s t g r e y v a l u e i n s i d e t h e c r o s s
s e c t i o n o f t h e l a s e r l i n e ( 4 M S B ) .
S T A T : S t a t u s i n f o r m a t i o n
3 D r o w
76543210
B i t s
D e s c r i p t i o n
0
P E A K [ 1 5 : 8 ]
D e t e c t e d l a s e r l i n e c o o r d i n a t e .
P E A K [ 1 5 : 4 ] : i n t e g e r p a r t , P E A K [ 3 : 0 ] : f r a c t i o n a l p a r t .
1
2
3
P E A K [ 7 : 0 ]
L L _ W I D T H [ 3 : 0 ]
L L _ H E I G H T [ 7 : 0 ]
S T A T
L L _ H E I G H T : h i g h e s t g r e y v a l u e i n s i d e t h e c r o s s
s e c t i o n o f t h e l a s e r l i n e ( 8 b i t ) .
L L _ W I D T H : l a s e r l i n e w i d t h ( 4 M S B ) . W i d t h o f l a s e r
l i n e i s 4 * L L _ W I D T H
S T A T : S t a t u s i n f o r m a t i o n
3 D r o w
76543210
B i t s
D e s c r i p t i o n
0P E A K [ 1 5 : 8 ]
D e t e c t e d l a s e r l i n e c o o r d i n a t e .
P E A K [ 1 5 : 4 ] : i n t e g e r p a r t , P E A K [ 3 : 0 ] : f r a c t i o n a l p a r t .
1
P E A K [ 7 : 0 ]
STAT value: the status (value) of some parameters and internal registers are placed here. The
status information is described in Section 5.8.2. In every pixel (column) 4 bits of this status
information send, starting with the LSB in the first column.
Figure 5.6: 3D data format, DataFormat3D=3
Figure 5.7: 3D data format. DataFormat3D=4
Figure 5.8: 3D data format. DataFormat3D=2
Calculation example (DataFormat3D=3): Suppose that the 3D data of image column n has the
following data: 14 / 140 / 10 / 128 (see also Fig. 5.9).
The position of the laser line is in this case 232.75: integer part is calculated from the 8 bits of
3D row 0 followed by the 4 MSB of 3D row 1: 0b000011101000 = 0x0e8 = dec 232. The
40 of 117MAN073 12/2016 V1.0
5.2 3D Features
P o s i t i o n i n t e g r a l p a r t
P o s i t i o n f r a c t i o n a l p a r t
L a s e r l i n e w i d t h
L L _ H E I G H T
3 D r o w
76543210
B i t s
0
1
2
3
00001110
10001100
00001010
10000000
fractional part is calculated from the 4 LSB of 3D row 1: 0b1100 = 0xc. This value must be
divided by 16: 0xc / 16 = 0.75.
The laser line width is 10 pixels (6 LSB of 3D row 2).
The LL_HEIGHT value is in this case 8 (=4 MSB of 3D row 3). This means that the 4 MSB of the
highest gray value have the value 8. At 12 bit resolution, the highest gray value lies between
0x800 (=2048) and 0x8ff (=2303).
Figure 5.9: 3D data calculation example
MAN073 12/2016 V1.041 of 117
5 Functionality
0
0
W
P e a k 0 _ 3 D H
R a w i m a g e
S c a n a r e a f o r p e a k 1
3 D d a t a3 D d a t a r o w s
3 D d a t a r o w s : 2 ( D a t a F o r m a t 3 D = 2 ) o r 4 ( D a t a F o r m a t 3 D = 0 o r 1 )
0
0
W
3 D d a t a3 D d a t a r o w s
3 D d a t a r o w s : 2 ( D a t a F o r m a t 3 D = 2 ) o r 4 ( D a t a F o r m a t 3 D = 0 o r 1 )
5.2.8Transmitted data in 2D&3D mode
The transmitted image in 2D&3D mode is shown in Fig. 5.10. The data is transmitted in the
following order:
•Raw image
•3D data
Resulting height in 2D&3D mode is:
Hres = Peak0_3DH + 4 (if DataFormat3D=3 or 4), or
Hres = Peak0_3DH + 2 (if DataFormat3D=2)
Figure 5.10: Transmitted image in 2D&3D mode
5.2.9Transmitted data in 3Donly mode
In 3Donly mode only the 3D data is transmitted. The FrameCombine feature (see Section
5.2.10) was added to lower the transmitted frame rate. For FrameCombine = f, the data for f
images are combined into one image.
Resulting height in 3Donly mode is therefore:
Hres = 4 (if DataFormat3D=3 or 4), or
Hres = 2 (if DataFormat3D=2)
Figure 5.11: Transmitted image in 3Donly mode
.
42 of 117MAN073 12/2016 V1.0
5.2 3D Features
I n d i v i d u a l i m a g e s
C o m b i n e d i m a g e
F r a m e C o m b i n e
5.2.10Frame Combine
Very high frame rates, that are well over 1000 fps, can be achieved in 3Donly mode. Every
frame (image) activates an interrupt in the GigE software which will issue a high CPU load or
the frame rate can not be handled at all by an overload of interrupts.
To solve this issue, the FrameCombine mode has been implemented in the
MV1-D1280-L01-3D05 camera. In this mode, the data of n images are bundled into one frame.
In the example shown in Fig. 5.12 4 frames are combined into one frame
(FrameCombineNrFrames=4). In this case there are 4 times less software interrupts that indicate a
new frame than without FrameCombine and the CPU load is significantly reduced. Instead of
receiving 4 images with 4 rows, only one image with 16 rows is received which reduces the
frame rate that sees the computer. Without FrameCombine, the CPU load on the computer
might be too high to receive all images and images might be dropped. The value n
(=FrameCombineNrFrames) can be set by the user. This value should be set so that the resulting
frame rate is well below 1000 fps (e.g. at 100 fps). E.g. if the camera shows a maximal frame
rate of 4000 (property AcquisitionFrameRateMax), then FrameCombineNrFrames could be set to 40 to
have a resulting frame rate of 100 fps. The PF 3D Suite supports this mode.
The FrameCombineNrFrames value should be set so that the resulting frame rate is
not too high. A recommended target value, for example, is 100 fps.
Figure 5.12: Example for FrameCombine with 4 frames
Frame Combine Timeout
There exist possibilities to transmit the combined frame even if there is not enough data to fill
it. E.g. It can be desirable to get the 3D data immediately after an item on the conveyor belt
has passed.
FrameCombine_Timeout A timeout can be specified after which the combined frame will be
transmitted, regardless if there was enough data to fill it. The timeout counter is reset
after each frame and counts until a new trigger has been detected or until the timeout is
reached.
A FrameCombine_Timeout value of 0 disables the FrameCombine timeout feature.
MAN073 12/2016 V1.043 of 117
5 Functionality
3 D d a t a f r a m e 0
1 8 7
6 800
. . .
0
1 8 76 800
. . .
0
1 8 76 800
. . .
0
1 8 76 800
. . .
0
. . .
. . .
. . .
. . .
. . .
. . .
3 D d a t a f r a m e 1
3 D d a t a f r a m e 2
D u m m y f r a m e s
}
FrameCombine_ForceTimeout The transmission of the combined frame is forced by writing to
the FrameCombine_ForceTimeout property.
When the FrameCombine is finished by a timeout, then the remaining data in the combined
frame will be filled with filler data: the first two pixels of every filler row have the values 0xBB
(decimal 187) and 0x44 (decimal 68). The remaining pixels of the filler rows have the value 0.
An example is shown in Fig. 5.13. The timeout occured after the second frame and the two
remaining frames are filled with dummy data. DataFormat3D=2 is used in this example.
The FrameCombine mode is only available in 3Donly mode.
When acquisition is stopped, then a pending combined frame will be discarded.
To get the pending combined frame, a FrameCombine_ForceTimeout command must
be sent prior to stopping the acquisition.
Figure 5.13: Example for timeout with dummy frames in FrameCombine with 5 frames
44 of 117MAN073 12/2016 V1.0
5.2 3D Features
P e a k F i l t e r H e i g h t M i n
I n t e n s i t y
y - d i r e c t i o n
T h r e s h o l d
o k
f i l t e r e d : h e i g h t t o o b i g
f i l t e r e d : h e i g h t t o o s m a l l
P e a k F i l t e r H e i g h t M a x
5.2.11Peak Filter
Peaks that are detected by the laser line detection algorithm can be filtered by applying the
parameters described in this section. A filtered peak appears as all 3D data set to 0, which is
the same as if no peak occurred.
Filtering peaks might increase the robustness of the 3D application by filtering peaks that were
caused by unwanted effects, such as reflections of the laser beam.
PeakFilter parameters:
Peak0_EnPeakFilter Enable peak filtering. If set to False, the PeakFilter settings are ignored.
Peak0_PeakFilterHeightMin Filters all peaks (columns) where 256*LL_HEIGHT <
Peak0_PeakFilterHeightMin (see Fig. 5.6 and Fig. 5.14).
Peak0_PeakFilterHeightMax Filters all peaks (columns) where 256*LL_HEIGHT >
Peak0_PeakFilterHeightMax (see Fig. 5.6 and Fig. 5.14).
Peak0_PeakFilterWidthMin Filters all peaks (columns) where LL_WIDTH <
Peak0_PeakFilterWidthMin (see Fig. 5.6 and Fig. 5.15).
Peak0_PeakFilterWidthMax Filters all peaks (columns) where LL_WIDTH >
Peak0_PeakFilterWidthMax (see Fig. 5.6 and Fig. 5.15).
An illustration of the PeakFilterHeight parameters is shown in Fig. 5.14. The red line denotes a
situation where the laser peak is filtered because the height is too big or too small. An
illustration of the PeakFilterWidth parameters is shown in Fig. 5.15. The red line denotes a
situation where the laser peak is filtered because the width is too big or too small.
Figure 5.14: Illustration of the PeakFilterHeight parameters
.
MAN073 12/2016 V1.045 of 117
5 Functionality
P e a k F i l t e r W i d t h M i n
I n t e n s i t y
y - d i r e c t i o n
T h r e s h o l d
W i d t h T h r e s h o l d
P e a k F i l t e r W i d t h M a x
o k
f i l t e r e d : w i d t h t o o b i g
f i l t e r e d : w i d t h
t o o s m a l l
Figure 5.15: Illustration of the PeakFilterWidth parameters
46 of 117MAN073 12/2016 V1.0
5.2 3D Features
5.2.12Absolute Coordinates
The 3D coordinates are given relative to the start of the 3D ROI as a default.
When the property Peak0_EnAbsCoordinate is set to True then the 3D coordinates are given
relative to the value of the property Peak0_AbsCoordinateBase. This is useful if the 3D-ROI is not
kept constant.
Example: Peak0_EnAbsCoordinate = True, Peak0_AbsCoordinateBase = 0, Peak0_3DY=200: If the peak
is detected in row 50 of the ROI, the value 250 (50+Peak0_3DY) would be given as resulting 3D
coordinate.
The value of Peak0_AbsCoordinateBase must fulfill the following conditions:
(Peak0_AbsCoordinateBase <= Peak0_3DY) and (Peak0_AbsCoordinateBase + 1024 >=
Peak0_3DY + Peak0_3DH).
MirrorPeak and absolute coordinates
If Peak0_Mirror =True and Peak0_EnAbsCoordinate = True then the formula to calculate the 3D
coordinate is:
where c is the original (relative) coordinate without mirroring. MAX_H=1024. This is the same as
mirroring in a ROI with Y=Peak0_AbsCoordinateBase and H=MAX_H.
MAN073 12/2016 V1.047 of 117
5 Functionality
5.3Reduction of Image Size
5.3.1Region of Interest (ROI) (2Donly mode)
This section describes the ROI features in the 2Donly mode where the camera behaves as a
standard area scan camera.
The maximal frame rate of the 2Donly mode is considerably lower than in the
3Donly mode.
Some applications do not need full image resolution. By reducing the image size to a certain
region of interest (ROI), the frame rate can be drastically increased. A region of interest can be
almost any rectangular window and is specified by its position within the full frame and its
width and height. Table 5.1 shows some numerical examples of how the frame rate can be
increased by reducing the ROI.
The ROI is determined by the following properties: OffsetX, Width, OffsetY, Height.
The maximal frame rate can be increased by decreasing the image width. The maximal frame
rate is reached at a width of 768 pixels. Below this value, the maximal frame rate does not
increase anymore.
ROI DimensionMV1-D1280-L01-3D05
1280 x 1024 (SXGA)85 fps
1280 x 768 (WXGA)110 fps
800 x 600 (SVGA)230 fps
640 x 480 (VGA)360 fps
Table 5.1: Frame rates of different ROI settings in 2Donly mode (minimal exposure time)
48 of 117MAN073 12/2016 V1.0
5.3 Reduction of Image Size
5.3.2Region of Interest (ROI) in 3D modes
The ROI is determined by the following properties: OffsetX, Width, Peak0_3DY, Peak0_3DH,
DataFormat3D.
The area for the detection of the laser line is defined by the parameters Peak0_3DY,
Peak0_3DH. These values can be freely set within the following limits:
1.The minimal height of the laser triangulation region (Peak0_3DH) is 16 pixels.
2.The maximal height of the laser triangulation region (Peak0_3DH) is 1024.
The maximal frame rate can be increased by decreasing the image width. The maximal frame
rate is reached at a width of 768 pixels. Below this value, the maximal frame rate does not
increase anymore.
The maximal frame rates for the MV1-D1280-L01-3D05 camera series are shown in Table 5.2.
Width x Peak0_3DH2D&3D mode (4 3D rows)3Donly (4 3D rows)3Donly (2 3D rows)
1280 x 164370 fps21880 fps43700 fps
1280 x 322430 fps21880 fps26500 fps
1280 x 641285 fps14100 fps14100 fps
1280 x 128660 fps7340 fps7340 fps
1280 x 256335 fps3740 fps3740 fps
1280 x 512165 fps1885 fps1885 fps
1280 x 102485 fps948 fps948 fps
768 x 168750 fps43760 fps68847 fps
768 x 324860 fps40775 fps40775 fps
768 x 642575 fps22460 fps22460 fps
Table 5.2: MV1-D1280-L01-3D05 Frame rates of different ROI settings in 3D modes (minimal exposure time,
free-running mode)
MAN073 12/2016 V1.049 of 117
5 Functionality
5.4Trigger and Strobe
5.4.1Trigger Source
The trigger signal can be configured to be active high or active low by the TriggerActivation
(category AcquisitionControl) property. One of the following trigger sources can be used:
Free running The trigger is generated internally by the camera. Exposure starts immediately
after the camera is ready and the maximal possible frame rate is attained, if
AcquisitionFrameRateEnable is disabled. Settings for free running trigger mode:
TriggerMode = Off. In Constant Frame Rate mode (AcquisitionFrameRateEnable = True),
exposure starts after a user-specified time has elapsed from the previous exposure start so
that the resulting frame rate is equal to the value of AcquisitionFrameRate.
Software Trigger The trigger signal is applied through a software command (TriggerSoftware
in category AcquisitionControl). Settings for Software Trigger mode: TriggerMode = On
and TriggerSource = Software.
Line1 Trigger The trigger signal is applied directly to the camera by the power supply
connector through pin ISO_IN1 (see also Section A.1). A setup of this mode is shown in
Fig. 5.16 and Fig. 5.17. The electrical interface of the trigger input and the strobe output
is described in Section 6.8. Settings for Line1 Trigger mode: TriggerMode = On and
TriggerSource = Line1.
PLC_Q4 Trigger The trigger signal is applied by the Q4 output of the PLC (see also Section 6.9).
Settings for PLC_Q4 Trigger mode: TriggerMode = On and TriggerSource = PLC_Q4.
ABTrigger Trigger from incremental encoder (see Section 5.4.9).
Some trigger signals are inverted. A schematic drawing is shown in Fig. 7.4.
The available acquisition modes are shown in Table 5.3.
The ContinuousRecording and ContinousReadout modes can be used if more than
one camera is connected to the same network and need to shoot images simultaneously. If all cameras are set to Continuous mode, then all will send the
packets at same time resulting in network congestion. A better way would be to
set the cameras in ContinuousRecording mode and save the images in the memory
of the IPEngine. The images can then be claimed with ContinousReadout from one
camera at a time avoid network collisions and congestion.
AcquisitionModeAfter the command AcquisitionStart is executed:
ContinuousCamera aquires image frames continuously. Acquisition can be
stopped by executing AcquisitionStop command.
SingleFrameCamera acquires one frame and acquisition stops.
MultiFrameCamera acquires n=AcquisitionFrameCount frames and acquisition
stops.
SingleFrameRecordingCamera saves one image on the on-board memory of the IP engine.
SingleFrameReadoutOne image is acquired from the IP engine’s on-board memory. The
image must have been saved in the SingleFrameRecording mode.
ContinuousRecordingCamera continuosly saves images on the on-board memory of the
IP engine until the memory is full. The size of available on-board
memory is 24 MB.
ContinousReadoutAll Images that have been previously saved by the
ContinuousRecording mode are acquired from the IP engine’s
on-board memory.
Table 5.3: AcquisitionMode and Trigger
.
52 of 117MAN073 12/2016 V1.0
5.4 Trigger and Strobe
e x t e r n a l t r i g g e r p u l s e i n p u t
t r i g g e r a f t e r i s o l a t o r
t r i g g e r p u l s e i n t e r n a l c a m e r a c o n t r o l
d e l a y e d t r i g g e r f o r s h u t t e r c o n t r o l
i n t e r n a l s h u t t e r c o n t r o l
d e l a y e d t r i g g e r f o r s t r o b e c o n t r o l
i n t e r n a l s t r o b e c o n t r o l
e x t e r n a l s t r o b e p u l s e o u t p u t
t
d - i s o - i n p u t
t
j i t t e r
t
t r i g g e r - d e l a y
t
e x p o s u r e
t
s t r o b e - d e l a y
t
d - i s o - o u t p u t
t
s t r o b e - d u r a t i o n
t
t r i g g e r - o f f s e t
t
s t r o b e - o f f s e t
5.4.3Exposure Time Control
The exposure time is defined by the camera. For an active high trigger signal, the camera starts
the exposure with a positive trigger edge and stops it when the programmed exposure time
has elapsed.
External Trigger
In the external trigger mode with camera controlled exposure time the rising edge of the
trigger pulse starts the camera states machine, which controls the sensor and optional an
external strobe output. Fig. 5.18 shows the detailed timing diagram for the external trigger
mode with camera controlled exposure time.
Figure 5.18: Timing diagram for the camera controlled exposure time
The rising edge of the trigger signal is detected in the camera control electronic which is
implemented in an FPGA. Before the trigger signal reaches the FPGA it is isolated from the
camera environment to allow robust integration of the camera into the vision system. In the
signal isolator the trigger signal is delayed by time t
FPGA which leads to a jitter of t
. The pulse can be delayed by the time t
jitter
can be configured by a user defined value via camera software. The trigger offset delay
t
trigger−offset
exposure time t
The trigger pulse from the internal camera control starts also the strobe control state machines.
The strobe can be delayed by t
the customer via software settings. The strobe offset delay t
results then from the synchronous design of the FPGA state machines. The
is controlled with an internal exposure time controller.
strobe−delay
exposure
d−iso−input
. This signal is clocked into the
trigger−delay
which
with an internal counter which can be controlled by
strobe−delay
results then from the
synchronous design of the FPGA state machines. A second counter determines the strobe
duration t
MAN073 12/2016 V1.053 of 117
strobe−duration
(strobe-duration). For a robust system design the strobe output is also
5 Functionality
isolated from the camera electronic which leads to an additional delay of t
d−iso−output
. Table
5.4 gives an overview over the minimum and maximum values of the parameters.
5.4.4Trigger Delay
The trigger delay is a programmable delay in milliseconds between the incoming trigger edge
and the start of the exposure. This feature may be required to synchronize the external strobe
with the exposure of the camera.
5.4.5Trigger Divider
The Trigger Divider reduces the trigger frequency that is applied to the camera. Every n-th
trigger is processed for a setting of TriggerDivider = n. If n=1, then every trigger is processed
(default behaviour). Fig. 7.4 shows the position of the TriggerDivider block.
TriggerDivider is ignored if trigger mode must be set to free-running Trigger
(TriggerMode = Off).
5.4.6Burst Trigger
The camera includes a burst trigger engine. When enabled, it starts a predefined number of
acquisitions after one single trigger pulse. The time between two acquisitions and the number
of acquisitions can be configured by a user defined value via the camera software. The burst
trigger feature works only in the mode "Camera controlled Exposure Time".
The burst trigger signal can be configured to be active high or active low. When the frequency
of the incoming burst triggers is higher than the duration of the programmed burst sequence,
then some trigger pulses will be missed. A missed burst trigger counter counts these events.
This counter can be read out by the user.
The burst trigger mode is only available when TriggerMode=On. Trigger source is determined by
the TriggerSource property.
The timing diagram of the burst trigger mode is shown in Fig. 5.19.
5.4.7Trigger Timing Values
Table 5.4 shows the values of the trigger timing parameters.
.
54 of 117MAN073 12/2016 V1.0
5.4 Trigger and Strobe
e x t e r n a l t r i g g e r p u l s e i n p u t
t r i g g e r a f t e r i s o l a t o r
t r i g g e r p u l s e i n t e r n a l c a m e r a c o n t r o l
d e l a y e d t r i g g e r f o r s h u t t e r c o n t r o l
i n t e r n a l s h u t t e r c o n t r o l
d e l a y e d t r i g g e r f o r s t r o b e c o n t r o l
i n t e r n a l s t r o b e c o n t r o l
e x t e r n a l s t r o b e p u l s e o u t p u t
t
d - i s o - i n p u t
t
j i t t e r
t
t r i g g e r - d e l a y
t
e x p o s u r e
t
s t r o b e - d e l a y
t
d - i s o - o u t p u t
t
s t r o b e - d u r a t i o n
t
t r i g g e r - o f f s e t
t
s t r o b e - o f f s e t
d e l a y e d t r i g g e r f o r b u r s t t r i g g e r e n g i n e
t
b u r s t - t r i g g e r - d e l a y
t
b u r s t - p e r i o d - t i m e
Figure 5.19: Timing diagram for the burst trigger mode
MAN073 12/2016 V1.055 of 117
5 Functionality
MV1-D1280-L01-3D05MV1-D1280-L01-3D05
Timing ParameterMinimumMaximum
t
d−iso−input
t
d−RS422−input
t
jitter
t
trigger−delay
t
burst−trigger−delay
t
burst−period−time
t
trigger−offset
t
trigger−offset
t
exposure
t
strobe−delay
t
strobe−offset
t
strobe−offset
t
strobe−duration
t
d−iso−output
t
trigger−pulsewidth
(non burst mode)100 nsduration of 1 row
(burst mode)125 ns125 ns
(non burst mode)100 ns100 ns
(burst mode)125 ns125 ns
depends on camera settings0.20 s
1 µs1.5 µs
65 ns185 ns
012.5 ns
00.20 s
00.20 s
10 µs0.41 s
600 ns0.34 s
200 ns0.34 s
150 ns350 ns
200 nsn/a
Number of bursts n130000
Table 5.4: Summary of timing parameters relevant in the external trigger mode using camera MV1-D1280L01-3D05
56 of 117MAN073 12/2016 V1.0
5.4 Trigger and Strobe
5.4.8Software Trigger
The software trigger enables to emulate an external trigger pulse by the camera software
through the serial data interface. It works with both burst mode enabled and disabled. As
soon as it is performed via the camera software, it will start the image acquisition(s),
depending on the usage of the burst mode and the burst configuration. The trigger mode
must be set to external Trigger (TriggerMode = On).
5.4.9A/B Trigger for Incremental Encoder
An incremental encoder with A/B outputs can be used to synchronize the camera triggers to
the speed of a conveyor belt. These A/B outputs can be directly connected to the camera and
appropriate triggers are generated inside the camera.
The A/B Trigger feature is is not available on all camera revisions, see Appendix
B for a list of available features.
In this setup, the output A is connected to the camera input ISO_INC0 (see also Section 6.8.4
and Section A.1) and the output B to ISO_INC1.
In the camera default settings the PLC is configured to connect the ISO_INC inputs to the A/B
camera inputs. This setting is listed in Section 7.10.3.
The following parameters control the A/B Trigger feature:
TriggerSource Set TriggerSource to ABTrigger to enable this feature
ABMode Determines how many triggers should be generated. Available modes: single,
double, quad (see description below)
ABTriggerDirection Determines in which direction a trigger should be generated: fwd: only
forward movement generates a trigger; bkwd: only backward movement generates a
trigger; fwdBkwd: forward and backward movement generate a trigger.
ABTriggerDeBounce Suppresses the generation of triggers when the A/B signal bounce.
ABTriggerDeBounce is ignored when ABTriggerDirection=fwdbkwd.
ABTriggerDivider Specifies a division factor for the trigger pulses. Value 1 means that all
internal triggers should be applied to the camera, value 2 means that every second
internal trigger is applied to the camera.
EncoderPosition (read only) Counter (signed integer) that corresponds to the position of
incremental encoder. The counter frequency depends on the ABMode. It counts up/down
pulses independent of the ABTriggerDirection. Writing to this property resets the counter
to 0.
A/B Mode
The property ABMode takes one of the following three values:
Single A trigger is generated on every A/B sequence (see Fig. 5.20). TriggerFwd is the trigger
that would be applied if ABTriggerDirection=fwd, TriggerBkwd is the trigger that would be
applied if ABTriggerDirection=bkwd, TriggerFwdBkwd is the trigger that would be applied if
ABTriggerDirection=fwdBkwd. GrayCounter is the Gray-encoded BA signal that is shown as an
aid to show direction of the A/B signals. EncoderCounter is the representation of the
current position of the conveyor belt. This value is available as a camera register.
MAN073 12/2016 V1.057 of 117
5 Functionality
A
B
G r a y C o u n t e r
E n c o d e r C o u n t e r
T r i g g e r F w d
T r i g g e r B k w d
01230123210321230
01210
T r i g g e r F w d B k w d
1
1
A
B
G r a y C o u n t e r
E n c o d e r C o u n t e r
T r i g g e r F w d
T r i g g e r B k w d
01230123210321230
012343212
T r i g g e r B k w d
1
3
A
B
G r a y C o u n t e r
E n c o d e r C o u n t e r
T r i g g e r F w d
T r i g g e r B k w d
01230123210321230
01234567654321234
T r i g g e r F w d B k w d
1
5
Double Two triggers are generated on every A/B sequence (see Fig. 5.21).
Quad Four triggers are generated on every A/B sequence (see Fig. 5.22).
.
There is a bug in the single A/B trigger mode in some camera revisions (see Appendix B). In this case when the encoder position moves back and forth by a
small amount, the EncoderCounter is incremented and the decrement is sometimes omitted, leading to a wrong EncoderPosition indication in the camera.
Therefore the single A/B trigger mode should not be used in the affected versions. To have the same behaviour as the single trigger mode, but without the
bug, use the double A/B mode and double the value of ABTriggerDivider.
Figure 5.20: Single A/B Mode
Figure 5.21: Double A/B Mode
Figure 5.22: Quad A/B Mode
.
58 of 117MAN073 12/2016 V1.0
5.4 Trigger and Strobe
B o u n c i n g
A
B
G r a y C o u n t e r
E n c o d e r C o u n t e r
T r i g g e r F w d
W a t e r m a r k
01331
012
220
33524
012543
0 3
4 3
Q u a d A / B M o d e , D e b o u n c i n g
f o r w a r d m o v e m e n t
h i g h w a t e r m a r k i s s a v e d
b a c k w a r d m o v e m e n tf o r w a r d m o v e m e n t
t r i g g e r w h e n w a t e r m a r k i s e x c e e d e d
A
B
G r a y C o u n t e r
E n c o d e r C o u n t e r
T r i g g e r F w d
W a t e r m a r k
01
01
2
01
3012301
2233445
5
03210123012301
66778899876545267891 01 11 21 31 4
1 01 11 21 31 4
A/B Trigger Debounce
A debouncing logic can be enabled by setting ABTriggerDeBounce=True. It is implemented with a
watermark value of the EncoderCounter (see Fig. 5.23). Suppose ABTriggerDirection=fwd, then
the watermark value is increased with the increments of the EncoderCounter. If
EncoderCounter decreases, e.g. Due to bouncing problems, the watermark value is hold
unchanged. Triggers are then only generated when the watermark value increases.
Figure 5.23: A/B Trigger Debouncing, example with ABMode=quad
The A/B Trigger Debounce mode can also be used for another issue:
In some applications the conveyor belt may stop between parts. In practice the conveyor belt
stops and retraces by a small amount which may cause a misalignment in the system. If
ABTriggerDirection=fwd is used and the Debounce mode is enabled and the conveyor belt starts
again in forward direction, no triggers are generated for the amount that the conveyor belt
retraced (see Fig. 5.24). The highest value of the EncoderCounter is stored as the watermark.
Triggers are only generated when the EncoderCounter is at the watermark level.
Figure 5.24: A/B Trigger Debouncing, example with ABMode=quad; example for encoder retracing
.
MAN073 12/2016 V1.059 of 117
5 Functionality
A
B
G r a y C o u n t e r
E n c o d e r C o u n t e r
I n t e r n a l T r i g g e r F w d
01230123210321230
0123456765432123415
A p p l i e d T r i g g e r F w d
A
E n c o d e r C o u n t e r
T r i g g e r F w d
T r i g g e r B k w d
01234
T r i g g e r F w d B k w d
5
A
E n c o d e r C o u n t e r
T r i g g e r F w d
T r i g g e r B k w d
012345679
T r i g g e r F w d B k w d
1 08
A/B Trigger Divider
if ABTriggerDivider>1 then not all internally generated triggers are applied to the camera logic.
E.g. If ABTriggerDivider=2, then every second trigger is applied to the camera (see Fig. 5.25).
Figure 5.25: A/B Trigger Divider, example with ABTriggerDivider=1, ABMode=quad
A Only Trigger
The camera supports the use of simple incremental decoders that only provide one input, by
enabling the property ABTriggerAOnly. The B-signal is ignored in this mode and information
about direction of the object movement is not available: if ABTriggerAOnly is enabled then the
encoder position is always incremented. Detailed diagrams are shown in Fig. 5.26 and Fig.
5.27. Note that the quad mode is not available when ABTriggerAOnly=true.
Figure 5.26: AOnly Trigger in Single A/B Mode
Figure 5.27: AOnly Trigger in Double A/B Mode
Encoder Position
The internal ABTrigger signal before the ABTriggerDivider is processed for the Encoder
Position: every TriggerFwd pulse increments the Encoder Position and every TriggerBkwd pulse
decrements its value. For details refer to the diagram of the corresponding mode.
The Encoder Position value can be accessed through the EncoderPosition property or through
the status info that is inserted into the image (see Section 5.8).
60 of 117MAN073 12/2016 V1.0
5.4 Trigger and Strobe
By default the Encoder Position is only generated when TriggerMode=On and
TriggerSource=ABTrigger. When the property ABTriggerCountAlways=True, then the Encoder
Position is generated regardless of the trigger mode.
5.4.10Counter Reset by an External Signal
The image counter and the real time counter (timestamp) (see Section 5.8.1) can be reset by an
external signal. Both counters can be embedded into the image by the status line (see Section
5.8) or their register can be read out. These counters may be used to check that no images are
lost or to ease the synchronisation of multiple cameras.
The external signal to reset the above mentionend counters is selected by the property
Counter_ResetCounterSource. Available choices are PLC_Q4 to PLC_Q7 (see Section 7.10), Line1
(ISO_IN1) and ExposureStart. ExposureStart resets the counters at the start of an exposure.
The property Counter_ResetCounterMode determines how often the selected source should reset
the counters. The setting Once works together with the property
Counter_ResetCounterOnNextTrigger.
If Counter_ResetCounterMode=Once, then the counters are reset on the next active edge of the
selected reset source (property Counter_ResetCounterSource) after the device is armed with
Counter_ResetCounterOnNextTrigger=True. The register Counter_ResetCounterOnNextTrigger is reset
after the resetting trigger is received.
The setting Counter_ResetCounterMode=Continuous resets the counters on every occurrence of an
active edge of the reset source without the requirement to arm the device first. This setting is
suited if the reset source signal is different than the camera trigger.
The active edge of the reset input can be set by the property Counter_ResetCounterSourceInvert.
If set to True, then the rising edge is the active edge, else the falling edge.
Counter reset by an external signal is important if you would like to synchronize
multiple cameras. One signal is applied to all cameras which resets the counters simultaneously. The timestamps of all cameras are then theoretically synchronous with each other. In practice every camera runs on its own clock source
which has a precision of +/- 30 ppm and therefore the values of the timestamp
(real time counter) of the cameras may diverge with time. If this is an issue, then
the counters could be reset periodically by the external signal.
Reset of Individual Counters (ResetCounter_Dual)
If the property ResetCounter_Dual is set to False or if this property is not available, then the
ResetCounter settings apply to the image counter and to the real time counter together.
If ResetCounter_Dual is set to True then CounterReset can be set separately for the image
counter and for the real time counter. In this case the settings without ’RTC’ are applied to the
image counter and the settings with ’RTC’ in its name are applied to the real time counter.
.
MAN073 12/2016 V1.061 of 117
5 Functionality
T r i g g e r A c q u i s i t i o n _ S t a r t
T r i g g e r I n
A p p l i e d T r i g g e r
I n t e r n a l T r i g g e r E n a b l e
d i s a b l e de n a b l e dd i s a b l e d
5.4.11Trigger Acquisition
The applied trigger can be enabled or disabled by one or two external signals in the
TriggerAcquisition mode. This mode works with free-running (internal) trigger and external
trigger.
The property TriggerAcquisition_Enable enables the TriggerAcquisition mode.
Level Triggered Trigger Acquisition
The Level Triggered mode is enabled by setting TriggerAcquisition_Mode to Level and
TriggerAcquisition_Enable=True. A signal acts as a trigger enable (see Fig. 5.28). This signal is
selected by TriggerAcquisition_StartSource. A high signal level enables triggering of the
camera and a low signal level disables all triggers.
To invert the TriggerAcquisition signal use one of the PLC_Q signal and select
the inverted signal as its source. Table 5.5 shows a setting that uses ISO_IN0 as
trigger enable signal: the inverted signal is used as ISO_IN0 is inverted in the
input logic (see Fig. 7.4).
The Edge Triggered mode is enabled by setting TriggerAcquisition_Mode to Edge and
TriggerAcquisition_Enable=True. Two signals act as trigger enable (see Fig. 5.29). A rising edge
on the start signal enables triggering. A rising edge on the stop signal disables all triggers. The
start/stop signals are selected by TriggerAcquisition_StartSource and
TriggerAcquisition_StopSource.
The TriggerAcquisition feature can also be used with free-running trigger (TriggerMode=Off).
TriggerAcquisition enables or disables in this case the generation of the free-running trigger.
5.4.12Strobe Output
The strobe output is an isolated output located on the power supply connector that can be
used to trigger a strobe. The strobe output can be used both in free-running and in trigger
mode. There is a programmable delay available to adjust the strobe pulse to your application.
The strobe output needs a separate power supply. Please see Section 6.8, Fig.
5.16 and Fig. 5.17 for more information.
MAN073 12/2016 V1.063 of 117
5 Functionality
I m a g e S e n s o r
D i g i t a l O f f s e t
D i g i t a l G a i n
L a s e r L i n e D e t e c t o r
( C O G )
I m a g e o u t p u t
T e s t i m a g e s i n s e r t i o n
( r a m p / L F S R )
L a s e r t e s t i m a g e
i n s e r t i o n
C o l u m n F P N
C o r r e c t i o n
5.5Data Path Overview
The data path is the path of the image from the output of the image sensor to the output of
the camera. The sequence of blocks is shown in figure Fig. 5.30.
Figure 5.30: camera data path
.
64 of 117MAN073 12/2016 V1.0
5.6 Column FPN Correction
5.6Column FPN Correction
The camera contains a correction to decrease the Column Fixed Pattern Noise (FPN) of the
sensor. By default the Column FPN Correction is enabled.
The Column FPN Correction of the camera is correctly calibrated at the Photonfocus production
facility. Although a new calibration is normally not required, the user can recalibrate the
Column FPN Correction. No light should be applied to the camera during calibration. The
average grey value of every column of the image is calculated. The difference to the average
grey value of the whole image is then calculated for every column and stored in internal
camera memory. After calibration, this difference is then subtracted column-wise from every
image to reduce the Column FPN. Detailed instructions on the calibration of the Column FPN
Correction is given in Section 7.7.
5.7Gain and Offset
There are different gain settings on the camera:
Gain (Digital Fine Gain) Digital fine gain accepts fractional values from 0.01 up to 15.99. It is
implemented as a multiplication operation. Colour camera models only: There is
additionally a gain for every RGB colour channel. The RGB channel gain is used to
calibrate the white balance in an image, which has to be set according to the current
lighting condition.
Digital Gain Digital Gain is a coarse gain with the settings x1, x2, x4 and x8. It is implemented
as a binary shift of the image data where ’0’ is shifted to the LSB’s of the gray values. E.g.
for gain x2, the output value is shifted by 1 and bit 0 is set to ’0’.
The resulting gain is the product all gain values, which means that the image data is multiplied
in the camera by this factor.
Digital Fine Gain and Digital Gain may result in missing codes in the output image data.
A user-defined value can be subtracted from the gray value in the digital offset block. If digital
gain is applied and if the brightness of the image is too big then the interesting part of the
output image might be saturated. By subtracting an offset from the input of the gain block it
is possible to avoid the saturation.
5.8Image Information and Status Information
There are camera properties available that give information about the acquired images, such
as an image counter and the number of missed trigger signals. These properties can be queried
by software.
5.8.1Counters
Image counter The image counter provides a sequential number of every image that is output.
After camera startup, the counter counts up from 0 (counter width 24 bit). The counter
can be reset by the camera control software.
Real Time counter (Time stamp) The time counter starts at 0 after camera start, and counts
real-time in units of 1 micro-second. The time counter can be reset by the software in the
SDK (Counter width 32 bit).
MAN073 12/2016 V1.065 of 117
5 Functionality
Missed trigger counter The missed trigger counter counts trigger pulses that were ignored by
the camera because they occurred within the exposure or read-out time of an image. In
free-running mode it counts all incoming external triggers (counter width 8 bit / no wrap
around).
Missed burst trigger counter The missed burst trigger counter counts trigger pulses that were
ignored by the camera in the burst trigger mode because they occurred while the camera
still was processing the current burst trigger sequence.
Missed FrameCombine trigger counter Counts missed triggers due to the FrameCombine
feature (see also Section 5.2.10). A missed FrameCombine trigger can occur if a trigger is
applied while filler rows are added to a frame due to a FrameCombine timeout.
5.8.2Status Information
Status information is inserted in the 4 LSB in the last 3D data row (see bits labeled STAT in Fig.
5.6). LSB are transmitted first (see Table 5.6). The status information is divided in fields of 32
bits each, where every information field corresponds to one information parameter (see Table
STAT[23:0]IMG_CNT[23:0]Image counter (see also Section 5.8.1)
STAT[63:32]RT_CNT[31:0]Real time counter (time stamp) (see also Section
5.8.1)
STAT[87:64]ENC_POS[23:0]Encoder position (see also Section 5.4.9)
STAT[103:96]M_TRIG[7:0]Missed trigger counter (see also Section 5.8.1)
STAT[135:128]M_BURST_TRIG[7:0]Missed burst trigger counter (see also Section 5.8.1)
STAT[167:160]M_FC_TRIG[7:0]Missed FrameCombine trigger counter (see also
Section 5.8.1)
STAT[195:192]M_TRIG_LEVEL[3:0]Trigger Level: signal level of the trigger input signal
(only available in some models, see Appendix B). Bit
0: PLC_Q4: Bit 1: Line1; Bit 2: PLC_Q6 (A-Trigger); Bit
3: PLC_Q7 (B-Trigger).
Table 5.7: Status fields
.
66 of 117MAN073 12/2016 V1.0
5.9 Laser test image
5.9Laser test image
A Laser Test Image has been added that resembles a moving laser line (see Fig. 5.31) and it is
placed just before the peak detection. Therefore it can be used to test if the 3D data is
correctly processed during application development.
Figure 5.31: Laser test image
5.10Test Images
Test images are generated in the camera FPGA, independent of the image sensor. They can be
used to check the transmission path from the camera to the acquisition software. Independent
from the configured grey level resolution, every possible grey level appears the same number
of times in a test image. Therefore, the histogram of the received image must be flat.
A test image is a useful tool to find data transmission errors or errors in the access
of the image buffers by the acquisition software.
The analysis of the test images with a histogram tool gives a flat histogram only
if the image width is a multiple of 256 (in 8 bit mode).
.
MAN073 12/2016 V1.067 of 117
5 Functionality
5.10.1Ramp
Depending on the configured grey level resolution, the ramp test image outputs a constant
pattern with increasing grey level from the left to the right side (see Fig. 5.32).
Figure 5.32: Ramp test images: 8 bit output
5.10.2LFSR
The LFSR (Linear Feedback Shift Register) test image outputs a constant pattern with a
pseudo-random grey level sequence containing every possible grey level that is repeated for
every row. The LFSR test pattern was chosen because it leads to a very high data toggling rate,
which stresses the interface electronic and the cable connection.
Figure 5.33: LFSR (linear feedback shift register) test image
In the histogram you can see that the number of pixels of all grey values are the same.
Please refer to application note [AN026] for the calculation and the values of the LFSR test
image.
68 of 117MAN073 12/2016 V1.0
5.10 Test Images
5.10.3Troubleshooting using the LFSR
To control the quality of your complete imaging system enable the LFSR mode, set the camera
window to 1024 x 1024 pixels (x=0 and y=0) and check the histogram. The camera window can
also be set to a multiple of this resolution (e.g. 2048 x 2048 or 4096 x 3072) if the camera
model supports this resolution. If your image acquisition application does not provide a
real-time histogram, store the image and use an image viewing tool (e.g. ImageJ) to display
the histogram.
In the LFSR (linear feedback shift register) mode the camera generates a constant
pseudo-random test pattern containing all grey levels. If the data transmission is correctly
received, the histogram of the image will be flat (Fig. 5.34). On the other hand, a non-flat
histogram (Fig. 5.35) indicates problems, that may be caused either by a defective camera, by
problems in the acquisition software or in the transmission path.
Figure 5.34: LFSR test pattern received and typical histogram for error-free data transmission
Figure 5.35: LFSR test pattern received and histogram containing transmission errors
In robots applications, the stress that is applied to the camera cable is especially high due to
the fast movement of the robot arm. For such applications, special drag chain capable cables
are available. Please contact the Photonfocus Support for consulting expertise.
MAN073 12/2016 V1.069 of 117
5 Functionality
70 of 117MAN073 12/2016 V1.0
6
E t h e r n e t J a c k ( R J 4 5 )
P o w e r S u p p l y
a n d I / O C o n n e c t o r
S t a t u s L E D
Hardware Interface
6.1GigE Connector
The GigE cameras are interfaced to external components via
•an Ethernet jack (RJ45) to transmit configuration, image data and trigger.
•a 12 pin subminiature connector for the power supply, Hirose HR10A-10P-12S (female) .
The connectors are located on the back of the camera. Fig. 6.1 shows the plugs and the status
LED which indicates camera operation.
Figure 6.1: Rear view of the GigE camera
6.2Power Supply Connector
The camera requires a single voltage input (see Table 4.4). The camera meets all performance
specifications using standard switching power supplies, although well-regulated linear power
supplies provide optimum performance.
It is extremely important that you apply the appropriate voltages to your camera.
Incorrect voltages will damage the camera.
MAN073 12/2016 V1.071 of 117
6 Hardware Interface
A suitable power supply can be ordered from your Photonfocus dealership.
For further details including the pinout please refer to Appendix A.
6.3Status Indicator (GigE cameras)
A dual-color LED on the back of the camera gives information about the current status of the
GigE CMOS cameras.
LED GreenIt blinks slowly when the camera is not grabbing images.When the camera is
grabbing images the LED blinks at a rate equal to the frame rate. At slow
frame rates, the LED blinks. At high frame rates the LED changes to an
apparently continuous green light, with intensity proportional to the ratio of
readout time over frame time.
LED RedRed indicates an active serial communication with the camera.
Table 6.1: Meaning of the LED of the GigE CMOS cameras
6.4Absolute Maximum Ratings
ParameterValue
Power Supply Voltage26.4 V
ESD Contact Discharge Power Supply4 kV
ESD Air Discharge Power Supply8 kV
Fast Transients/Bursts Power Supply2 kV
Camera Control Input Signal Voltage Single Ended-30 V ... +30 V
Camera Control Input Signal Voltage RS422-25 V ... +25 V
Camera Control Input Signal Voltage HTL10 V ... 30 V
Common Mode Range Voltage RS422-10 V ... +13 V
Camera Control Output Signal Voltage Single Ended0 V ... +30 V
Camera Control Output Signal Output Current Single Ended0.5 A
Camera Control Output Signal Output Power Single Ended0.5 W
ESD Contact Discharge Camera Control Signals4 kV
ESD Air Discharge Camera Control Signals8 kV
Fast Transients/Bursts Data and Camera Control Signals1 kV
Surge immunity Data and Camera Control Signals1 kV
Table 6.2: Absolute Maximum Ratings
.
72 of 117MAN073 12/2016 V1.0
6.5 Electrical Characteristics
6.5Electrical Characteristics
ParameterValue
Camera Power Supply+12 V (-10%) ... +24 V (+10%)
Camera Control Input Single Ended+5 V ... +30 V
Camera Control Input RS422 Receiver Sensitivity+/- 200 mV
Camera Control Input RS422 Maximum Common Mode Range-7 V ... +7 V
Camera Control Input RS422 Maximum Differential Input Level10 V
Camera Control Input RS422 Minimum Differential Input Level2 V
ISO power RS-4225 V (-10%) ... 24 V (+10%)
ISO power HTL10 V (-10%) ... 30V (+10%)
Table 6.3: Electrical Characteristics
6.6Power and Ground Connection for GigE G2 Cameras
The interface electronics is isolated from the camera electronics and the power supply
including the line filters and camera case. Fig. 6.2 shows a schematic of the power and ground
connections in the G2 camera models.
.
MAN073 12/2016 V1.073 of 117
6 Hardware Interface
P o w e r S u p p l y
2
P O W E R _ R E T U R N
1
C A S E
G N D
I n t e r n a l P o w e r S u p p l y
D C / D C
V C C _ 3
+
P O W E R
R X R S 4 2 2
I S O _ I N C 0 _ P
I S O _ I N C 0 _ N
I S O _ I N C 1 _ P
I S O _ I N C 1 _ N
I S O _ I N 0
I S O _ I N 1
I S O _ O U T 0
I S O _ O U T 1
I s o l a t e d I n t e r f a c e
C a m e r a E l e c t r o n i c
I S O L A T O R
I S O _ G N D
I S O _ P W R
1 2
1 2 p o l . H i r o s e C o n n e c t o r
6
8
3
9
7
1 0
1 1
4
5
+
I / O a n d T r i g g e r I n t e r f a c e
D C / D C
D C / D C
V C C _ 2
V C C _ 1
E S D
P r o t e c t i o n
E S D
P r o t e c t i o n
C a m e r a E l e c t r o n i c
L i n e
F i l t e r
Y O U R _ G N D
Y O U R _ P W R
+
H i r o s e C o n n e c t o r
C A S E
G N D
C a m e r a
Figure 6.2: Schematic of power and ground connections in G2 camera models
74 of 117MAN073 12/2016 V1.0
6.7 Power and Ground Connection for GigE H2 Cameras
P o w e r S u p p l y
2
P O W E R _ R E T U R N
1
C A S E
G N D
I n t e r n a l P o w e r S u p p l y
D C / D C
V C C _ 3
+
P O W E R
R X H T L
I S O _ I N C 0 _ P
I S O _ I N C 0 _ N
I S O _ I N C 1 _ P
I S O _ I N C 1 _ N
I S O _ I N 0
I S O _ I N 1
I S O _ O U T 0
I S O _ O U T 1
I s o l a t e d I n t e r f a c e
C a m e r a E l e c t r o n i c
I S O L A T O R
I S O _ G N D
I S O _ P W R
1 2
1 2 p o l . H i r o s e C o n n e c t o r
6
8
3
9
7
1 0
1 1
4
5
+
I / O a n d T r i g g e r I n t e r f a c e
D C / D C
D C / D C
V C C _ 2
V C C _ 1
E S D
P r o t e c t i o n
E S D
P r o t e c t i o n
C a m e r a E l e c t r o n i c
L i n e
F i l t e r
Y O U R _ G N D
Y O U R _ P W R
+
H i r o s e C o n n e c t o r
C A S E
G N D
C a m e r a
6.7Power and Ground Connection for GigE H2 Cameras
The interface electronics is isolated from the camera electronics and the power supply
including the line filters and camera case. Fig. 6.3 shows a schematic of the power and ground
connections in H2 camera models.
Figure 6.3: Schematic of power and ground connections in H2 camera models
.
MAN073 12/2016 V1.075 of 117
6 Hardware Interface
6.8Trigger and Strobe Signals for GigE Cameras
6.8.1Overview
The 12-pol. Hirose power connector contains two external trigger inputs, two strobe outputs
and two differential inputs (G2 models: RS-422, H2 models: HTL). All inputs and outputs are
connected to the Programmable Logic Controller (PLC) (see also Section 6.9) that offers
powerful operations.
The pinout of the power connector is described in Section A.1.
G2 models: ISO_INC0 and ISO_INC1 RS-422 inputs have -10 V to +13 V extended
common mode range.
H2 models: The voltage level for the HTL interface should be given by the user by
means of connecting the encoder power pin (HTL_ENC_PWR) and the ISO_PWR
pin to the same power supply within a range between 10 and 30V. In the same
way, encoder ground (HTL_ENC_GND) and ISO_GND signals should be connected
to the same ground in order to guarantee the good reception of the differential
signals.
ISO_OUT0 and ISO_OUT1 have different output circuits (see also Section 6.8.2).
A suitable trigger breakout cable for the Hirose 12 pol. connector can be ordered
from your Photonfocus dealership.
Simulation with LTSpice is possible, a simulation model can be downloaded from
our web site www.photonfocus.com on the software download page (in Support
section). It is filed under "Third Party Tools".
Don’t connect single-ended signals to the differential inputs ISO_INC0 and
ISO_INC1.
Fig. 6.4 shows the schematic of the inputs and outputs for the G2 models and Fig. 6.5 for the
H2 models. All inputs and outputs are isolated. ISO_VCC is an isolated, internally generated
voltage.
.
76 of 117MAN073 12/2016 V1.0
6.8 Trigger and Strobe Signals for GigE Cameras
I S O _ G N D
R X R S 4 2 2
I S O _ I N C 0 _ P
I S O _ I N C 0 _ N
M A X 3 0 9 8
I S O _ I N C 1 _ P
I S O _ I N C 1 _ N
I S O _ G N D
I S O _ V C C
e n h a n c e d
P o w e r F E T
4 . 7 V
1 0 k
I S O _ I N 0
G N D
I S O _ G N D
I S O _ V C C
e n h a n c e d
P o w e r F E T
4 . 7 V
1 0 k
I S O _ I N 1
I S O _ G N D
I S O _ P W R
P o w e r
M O S F E T
I S O _ O U T 0
P T C
4 k 7
M a x . 3 0 V
M a x . 0 . 5 A
M a x . 0 . 5 W
I S O _ G N D
P o w e r
M O S F E T
I S O _ O U T 1
P T C
M a x . 3 0 V
M a x . 0 . 5 A
M a x . 0 . 5 W
I s o l a t e d I n t e r f a c e
C a m e r a E l e c t r o n i c
- 1 0 V t o + 1 3 V e x t e n d e d
C o m m o n M o d e R a n g e
I S O L A T O R
I S O _ G N D
I S O _ P W R
1 2
1 2 p o l . H i r o s e C o n n e c t o r
6
8
3
9
7
1 0
1 1
4
5
+
+
+
+
C a m e r a
M i n . - 3 0 V
M a x . 3 0 V
M i n . - 3 0 V
M a x . 3 0 V
I S O _ V C C
+
Figure 6.4: Schematic of inputs and output (G2 models)
MAN073 12/2016 V1.077 of 117
6 Hardware Interface
I S O _ G N D
R X H T L : i n p u t r a n g e : 1 0 V t o 3 0 V
I S O _ I N C 0 _ P
I S O _ I N C 0 _ N
I S O _ I N C 1 _ P
I S O _ I N C 1 _ N
I S O _ G N D
I S O _ V C C
e n h a n c e d
P o w e r F E T
4 . 7 V
1 0 k
I S O _ I N 0
G N D
I S O _ G N D
I S O _ V C C
e n h a n c e d
P o w e r F E T
4 . 7 V
1 0 k
I S O _ I N 1
I S O _ G N D
I S O _ P W R
P o w e r
M O S F E T
I S O _ O U T 0
P T C
4 k 7
M a x . 3 0 V
M a x . 0 . 5 A
M a x . 0 . 5 W
I S O _ G N D
P o w e r
M O S F E T
I S O _ O U T 1
P T C
M a x . 3 0 V
M a x . 0 . 5 A
M a x . 0 . 5 W
I s o l a t e d I n t e r f a c e
C a m e r a E l e c t r o n i c
I S O L A T O R
I S O _ G N D
I S O _ P W R
1 2
1 2 p o l . H i r o s e C o n n e c t o r
6
8
3
9
7
1 0
1 1
4
5
+
+
+
+
C a m e r a
M i n . - 3 0 V
M a x . 3 0 V
M i n . - 3 0 V
M a x . 3 0 V
I S O _ V C C
+
H T L _ E N C _ P W R
H T L _ E N C _ G N D
c o n n e c t t o :
H T L i n p u t r a n g e : 1 0 V . . . 3 0 V
Figure 6.5: Schematic of inputs and output (H2 models)
78 of 117MAN073 12/2016 V1.0
6.8 Trigger and Strobe Signals for GigE Cameras
I S O _ G N D
I S O _ G N D
I S O _ V C C
e n h a n c e d
P o w e r F E T
4 . 7 V
1 0 k
I S O _ I N 0
C a m e r a
7
1 2 p o l . H i r o s e
C o n n e c t o r
I S O _ G N D
1 2
Y O U R _ G N D
Y O U R _ G N D
I n p u t V o l t a g e
M a x . + 3 0 V D C
M i n . - 3 0 V D C
+
I S O _ G N D
I S O _ G N D
I S O _ V C C
e n h a n c e d
P o w e r F E T
4 . 7 V
1 0 k
I S O _ I N 0
C a m e r a
7
1 2 p o l . H i r o s e
C o n n e c t o r
I S O _ G N D
1 2
Y O U R _ G N D
Y O U R _ G N D
C o n t r o l L o g i c
&
Y O U R _ V C C
+
+
6.8.2Single-ended Inputs
ISO_IN0 and ISO_IN1 are single-ended isolated inputs. The input circuit of both inputs is
identical (see Fig. 6.4).
Fig. 6.6 shows a direct connection to the ISO_IN inputs.
In the camera default settings the PLC is configured to connect the ISO_IN0 to
the PLC_Q4 camera trigger input. This setting is listed in Section 7.10.2.
Figure 6.6: Direct connection to ISO_IN
Fig. 6.7 shows how to connect ISO_IN to TTL logic output device.
Figure 6.7: Connection to ISO_IN from a TTL logic device
.
MAN073 12/2016 V1.079 of 117
6 Hardware Interface
I S O _ G N D
I S O _ P W R
P o w e r
M O S F E T
I S O _ O U T 0
P T C
4 k 7
C a m e r a
3
1 2 p o l . H i r o s e
C o n n e c t o r
I S O _ G N D
1 2
Y O U R _ G N D
I S O _ P W RY O U R _ P W R
Y O U R _ G N D
C o n t r o l L o g i c
&
Y O U R _ P W R
+
+
+
+
6
M a x . 3 0 V
M a x . 0 . 5 A
M a x . 0 . 5 W
I S O _ G N D
P o w e r
M O S F E T
I S O _ O U T 1
P T C
C a m e r a
8
1 2 p o l . H i r o s e
C o n n e c t o r
I S O _ G N D
1 2
Y O U R _ G N D
Y O U R _ G N D
C o n t r o l L o g i c
&
Y O U R _ P W R
+
4 k 7
+
Y O U R _ P W R
M a x . 3 0 V
M a x . 0 . 5 A
M a x . 0 . 5 W
6.8.3Single-ended Outputs
ISO_OUT0 and ISO_OUT1 are single-ended isolated outputs.
ISO_OUT0 and ISO_OUT1 have different output circuits: ISO_OUT1 doesn’t have
a pullup resistor and can be used as additional Strobe out (by adding Pull up) or
as controllable switch. Maximal ratings that must not be exceeded: voltage: 30
V, current: 0.5 A, power: 0.5 W.
Fig. 6.8 shows the connection from the ISO_OUT0 output to a TTL logic device. PTC is a current
limiting device.
Figure 6.8: Connection example to ISO_OUT0
Fig. 6.9 shows the connection from ISO_OUT1 to a TTL logic device. PTC is a current limiting
device.
Figure 6.9: Connection from the ISO_OUT1 output to a TTL logic device
.
80 of 117MAN073 12/2016 V1.0
Fig. 6.10 shows the connection from ISO_OUT1 to a LED.
Y O U R _ P W R
I S O _ G N D
P o w e r
M O S F E T
I S O _ O U T 1
P T C
R
C a m e r a
8
1 2 p o l . H i r o s e
C o n n e c t o r
I S O _ G N D
1 2
Y O U R _ G N D
+
Y O U R _ P W R
I S O _ G N D
P o w e r
M O S F E T
I S O _ O U T 1
P T C
L
C a m e r a
8
1 2 p o l . H i r o s e
C o n n e c t o r
I S O _ G N D
1 2
Y O U R _ G N D
Y O U R _ P W R
L
D
D
D
1
2
M a x . 3 0 V
M a x . 0 . 5 A
M a x . 0 . 5 W
+
+
R e s p e c t t h e l i m i t s o f t h e P O W E R M O S F E T !
Figure 6.10: Connection from ISO_OUT1 to a LED
Respect the limits of the POWER MOSFET in the connection to ISEO_OUT1. Maximal ratings that must not be exceeded: voltage: 30 V, current: 0.5 A, power: 0.5
W. (see also Fig. 6.11). The type of the Power MOSFET is: International Rectifier
IRLML0100TRPbF.
6.8 Trigger and Strobe Signals for GigE Cameras
Figure 6.11: Limits of ISO_OUT1 output
.
MAN073 12/2016 V1.081 of 117
6 Hardware Interface
R X R S 4 2 2
I S O _ I N C x _ P
I S O _ I N C x _ N
1 2 p o l . H i r o s e
C o n n e c t o r
Y O U R _ G N D
5 V T T L L o g i c L e v e l
C a m e r a
I S O _ G N D
I S O _ P W R
P o w e r
M O S F E T
I S O _ O U T 0
P T C
4 k 7
I S O _ G N D
I S O _ G N D
I S O _ V C C
e n h a n c e d
P o w e r F E T
4 . 7 V
1 0 k
I S O _ I N 0
M a s t e r C a m e r a
S l a v e C a m e r a
3
7
H i r o s e
C o n n e c t o r s
+
+
I S O _ G N DI S O _ G N D
1 21 2
6
I S O _ P W R
6.8.4Differential RS-422 Inputs (G2 models)
ISO_INC0 and ISO_INC1 are isolated differential RS-422 inputs (see also Fig. 6.4). They are
connected to a Maxim MAX3098 RS-422 receiver device. Please consult the data sheet of the
MAX3098 for connection details.
Don’t connect single-ended signals to the differential inputs ISO_INC0 and
ISO_INC1 (see also Fig. 6.12).
Figure 6.12: Incorrect connection to ISO_INC inputs
6.8.5Master / Slave Camera Connection
The trigger input of one Photonfocus G2 camera can easily connected to the strobe output of
another Photonfocus G2 camera as shown in Fig. 6.13. This results in a master/slave mode
where the slave camera operates synchronously to the master camera.
Figure 6.13: Master / slave connection of two Photonfocus G2 cameras
.
82 of 117MAN073 12/2016 V1.0
6.8 Trigger and Strobe Signals for GigE Cameras
I s o l a t o r
C A M _ G N D
I S O _ I N
I S O _ G N D
G r o u n d p l a n e v o l t a g e d i f f e r e n c e
I S O _ G N D
S e p a r a t e g r o u n d
n o g r o u n d l o o p
6.8.6I/O Wiring
The Photonfocus cameras include electrically isolated inputs and outputs. Take great care when
wiring trigger and strobe signals to the camera, specially over big distances (a few meters) and
in noisy environments. Improper wiring can introduce ground loops which lead to malfunction
of triggers and strobes.
There are two roads to avoid ground loops:
•Separating I/O ground and power supply (ISO_GND and ISO_PWR) from camera power
(CAM_GND, CAM_PWR)
•Using a common power supply for camera and I/O signals with star-wiring
Separate Grounds
To separate the signal and ground connections of the camera (CAM_GND, CAM_PWR, data
connections) from the I/O connections (ISO_GND, ISO_PWR, ISO_IN, ISO_OUT) is one way to
avoid ground loops. Fig. 6.14 shows a schematic of this setup. In this setup the power supplies
for the camera and for ISO power must be separate devices.
Figure 6.14: I/O wiring using separate ground
.
MAN073 12/2016 V1.083 of 117
6 Hardware Interface
D e v i c e 1
+
-
D e v i c e 2
+-
D e v i c e 3
+
-
D e v i c e 4
+
-
D e v i c e n
+-
. . .
S t a r P o i n t
G N DP W R
I s o l a t o r
C A M _ G N D
I S O _ I N
I S O _ G N D
S t a r w i r i n i g
n o g r o u n d l o o p
Common Grounds with Star Wiring
Ground loops can be avoided using "star wiring", i.e. the wiring of power and ground
connections originate from one "star point" which is typically a power supply. Fig. 6.15 shows
a schematic of the star-wiring concept.
Fig. 6.16 shows a schematic of the star-wiring concept applied to a Photonfocus GigE
camera.The power supply and ground connections for the camera and for the I/O are
connected to the same power supply which acts as the "Star Point".
Figure 6.15: Star-wiring principle
Figure 6.16: I/O wiring using star-wiring
.
84 of 117MAN073 12/2016 V1.0
6.8 Trigger and Strobe Signals for GigE Cameras
P o w e r S u p p l y
+
-
C a m e r a
F l a s h
M a c h i n e V i s i o n
S y s t e m P C
E t h e r n e t D a t a C a b l e
S T R
+
-
C A M _ P W R
C A M _ G N D
I S O _ O U T
I S O _ P W R
I S O _ G N D
I S O _ I N
S t a r t P o i n t
+
-
L i g h t B a r r i e r
Fig. 6.17 shows an example of how to connect a flash light and a trigger source to the camera
using star-wiring. The trigger in this example is generated from a light barrier. Note how the
power and ground cables are connected to the same power supply.
Figure 6.17: I/O wiring using star-wiring example
.
MAN073 12/2016 V1.085 of 117
6 Hardware Interface
I s o l a t o r
C A M _ G N D
I S O _ I N
I S O _ G N D
G r o u n d p l a n e v o l t a g e d i f f e r e n c e
C o n n e c t i n g C A M _ G N D a n d
I S O _ G N D t h e w r o n g w a y
G r o u n d l o o p
G r o u n d l o o p
An example of improper wiring that causes a ground loop is shown in Fig. 6.18.
Figure 6.18: Improper I/O wiring causing a ground loop
86 of 117MAN073 12/2016 V1.0
6.9 PLC connections
6.9PLC connections
The PLC (Programmable Logic Controller) is a powerful device where some camera inputs and
outputs can be manipulated and software interrupts can be generated. Sample settings and an
introduction to PLC are shown in Section 7.10. PLC is described in detail in the document [PLC].
NameDirectionDescription
A0 (Line0)Power connector -> PLCISO_IN0 input signal
A1(Line1)Power connector -> PLCISO_IN1 input signal
A2 (Line2)Power connector -> PLCISO_INC0 input signal
A3 (Line3)Power connector -> PLCISO_INC1 input signal
A4camera head -> PLCFVAL (Frame Valid) signal
A5camera head -> PLCLVAL (Line Valid) signal
A6camera head -> PLCDVAL (Data Valid) signal
A7camera head -> PLCReserved (CL_SPARE)
Q0PLC ->not connected
Q1PLC -> power connectorISO_OUT1 output signal (signal is inverted)
Q2PLC ->not connected
Q3PLC ->not connected
Q4PLC -> camera headPLC_Q4 camera trigger
Q5PLC -> camera headReserved for future use
Q6PLC -> camera headIncremental encoder A signal
Q7PLC -> camera headIncremental encoder B signal
Table 6.4: Connections to/from PLC
MAN073 12/2016 V1.087 of 117
6 Hardware Interface
88 of 117MAN073 12/2016 V1.0
7
Software
7.1Software for MV1-D1280-L01-3D05
Various software packages for Photonfocus the MV1-D1280-L01-3D05 camera series are
available on the Photonfocus website:
eBUS SDK Contains the Pleora SDK and the Pleora GigE filter drivers. Many examples of the
SDK are included.
PFInstaller Contains the PF_GEVPlayer, the PF 3D Suite and SDK, a property list for every GigE
camera and additional documentation and examples.
PF 3D Suite and SDK Visualization tool for Photonfocus 3D cameras. This tool is described in a
separate manual [MAN053] and is included in the PFInstaller.
7.2PF_GEVPlayer
The camera parameters can be configured by a Graphical User Interface (GUI) tool for Gigabit
Ethernet Vision cameras or they can be programmed with custom software using the SDK.
A GUI tool that can be downloaded from Photonfocus is the PF_GEVPlayer. How to obtain and
install the software and how to connect the camera is described in Chapter 3.
After connecting to the camera, the camera properties can be accessed by clicking on the GEVDevice control button (see also Section 7.2.2).
The PF_GEVPlayer is described in more detail in the GEVPlayer Quick Start Guide
[GEVQS] which is included in the PFInstaller.
There is also a GEVPlayer in the Pleora eBUS package. It is recommended to
use the PF_GEVPlayer as it contains some enhancements for Photonfocus GigE
cameras such as decoding the image stream in DR1 cameras.
MAN073 12/2016 V1.089 of 117
7 Software
7.2.1PF_GEVPlayer main window
After connecting the camera (see Chapter 3), the main window displays the following controls
(see Fig. 7.1):
Disconnect Disconnect the camera
Mode Acquisition mode
Play Start acquisition
Stop Stop acquisition
Acquisition Control Mode Continuous, Single Frame or Multi Frame modes. The number of
frames that are acquired in Multi Frame mode can be set in the GEV Device Control with
AcquisitionFrameCount in the AcquisitionControl category.
Communication control Set communication properties.
GEV Device control Set properties of the camera head, IP properties and properties of the PLC
(Programmable Logic Controller, see also Section 6.9 and document [PLC]).
Image stream control Set image stream properties and display image stream statistics.
Figure 7.1: PF_GEVPlayer main window
Below the image display there are two lines with status information
7.2.2GEV Control Windows
This section describes the basic use of the GEV Control windows, e.g. the GEV Device Control
window.
The view of the properties in the control window can be changed as described below. At start
the properties are grouped in categories which are expanded and whose title is displayed in
bold letters. An overview of the available view controls of the GEV Control windows is shown
in Fig. 7.2.
90 of 117MAN073 12/2016 V1.0
7.2 PF_GEVPlayer
T o g g l e c a t e g o r y /
a l p h a b e t i c a l v i e w
E x p a n d a l l
c a t e g o r i e s
C o l l a p s e a l l
c a t e g o r i e s
V i s i b i l i t y
s e l e c t i o n
E x p a n d
c a t e g o r y
C o l l a p s e
c a t e g o r y
P r o p e r t y
d e s c r i p t i o n
P a r a m e t e r
e r r o r
i n d i c a t i o n
To have a quick overview of the available categories, all categories should be collapsed. The
categories of interest can then be expanded again. If the name of the property is known, then
the alphabetical view is convenient. If this is the first time that you use a Photonfocus GigE
camera, then the visibility should be left to Beginner.
The description of the currently selected property is shown at the bottom ot the window.
After selecting a property from a drop-down box it is necessary to press <Enter>
or to click with the mouse on the control window to apply the property value to
the camera.
A red cross at the upper right corner of the GEV Control Window indicates a
parameter error, i.e. a parameter is not correctly set. In this case you should
check all properties. A red exclamation mark (!) at the right side of a parameter
value indicates that this parameters has to be set correctly.
Figure 7.2: PF_GEVPlayer Control Window
.
MAN073 12/2016 V1.091 of 117
7 Software
7.2.3Display Area
The images are displayed in the main window in the display area. A zoom menu is available
when right clicking in the display area. Another way to zoom is to press the Ctrl button while
using the mouse wheel.
7.2.4White Balance (Color cameras only)
A white balance utility is available in the PF_GEVPlayer in Tools -> Image Filtering (see Fig.
7.3). The gain of the color channels can be adjusted manually by sliders or an auto white
balance of the current image can be set by clicking on the White Balance button. To have a
correct white balance setting, the camera should be pointed to a neutral reference (object that
reflects all colors equally), e.g. a special grey reference card while clicking on the White
Balance button.
The white balance settings that were made as described in this section, are applied by the PF_GEVPlayer software and are not stored in the camera. To store
the color gain values in the camera, the Gain settings in the GEV Device Control
(in AnalogControl) must be used. If the gain properties in the camera are used,
then the PF_GEVPlayer RGB Filtering should be disabled.
Figure 7.3: PF_GEVPlayer image filtering dialog
7.2.5Save camera setting to a file
The current camera settings can be saved to a file with the PF_GEVPlayer (File -> Save or Save
As...). This file can later be applied to camera to restore the saved settings (File -> Open), Note,
that the Device Control window must not be open to do this.
The MROI and LUT settings are not saved in the file.
92 of 117MAN073 12/2016 V1.0
7.3 Pleora SDK
7.2.6Get feature list of camera
A list of all features of the Photonfocus GigE cameras in HTML format can be found in the
GenICam_Feature_Lists sub-directory (in Start -> All Programs -> Photonfocus -> GigE_Tools).
Alternatively, the feature list of the connected camera can be retrieved with the PF_GEVPlayer
(Tools -> Save Camera Features as HTML...).
7.3Pleora SDK
The eBUS package provides the PureGEV C++ SDK for image acquisition and the setting of
properties. A help file is installed in the Pleora installation directory, e.g. C:\ProgramFiles\Pleora Technologies Inc\eBUS SDK\Documentation.
Various code samples are installed in the installation directory, e.g. C:\Program Files\PleoraTechnologies Inc\eBUS SDK\Samples. The sample PvPipelineSample is recommended to start with.
Samples that show how to set device properties are included in the PFInstaller that can be
downloaded from the Photonfocus webpage.
7.4Frequently used properties
The following list shows some frequently used properties that are available in the Beginner
mode. The category name is given in parenthesis.
Width (ImageFormatControl) Width of the camera image ROI (region of interest)
Height (ImageFormatControl) Width of the camera image ROI
OffsetX, OffsetY (ImageFormatControl) Start of the camera image ROI
ExposureTime (AcquisitionControl) Exposure time in microseconds
TriggerSource (AcquisitionControl) Trigger source if external triggered mode is selected
Header_Serial (Info / CameraInfo) Serial number of the camera
UserSetSave (UserSetControl) Saves the current camera settings to non-volatile flash memory.
7.5Height setting
The Height property must be set manually to the value of HeightInterface whenever a property
relevant to the height setting is modified (an example for this can be found in Section 7.6). The
height relevant properties are:
•LineFinder_Mode
•Peak0_3DH
•FrameCombine
The height can be directly written to the Height property in 2Donly mode.
MAN073 12/2016 V1.093 of 117
7 Software
7.63D (Laser Line Detector) settings
This section describes how to the set the 3D properties. These properties are described in
Section 5.2.
1.Set threshold value for laser line with property Peak0_Threshold (in category
LineFinder/Peak0) (see also note in Section 5.2.3).
2.Set scan area by setting the start row (Peak0_3DY in category LineFinder/Peak0) and the
height (Peak0_3DH in category LineFinder/Peak0) (see also Section 5.3.2).
3.Set threshold for laser line width calculation. Its value must not be smaller than
Peak0_Threshold.
4.Set LineFinder_Mode (in category LineFinder) to Mode_3Donly or to Mode_2Dand3D. Note that
Mode_3Donly should be selected for maximal frame rate.
5.If LineFinder_Mode is set to Mode_2Dand3D then skip steps 6 to 8 and continue at step 9.
6.The number of frames of the FrameCombine feature (FrameCombine_NrOfFrames) should be
set to a value that the resulting frame rate is below 200 for most applications (see also
Section 5.2.10). The resulting frame rate is the trigger rate divided by
FrameCombine_NrOfFrames. In free running mode (TriggerMode = Off) the frame rate can be
read from the property AcquisitionFrameRateMax (in category AcquisitionControl). The
lower the resulting frame rate, the fewer interrupts are generated by the GigE driver and
the less load is produced on the computer’s CPU. E.g. if the trigger rate is 4000 fps then
FrameCombine_NrOfFrames should be set to 20 or more.
7.If FrameCombine is used, then the parameter FrameCombine_Timeout (in microseconds)
should be set (see also Section 5.2.10. The value should be higher than the longest time
between triggers, e.g. if the trigger rate is constant, then it could be set to twice the time
between triggers.
8.If FrameCombine is used then FrameCombine_Enable should be set to True.
9.Read the value of the parameter HeightInterface and set Height to this value.
7.7Column FPN Correction
Due to the readout structure of the image sensors there is a column-wise fixed pattern noise
(FPN). The Column FPN Correction (ColCorrection) adds or subtracts a fixed value to a column.
These values are obtained by a calibration process. The ColCorrection of the camera was
calibrated at Photonfocus production facility.
7.7.1Enable / Disable the Column FPN Correction
The Column FPN Correction can be enabled or disabled with the property ColCorrection_Enable
(in category Correction/ColCorrection). By default the correction is enabled.
7.7.2Calibration of the Column FPN Correction
The Column FPN Correction of the camera is correctly calibrated at Photonfocus production
facility. Although a new calibration is normally not required, you can recalibrate the Column
FPN Correction with the following instructions:
1.Setup the camera to the mode where it will be usually used. (Exposure time, ...). The
width should be set to its maximal value. Due to the internal structure of the camera, best
performance of calibration will be achieved when calibrating under "real conditions".
94 of 117MAN073 12/2016 V1.0
7.7 Column FPN Correction
When the camera will be run in 3Donly mode, which is required for maximal frame rate,
then the following settings should be applied: LineFinder_Mode = Mode_3Donly (category
LineFinder), Peak0_3DY = 0 (category LineFinder / Peak0 / Peak0_3D), Peak0_3DH = 1024
(category LineFinder / Peak0 / Peak0_3D, Width = 1280 (category ImageFormatControl), Height =
4 (category ImageFormatControl)
If different exposure times will be used, calibrate the camera under the longest
exposure time.
2.Put the camera in free-running mode by setting the property TriggerMode to Off.
3.Start grabbing of the camera by clicking on the Play button.
4.Wait until the camera has achieved working temperature.
5.Close the lens of the camera or put a cap on the lens. The calibration requires an uniform
dark image. The black level offset should be set so that all pixels (except defect pixels)
have values above 0.
6.Set the visibility of the Device Control window of the PF_GEVPlayer to Expert.
7.Run the command ColCorrection_CalibrateBlack (category ColCorrection) by clicking on
the corresponding button. The camera transmits a test image during calibration.
8.Run the command ColCorrection_Update by clicking on the corresponding button. Read the
ColCorrection_Busy value which should be 0 after the calibration has finished. Repeat this
step if its value is not 0. If the ColCorrection_Busy value doesn’t show 0 after various tries,
check if the camera receive triggers or set the TriggerMode of the camera to Off.
9.Check the values of the properties ColCorrection_Overflow and ColCorrection_Underflow. Both
should have the value 0 after calibration. If ColCorrection_Overflow is not 0, then decrease
BlackLevel (in category AnalogControl) and re-run the procedure from step 6 on. If
ColCorrection_Underflow is not 0, then increase BlackLevel (in category AnalogControl) and
re-run the procedure from step 6 on.
10. The Column FPN correction is now calibrated. The calibration values are stored in the
camera’s RAM and these values are lost when the camera power is turned off. To store the
calibration values to permanent memory see Section 7.7.3.
7.7.3Storing the calibration in permanent memory
After running the calibration procedure (see Section 7.7.2) the calibration values are stored in
RAM. When the camera is turned off, their values are lost.
To prevent this, the calibration values must be stored in flash memory. This can be done by
clicking on the property ColCorrection_SaveToFlash (in category Calibration). Wait until the
command has been finished, i.e.the property ColCorrection_Busy (category Correction /
ColCorrection) is 0. ColCorrection_Busy can be updated by clicking on the property
ColCorrection_Update (in category Calibration).
Storing the calibration in permanent memory overwrites the factory calibration.
MAN073 12/2016 V1.095 of 117
7 Software
7.8Permanent Parameter Storage / Factory Reset
The property UserSetSave (in category UserSetControl) stores the current camera settings in the
non-volatile flash memory. At power-up these values are loaded.
The property UserSetSave (in category UserSetControl) overwrites the current camera settings
with the settings that are stored in the flash memory.
The command CameraHeadFactoryReset (in category PhotonfocusMain) restores the settings of the
camera head
The property CameraHeadStoreDefaults (in category PhotonfocusMain) stores only
the settings of the camera head in the flash memory. It is recommended to use
UserSetSave instead, as all properties are stored.
The calibration values of the FPN calibration are not stored with UserSetSave (or
CameraHeadStoreDefaults). Use the command Correction_SaveToFlash for this (see
Correction_SaveToFlash).
7.9Persistent IP address
It is possible to set a persistent IP address:
1.Set GevPersistentIPAddress (in category TransportLayerControl) to the desired IP address.
2.Set GevPersistentSubnetMask (in category TransportLayerControl) to the sub net mask.
3.Set GevCurrentIPConfigurationPersistent (in category TransportLayerControl) to True.
4.Set GevCurrentIPConfigurationDHCP (in category TransportLayerControl) to False.
5.The selected persistent IP address will be applied after a reboot of the camera.
.
96 of 117MAN073 12/2016 V1.0
7.10 PLC Settings
A 0 ( L i n e 0 )
A 1 ( L i n e 1 )
S i g n a l
R o u t i n g
B l o c k
A 4
A 5
A 6
A 7
P L C
7
I S O _ I N 0
9
I S O _ I N 1
I S O _ I N C 0 _ P
5
4
I S O _ I N C 0 _ N
I S O _ I N C 1 _ P
1 1
1 0
I S O _ I N C 1 _ N
P o w e r C o n n e c t o r
I / O d e c o u p l i n g
F V A L
L V A L
D V A L
R E S E R V E D
P L C _ c t r l 0
P L C _ c t r l 1
P L C _ c t r l 2
P L C _ c t r l 3
Q 2
Q 3
Q 6
Q 7
p g 0 _ o u t
p g 1 _ o u t
p g 2 _ o u t
p g 3 _ o u t
d e l _ o u t
r s l _ o u t
g p _ c n t _ e q
g p _ c n t _ g t
t s _ t r i g 0
t s _ t r i g 1
t s _ t r i g 2
t s _ t r i g 3
L o o k u p
T a b l e
I 1
I 2
I 3
I 4
I 5
I 6
I 7
I 0
E n h a n c e d
F u n c t i o n
B l o c k
Q 0
Q 1
Q 2
Q 3
Q 4
Q 5
Q 6
Q 7
Q 8
Q 9
Q 1 0
Q 1 1
Q 1 5
Q 1 6
Q 1 7
R e m o t e
C o n t r o l
B l o c k
f r o m
h o s t P C
8
I S O _ O U T 1
I m a g e
C o n t r o l
B l o c k
Q 1 2
Q 1 3
Q 1 4
T r i g g e r S o f t w a r e
T r i g g e r S o u r c e
F r e e - r u n n i n g t r i g g e r
I n t e r n a l
c a m e r a
t r i g g e r
T r i g g e r M o d e
3
I S O _ O U T 0
S t r o b e
A 2 ( L i n e 2 )
A 3 ( L i n e 3 )
L i n e 1
P L C _ Q 4
S o f t w a r e
O f f
O n
1
C A M E R A _ G N D
2
C A M E R A _ P W R
6
I S O _ P W R
1 2
I S O _ G N D
I / O d e c o u p l i n g , i n v e r t i n g
T r i g g e r
D i v i d e r
A B
T r i g g e r
A
B
A B T r i g g e r D i v i d e r
A B T r i g g e r M o d e
A B T r i g g e r D i r e c t i o n
A B T r i g g e r D e b o u n c e
A B T r i g g e r A O n l y
A B T r i g g e r
T r i g g e r S o u r c e
7.10PLC Settings
7.10.1Introduction
The Programmable Logic Controller (PLC) is a powerful tool to generate triggers and software
interrupts. A functional diagram of the PLC tool is shown in Fig. 7.4. The PLC tool is described
in detail with many examples in the [PLC] manual which is included in the PFInstaller.
The AB Trigger feature is not available on all camera revisions, see Appendix B
for a list of available features.
Figure 7.4: PLC functional overview and trigger connections
The simpliest application of the PLC is to connect a PLC input to a PLC output. The connection
of the ISO_IN0 input to the PLC_Q4 camera trigger is given as an example. The resulting
configuration is shown in Section 7.10.2.
MAN073 12/2016 V1.097 of 117
7 Software
1.Identify the PLC notation of the desired input. A table of the PLC mapping is given in
Section 6.9. In our example, ISO_IN0 maps to A0 or Line0.
2.Select a Signal Routing Block (SRB) that has a connection to the desired PLC input and
connect it to the PLC input. In our example, SRB PLC_I0 will be used as it has a connection
to Line0. To connect the SRB to input, set PLC_I<x> to the input. In the example, set PLC_I0
to Line0.
3.Identify the PLC notation of the desired output. A table of the PLC mapping is given in
Section 6.9. In the example Q4 is the desired output.
4.Connect the LUT that corresponds to the desired output to the SRB from step 2. In the
example, PLC_Q4 is connected to PLC_I0. Note that every LUT has the capability to connect
up to 4 inputs. In the example only the first input (PLC_Q4_Variable0) is used. The other
inputs are ignored by setting the PLC_Q4_Variable to Zero and the PLC_Q4_Operator to
Or for inputs 1 to 3.
5.If a PLC output is used to connect to a camera trigger, then the corresponding Trigger
Source must be activated. In the example, TriggerSource is set to PLC_Q4 and TriggerMode
is set to On.
7.10.2PLC Settings for ISO_IN0 to PLC_Q4 Camera Trigger
This setting connects the ISO_IN0 to the internal camera trigger, see Table 7.1 (the visibility in
the PF_GEVPlayer must be set to Guru for this purpose).
7.10.3PLC Settings for A/B Trigger from differential inputs
This settings connects the ISO_INC differential inputs to the A/B camera inputs. ISO_INC0 is
mapped to the A signal and ISO_INC1 to the B signal, see Table 7.2 (the visibility in the
PF_GEVPlayer must be set to Guru for this purpose).
The AB Trigger feature is not available on all camera revisions, see Appendix B
for a list of available features.
FeatureValueCategory
TriggerModeOnAcquisitionControl
TriggerSourceABTriggerAcquisitionControl
PLC_I2Line2<PLC>/SignalRoutingBlock
PLC_I3Line3<PLC>/SignalRoutingBlock
PLC_Q6_Variable0PLC_I2<PLC>/LookupTable/Q6
PLC_Q6_Operator0Or<PLC>/LookupTable/Q6
PLC_Q6_Variable1Zero<PLC>/LookupTable/Q6
PLC_Q6_Operator1Or<PLC>/LookupTable/Q6
PLC_Q6_Variable2Zero<PLC>/LookupTable/Q6
PLC_Q6_Operator2Or<PLC>/LookupTable/Q6
PLC_Q6_Variable3Zero<PLC>/LookupTable/Q6
PLC_Q7_Variable0PLC_I3<PLC>/LookupTable/Q7
PLC_Q7_Operator0Or<PLC>/LookupTable/Q7
PLC_Q7_Variable1Zero<PLC>/LookupTable/Q7
PLC_Q7_Operator1Or<PLC>/LookupTable/Q7
PLC_Q7_Variable2Zero<PLC>/LookupTable/Q7
PLC_Q7_Operator2Or<PLC>/LookupTable/Q7
PLC_Q7_Variable3Zero<PLC>/LookupTable/Q7
Table 7.2:PLC Settings for A/B Trigger from differential inputs (<PLC> = in category
IPEngine/ProgrammableLogicController)
.
MAN073 12/2016 V1.099 of 117
7 Software
7.10.4PLC Settings for A/B Trigger from single-ended inputs
This configuration maps the single-ended inputs to the A/B camera inputs: ISO_IN0 is mapped
to the A signal and ISO_IN1 to the B signal see Table 7.3 (the visibility in the PF_GEVPlayer must
be set to Guru for this purpose).
The AB Trigger feature is not available on all camera revisions, see Appendix B
for a list of available features.
FeatureValueCategory
TriggerModeOnAcquisitionControl
TriggerSourceABTriggerAcquisitionControl
PLC_I0Line0<PLC>/SignalRoutingBlock
PLC_I1Line1<PLC>/SignalRoutingBlock
PLC_Q6_Variable0PLC_I0<PLC>/LookupTable/Q6
PLC_Q6_Operator0Or<PLC>/LookupTable/Q6
PLC_Q6_Variable1Zero<PLC>/LookupTable/Q6
PLC_Q6_Operator1Or<PLC>/LookupTable/Q6
PLC_Q6_Variable2Zero<PLC>/LookupTable/Q6
PLC_Q6_Operator2Or<PLC>/LookupTable/Q6
PLC_Q6_Variable3Zero<PLC>/LookupTable/Q6
PLC_Q7_Variable0PLC_I1<PLC>/LookupTable/Q7
PLC_Q7_Operator0Or<PLC>/LookupTable/Q7
PLC_Q7_Variable1Zero<PLC>/LookupTable/Q7
PLC_Q7_Operator1Or<PLC>/LookupTable/Q7
PLC_Q7_Variable2Zero<PLC>/LookupTable/Q7
PLC_Q7_Operator2Or<PLC>/LookupTable/Q7
PLC_Q7_Variable3Zero<PLC>/LookupTable/Q7
Table 7.3:PLC Settings for A/B Trigger from single-ended inputs (<PLC> = in category
IPEngine/ProgrammableLogicController)
100 of 117MAN073 12/2016 V1.0
Loading...
+ hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.