Full text: Proceedings; XXI International Congress for Photogrammetry and Remote Sensing (Part B5-2)

MAPPPING DIGITAL IMAGE TEXTURE ONTO 3D MODEL FROM LIDAR DATA 
Chunmei Hu a ’,Yanmin Wang a ,Wentao Yu b 
a School of Geomatics and Urban Information, Beijing University of Civil Engineering and Architecture, 1# 
Zhanlanguan Road, Beijing, 100044, China,-(huchunmei81)@ 163.com 
b The Second Engineering Department, Heilongjiang Bureau of Surveying and Mapping, Cehui Road, Harbin, 150086, 
China, - yuwentao81@163.com 
Wgs-PS WGV/3 
KEY WORDS: LIDAR Data, Point Cloud, 3D TIN Model, Digital Image, Texture Mapping, Error Correction 
ABSTRACT: 
In this paper, an experiment system is developed to settle the problem about how to mapping a digital image onto 3D model from 
LIDAR data. Firstly, it chooses the corresponding points between point cloud and digital image, then, uses these corresponding 
points to calculate the exterior and interior orientation elements and systematic error corrections of image. For the purpose of 
improving the precision of these parameters, paper uses the iterative least squares computational method. The digital image comes 
from non-metric camera, and the initial values of orientation elements can not been obtained, so, paper replaces the collinearity 
equation of photogrammetry to calculate these parameters with the DLT. Finally, by using the texture mapping technology, paper 
maps the corresponding image onto point cloud model and TIN model, makes out textured point cloud model and textured TIN 
model. At the same time, it projects this image to the basic level of range image to make out orthophoto map. 
1. INTRODUCTION 
In recent years, the Laser Scanning technology has made a great 
progress in many application fields, such as historic buildings 
protection, 3D reconstruction and other aspects, and the texture 
mapping of 3D model is an important content in these fields. At 
present, some software such as Texcapture, RapidForm have 
actualized texture mapping function. However, it has been 
proved that these software’s functions are limited for the 3D 
model from LIDAR. On one hand, RapidForm’s precision can 
not achieve request. On the other hand, although Texcapture’s 
precision has achieved request, the result of texture mapping is 
not satisfying. Sometimes the edge of texture mapping would 
be declining, and for the large model data, the result of texture 
mapping can not be displayed by the software. So, it is 
significant to develop software of texture mapping based on the 
Laser scanning data and digital image to realize the data fusion. 
Paper develops an experiment system to realize the data fusion, 
and the principle and method about this experiment system are 
introduced in follow sections. 
2. SELECTING CORRESPONDING POINT 
The system need picking up corresponding points in the screen 
from digital image and point cloud model to calculate the 
orientation parameters between image and point cloud. For 
calculating these parameters, at least 6 pairs of corresponding 
points should be selected. There is a valuable noted point that 
corresponding points should be distributed around the image 
and point cloud with image bound as the first. These can be 
considered by the principle of photogrammetry. 
2.1 Selecting Character Points in Digital Image 
There has used the function of screen coordinates selection. 
When the mouse points the character point, system get the 
screen coordinates. Then, we should convert the coordinates 
from screen to image. For improving the selected point 
precision, the image had better to be magnified to the pixel 
level. 
2.2 Selecting Corresponding Character Point in Point Cloud 
Because the point cloud model is 3D, the paper utilizes the 
projection radial method to pick the 3D coordinate from screen 
point. First, by using the two points of intersection which come 
from the intersection of screen point projection radial and near 
and far cross sections, we can get this line equation. Then, 
according to calculating the distance between the line and the 
model points, the points within a threshold can be selected. At 
last, we chose a nearest distance point as the 3D coordinate of 
screen point after calculating the distance between the selected 
points and the viewpoint. Figure 1 is a pair of corresponding 
point. 
Figure 1 Corresponding Point 
3. CALCULATING THE PARAMETERS OF 
ORIENTATION AND LENS DISTORTION 
For actualizing the registration between digital image and point 
cloud, we should firstly calculate the parameters of orientation 
and lens distortion. The digital image comes from the non 
metric camera, which can not supply the initial value of
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.