Retrodigitalisierung Logo Full screen
  • First image
  • Previous image
  • Next image
  • Last image
  • Show double pages
Use the mouse to select the image area you want to share.
Please select which information should be copied to the clipboard by clicking on the link:
  • Link to the viewer page with highlighted frame
  • Link to IIIF image fragment

CMRT09

Access restriction

There is no access restriction for this record.

Copyright

CC BY: Attribution 4.0 International. You can find more information here.

Bibliographic data

fullscreen: CMRT09

Monograph

Persistent identifier:
856955019
Author:
Stilla, Uwe
Title:
CMRT09
Sub title:
object extraction for 3D city models, road databases, and traffic monitoring ; concepts, algorithms and evaluation ; Paris, France, September 3 - 4, 2009 ; [joint conference of ISPRS working groups III/4 and III/5]
Scope:
X, 234 Seiten
Year of publication:
2009
Place of publication:
Lemmer
Publisher of the original:
GITC
Identifier (digital):
856955019
Illustration:
Illustrationen, Diagramme, Karten
Language:
English
Usage licence:
Attribution 4.0 International (CC BY 4.0)
Publisher of the digital copy:
Technische Informationsbibliothek Hannover
Place of publication of the digital copy:
Hannover
Year of publication of the original:
2016
Document type:
Monograph
Collection:
Earth sciences

Chapter

Title:
TEXT EXTRACTION FROM STREET LEVEL IMAGES J. Fabrizio, M. Cord, B. Marcotegui
Document type:
Monograph
Structure type:
Chapter

Contents

Table of contents

  • CMRT09
  • Cover
  • ColorChart
  • Title page
  • Workshop Committees
  • Program Committee:
  • Preface
  • Contents
  • EFFICIENT ROAD MAPPING VIA INTERACTIVE IMAGE SEGMENTATION O. Barinova, R. Shapovalov, S. Sudakov, A. Velizhev, A. Konushin
  • SURFACE MODELLING FOR ROAD NETWORKS USING MULTI-SOURCE GEODATA Chao-Yuan Lo, Liang-Chien Chen, Chieh-Tsung Chen, and Jia-Xun Chen
  • AUTOMATIC EXTRACTION OF URBAN OBJECTS FROM MULTI-SOURCE AERIAL DATA Adriano Mancini, Emanuele Frontoni and Primo Zingaretti
  • ROAD ROUNDABOUT EXTRACTION FROM VERY HIGH RESOLUTION AERIAL IMAGERY M. Ravenbakhsh, C. S. Fraser
  • ASSESSING THE IMPACT OF DIGITAL SURFACE MODELS ON ROAD EXTRACTION IN SUBURBAN AREAS BY REGION-BASED ROAD SUBGRAPH EXTRACTION Anne Grote, Franz Rottensteiner
  • VEHICLE ACTIVITY INDICATION FROM AIRBORNE LIDAR DATA OF URBAN AREAS BY BINARY SHAPE CLASSIFICATION OF POINT SETS W. Yaoa, S. Hinz, U. Stilla
  • TRAJECTORY-BASED SCENE DESCRIPTION AND CLASSIFICATION BY ANALYTICAL FUNCTIONS D. Pfeiffer, R. Reulke
  • 3D BUILDING RECONSTRUCTION FROM LIDAR BASED ON A CELL DECOMPOSITION APPROACH Martin Kada, Laurence McKinle
  • A SEMI-AUTOMATIC APPROACH TO OBJECT EXTRACTION FROM A COMBINATION OF IMAGE AND LASER DATA S. A. Mumtaz, K. Mooney
  • COMPLEX SCENE ANALYSIS IN URBAN AREAS BASED ON AN ENSEMBLE CLUSTERING METHOD APPLIED ON LIDAR DATA P. Ramzi, F. Samadzadegan
  • EXTRACTING BUILDING FOOTPRINTS FROM 3D POINT CLOUDS USING TERRESTRIAL LASER SCANNING AT STREET LEVEL Karim Hammoudi, Fadi Dornaika and Nicolas Paparoditis
  • DETECTION OF BUILDINGS AT AIRPORT SITES USING IMAGES & LIDAR DATA AND A COMBINATION OF VARIOUS METHODS Demir, N., Poli, D., Baltsavias, E.
  • DENSE MATCHING IN HIGH RESOLUTION OBLIQUE AIRBORNE IMAGES M. Gerke
  • COMPARISON OF METHODS FOR AUTOMATED BUILDING EXTRACTION FROM HIGH RESOLUTION IMAGE DATA G. Vozikis
  • SEMI-AUTOMATIC CITY MODEL EXTRACTION FROM TRI-STEREOSCOPIC VHR SATELLITE IMAGERY F. Tack, R. Goossens, G. Buyuksalih
  • AUTOMATED SELECTION OF TERRESTRIAL IMAGES FROM SEQUENCES FOR THE TEXTURE MAPPING OF 3D CITY MODELS Sébastien Bénitez and Caroline Baillard
  • CLASSIFICATION SYSTEM OF GIS-OBJECTS USING MULTI-SENSORIAL IMAGERY FOR NEAR-REALTIME DISASTER MANAGEMENT Daniel Frey and Matthias Butenuth
  • AN APPROACH FOR NAVIGATION IN 3D MODELS ON MOBILE DEVICES Wen Jiang, Wu Yuguo, Wang Fan
  • GRAPH-BASED URBAN OBJECT MODEL PROCESSING Kerstin Falkowski and Jürgen Ebert
  • A PROOF OF CONCEPT OF ITERATIVE DSM IMPROVEMENT THROUGH SAR SCENE SIMULATION D. Derauw
  • COMPETING 3D PRIORS FOR OBJECT EXTRACTION IN REMOTE SENSING DATA Konstantinos Karantzalos and Nikos Paragios
  • OBJECT EXTRACTION FROM LIDAR DATA USING AN ARTIFICIAL SWARM BEE COLONY CLUSTERING ALGORITHM S. Saeedi, F. Samadzadegan, N. El-Sheimy
  • BUILDING FOOTPRINT DATABASE IMPROVEMENT FOR 3D RECONSTRUCTION: A DIRECTION AWARE SPLIT AND MERGE APPROACH Bruno Vallet and Marc Pierrot-Deseilligny and Didier Boldo
  • A TEST OF AUTOMATIC BUILDING CHANGE DETECTION APPROACHES Nicolas Champion, Franz Rottensteiner, Leena Matikainen, Xinlian Liang, Juha Hyyppä and Brian P. Olsen
  • CURVELET APPROACH FOR SAR IMAGE DENOISING, STRUCTURE ENHANCEMENT, AND CHANGE DETECTION Andreas Schmitt, Birgit Wessel, Achim Roth
  • RAY TRACING AND SAR-TOMOGRAPHY FOR 3D ANALYSIS OF MICROWAVE SCATTERING AT MAN-MADE OBJECTS S. Auer, X. Zhu, S. Hinz, R. Bamler
  • THEORETICAL ANALYSIS OF BUILDING HEIGHT ESTIMATION USING SPACEBORNE SAR-INTERFEROMETRY FOR RAPID MAPPING APPLICATIONS Stefan Hinz, Sarah Abelen
  • FUSION OF OPTICAL AND INSAR FEATURES FOR BUILDING RECOGNITION IN URBAN AREAS J. D. Wegner, A. Thiele, U. Soergel
  • FAST VEHICLE DETECTION AND TRACKING IN AERIAL IMAGE BURSTS Karsten Kozempel and Ralf Reulke
  • REFINING CORRECTNESS OF VEHICLE DETECTION AND TRACKING IN AERIAL IMAGE SEQUENCES BY MEANS OF VELOCITY AND TRAJECTORY EVALUATION D. Lenhart, S. Hinz
  • UTILIZATION OF 3D CITY MODELS AND AIRBORNE LASER SCANNING FOR TERRAIN-BASED NAVIGATION OF HELICOPTERS AND UAVs M. Hebel, M. Arens, U. Stilla
  • STUDY OF SIFT DESCRIPTORS FOR IMAGE MATCHING BASED LOCALIZATION IN URBAN STREET VIEW CONTEXT David Picard, Matthieu Cord and Eduardo Valle
  • TEXT EXTRACTION FROM STREET LEVEL IMAGES J. Fabrizio, M. Cord, B. Marcotegui
  • CIRCULAR ROAD SIGN EXTRACTION FROM STREET LEVEL IMAGES USING COLOUR, SHAPE AND TEXTURE DATABASE MAPS A. Arlicot, B. Soheilian and N. Paparoditis
  • IMPROVING IMAGE SEGMENTATION USING MULTIPLE VIEW ANALYSIS Martin Drauschke, Ribana Roscher, Thomas Läbe, Wolfgang Förstner
  • REFINING BUILDING FACADE MODELS WITH IMAGES Shi Pu and George Vosselman
  • AN UNSUPERVISED HIERARCHICAL SEGMENTATION OF A FAÇADE BUILDING IMAGE IN ELEMENTARY 2D - MODELS Jean-Pascal Burochin, Olivier Tournaire and Nicolas Paparoditis
  • GRAMMAR SUPPORTED FACADE RECONSTRUCTION FROM MOBILE LIDAR MAPPING Susanne Becker, Norbert Haala
  • Author Index
  • Cover

Full text

In: Stilla U, Rottensteiner F, Paparoditis N (Eds) CMRT09. IAPRS, Vol. XXXVIII, Part 3A/V4 — Paris, France, 3-4 September, 2009 
TEXT EXTRACTION FROM STREET LEVEL IMAGES 
J. Fabrizio 12 , M. Cord 1 , B. Marcotegui 2 
1 UPMC Univ Paris 06 
Laboratoire d’informatique de Paris 6, 75016 Paris, France 
2 MINES Paristech, CMM- Centre de morphologie mathématique. Mathématiques et Systèmes, 
35 rue Saint Honoré - 77305 Fontainebleau cedex, France 
KEY WORDS: Urban, Text, Extraction, Localization, Detection, Learning, Classification 
ABSTRACT 
We offer in this article, a method for text extraction in images issued from city scenes. This method is used in the 
French iTowns project (iTowns ANR project, 2008) to automatically enhance cartographic database by extracting text 
from geolocalized pictures of town streets. This task is difficult as 1. text in this environment varies in shape, size, 
color, orientation... 2. pictures may be blurred, as they are taken from a moving vehicle, and text may have perspective 
deformations, 3. all pictures are taken outside with various objects that can lead to false positives and in unconstrained 
conditions (especially light varies from one picture to the other). Then, we can not make the assumption on searched 
text. The only supposition is that text is not handwritten. Our process is based on two main steps: a new segmentation 
method based on morphological operator and a classification step based on a combination of multiple SVM classifiers. 
The description of our process is given in this article. The efficiency of each step is measured and the global scheme is 
illustrated on an example. 
1 INTRODUCTION 
Automatic text localization in images is a major task in 
computer vision. Applications of this task are various (au 
tomatic image indexing, visual impaired people assistance 
or optical character reading...). Our work deals with text 
localization and extraction from images in an urban en 
vironment and is a part of iTowns project (iTowns ANR 
project, 2008). This project has two main goals : 1. al 
lowing a user to navigate freely within the image flow of 
a city, 2. Extracting features automatically from this im 
age flow to automatically enhance cartographic databases 
and to allow the user to make high level queries on them 
(go to a given address, generate relevant hybrid text-image 
navigation maps (itinerary), find the location of an orphan 
image, select the images that contain an object, etc.). To 
achieve this work, geolocalized set of pictures are taken 
every meter. All images are processed off line to extract as 
many semantic data as possible and cartographic databases 
are enhanced with these data. At the same time, each mo 
saic of pictures is assembled into a complete immersive 
panorama (Figure 1). 
Many studies focus on text detection and localization in 
images. However, most of them are specific to a con 
strained context such as automatic localization of postal 
addresses on envelopes (Palumbo et al., 1992), license plate 
localization (Arth et al., 2007), text extraction in video 
sequences (Wolf et al., 2002), automatic forms reading 
(Kavallieratou et al., 2001) and more generally "documents” 
(Wahl et al., 1982). In such context, strong hypothesis 
may be asserted (blocks of text, alignments, temporal re 
dundancy for video sequences...). In our context (natural 
scenes in an urban environment), text comes from vari 
ous sources (road sign, storefront, advertisements...). Its 
extraction is difficult: no hypothesis can be made on text 
(style, position, orientation, lighting, perspective deforma 
tions...) and the amount of data is huge. Today, we work 
on 1 TB for a part of a single district in Paris. Next year, 
more districts will be processed (more than 4 TB). Differ- 
Segmentation ^R Fast filters Classification ^R Grouping 
Figure 2: General principle of our system. 
ent approaches already exist for text localization in natu 
ral scenes. States of the art are found in (Mancas-Thillou, 
2006, Retomaz and Marcotegui, 2007, Jung et al., 2004, 
Jian Liang et al., 2005). Even if preliminary works ex 
ist in natural scene (Retomaz and Marcotegui, 2007, Chen 
and Yuille, 2004), no standard solution really emerges and 
they do not focus on urban context. 
The paper presents our method and is organized as follows: 
the text localization process is presented and every step is 
detailed followed by the evaluation of main steps. In the 
last part, results are presented. Then comes the conclusion. 
2 SEGMENTATION BASED STRATEGY 
The goal of our system is to localize text. Once the lo 
calization is performed, the text recognition is carried out 
by an external O.C.R. (but the system may improve the 
quality of the region by correcting perspective deforma 
tions for example). Our system is a region based approach 
and starts by isolating letters, then groups them to restore 
words and text zones. Region based approach seems to be 
more efficient, such approach was ranked first (Retomaz 
and Marcotegui, 2007) during ImagEval campaign (Im- 
agEval, 2006). Our process is composed of a cascade of 
filters (Figure 2). It segments the image. Each region is 
analysed to determine whether the region corresponds to 
text or not. First stages during selection eliminate a part 
of non text regions but try to keep as many text region as 
possible (at the price of a lot of false positives). At the 
end, detected regions that are close to other text regions are 
grouped all together. Isolated text regions are canceled. 
199
	        

Cite and reuse

Cite and reuse

Here you will find download options and citation links to the record and current image.

Monograph

METS MARC XML Dublin Core RIS Mirador ALTO TEI Full text PDF DFG-Viewer OPAC
TOC

Chapter

PDF RIS

Image

PDF ALTO TEI Full text
Download

Image fragment

Link to the viewer page with highlighted frame Link to IIIF image fragment

Citation links

Citation links

Monograph

To quote this record the following variants are available:
Here you can copy a Goobi viewer own URL:

Chapter

To quote this structural element, the following variants are available:
Here you can copy a Goobi viewer own URL:

Image

To quote this image the following variants are available:
Here you can copy a Goobi viewer own URL:

Citation recommendation

Stilla, Uwe. CMRT09. GITC, 2009.
Please check the citation before using it.

Image manipulation tools

Tools not available

Share image region

Use the mouse to select the image area you want to share.
Please select which information should be copied to the clipboard by clicking on the link:
  • Link to the viewer page with highlighted frame
  • Link to IIIF image fragment

Contact

Have you found an error? Do you have any suggestions for making our service even better or any other questions about this page? Please write to us and we'll make sure we get back to you.

What is the first letter of the word "tree"?:

I hereby confirm the use of my personal data within the context of the enquiry made.