scholarly journals A Rapid Method for Information Extraction from Borehole Log Images

2020 ◽  
Vol 10 (16) ◽  
pp. 5520
Author(s):  
Junqiang Zhang ◽  
Yi Zhang ◽  
Yiping Tian ◽  
Gang Liu ◽  
Lirui Xu ◽  
...  

Borehole logs are very important for geological analysis and application. Extracting structured information from borehole logs in the image format is the key to any analysis and application based on borehole data. The current method has defects in solving the beard phenomenon of the borehole log and the identification of special geological symbols. This paper proposes an automatic extraction method for borehole log information by combining the structural analysis based on the corner mark, as well as the structural understanding based on deep learning. The principles and key technologies of the method are described in detail. The performance of the method was tested by specific examples. This method is implemented on a geological information platform called QuantyView. The information extraction of 100 borehole logs with the same specification is used to verify the effectiveness of the proposed method. The results show that the method can not only effectively solve the inconsistency between the thickness and the description information in the borehole log but it can also address the low recognition efficiency of professional vocabulary, which can improve the extraction efficiency and accuracy of the borehole log information.

2013 ◽  
Vol 7 (2) ◽  
pp. 574-579 ◽  
Author(s):  
Dr Sunitha Abburu ◽  
G. Suresh Babu

Day by day the volume of information availability in the web is growing significantly. There are several data structures for information available in the web such as structured, semi-structured and unstructured. Majority of information in the web is presented in web pages. The information presented in web pages is semi-structured.  But the information required for a context are scattered in different web documents. It is difficult to analyze the large volumes of semi-structured information presented in the web pages and to make decisions based on the analysis. The current research work proposed a frame work for a system that extracts information from various sources and prepares reports based on the knowledge built from the analysis. This simplifies  data extraction, data consolidation, data analysis and decision making based on the information presented in the web pages.The proposed frame work integrates web crawling, information extraction and data mining technologies for better information analysis that helps in effective decision making.   It enables people and organizations to extract information from various sourses of web and to make an effective analysis on the extracted data for effective decision making.  The proposed frame work is applicable for any application domain. Manufacturing,sales,tourisum,e-learning are various application to menction few.The frame work is implemetnted and tested for the effectiveness of the proposed system and the results are promising.


Geophysics ◽  
1992 ◽  
Vol 57 (2) ◽  
pp. 334-342 ◽  
Author(s):  
Larry R. Lines ◽  
Kenneth R. Kelly ◽  
John Queen

Layered geological formations with large seismic velocity contrasts can effectively create channel waves in cross‐borehole seismic data. The existence of channel waves for such waveguides can be confirmed by ray tracing, wave equation modeling, and modal analysis. Channel wave arrivals are identified in cross‐borehole data recorded at Conoco’s Newkirk test facility. For these data, where velocity contrasts are about 2 to 1, tomography based on first arrival traveltimes, is limited due to problems with extreme ray bending and seismic shadow zones. However, it may be possible to extract geological information using channel wave information. The seismometer differencing method appears to be a promising approach for detecting waveguide boundaries by use of cross‐borehole data.


2020 ◽  
Vol 8 (6) ◽  
pp. 3281-3287

Text is an extremely rich resources of information. Each and every second, minutes, peoples are sending or receiving hundreds of millions of data. There are various tasks involved in NLP are machine learning, information extraction, information retrieval, automatic text summarization, question-answered system, parsing, sentiment analysis, natural language understanding and natural language generation. The information extraction is an important task which is used to find the structured information from unstructured or semi-structured text. The paper presents a methodology for extracting the relations of biomedical entities using spacy. The framework consists of following phases such as data creation, load and converting the data into spacy object, preprocessing, define the pattern and extract the relations. The dataset is downloaded from NCBI database which contains only the sentences. The created model evaluated with performance measures like precision, recall and f-measure. The model achieved 87% of accuracy in retrieving of entities relation.


Author(s):  
Erma Susanti ◽  
Khabib Mustofa

AbstrakEkstraksi  informasi  merupakan suatu bidang ilmu untuk pengolahan bahasa alami, dengan cara mengubah teks tidak terstruktur menjadi informasi dalam bentuk terstruktur. Berbagai jenis informasi di Internet ditransmisikan secara tidak terstruktur melalui website, menyebabkan munculnya kebutuhan akan suatu teknologi untuk menganalisa teks dan menemukan pengetahuan yang relevan dalam bentuk informasi terstruktur. Contoh informasi tidak terstruktur adalah informasi utama yang ada pada konten halaman web. Bermacam pendekatan untuk ekstraksi informasi telah dikembangkan oleh berbagai peneliti, baik menggunakan metode manual atau otomatis, namun masih perlu ditingkatkan kinerjanya terkait akurasi dan kecepatan ekstraksi. Pada penelitian ini diusulkan suatu penerapan pendekatan ekstraksi informasi dengan mengkombinasikan pendekatan bootstrapping dengan Ontology-based Information Extraction (OBIE). Pendekatan bootstrapping dengan menggunakan sedikit contoh data berlabel, digunakan untuk memimalkan keterlibatan manusia dalam proses ekstraksi informasi, sedangkan penggunakan panduan ontologi untuk mengekstraksi classes (kelas), properties dan instance digunakan untuk menyediakan konten semantik untuk web semantik. Pengkombinasian kedua pendekatan tersebut diharapkan dapat meningkatan kecepatan proses ekstraksi dan akurasi hasil ekstraksi. Studi kasus untuk penerapan sistem ekstraksi informasi menggunakan dataset “LonelyPlanet”. Kata kunci—Ekstraksi informasi, ontologi, bootstrapping, Ontology-Based Information Extraction, OBIE, kinerja Abstract Information extraction is a field study of natural language processing by converting unstructured text into structured information. Several types of information on the Internet is transmitted through unstructured information via websites, led to emergence of the need a technology to analyze text and found relevant knowledge into structured information. For example of unstructured information is existing main information on the content of web pages. Various approaches  for information extraction have been developed by many researchers, either using manual or automatic method, but still need to be improved performance related accuracy and speed of extraction. This research proposed an approach of information extraction that combines bootstrapping approach with Ontology-Based Information Extraction (OBIE). Bootstrapping approach using small seed of labelled data, is used to minimize human intervention on information extraction process, while the use of guide ontology for extracting classes, properties and instances, using for provide semantic content for semantic web. Combining both approaches expected to increase speed of extraction process and accuracy of extraction results. Case study to apply information extraction system using “LonelyPlanet” datasets. Keywords— Information extraction, ontology, bootstrapping, Ontology-Based Information Extraction, OBIE, performance


2011 ◽  
pp. 2048-2081
Author(s):  
Gijs Geleijnse ◽  
Jan Korst

In this chapter we discuss approaches to find, extract, and structure information from natural language texts on the Web. Such structured information can be expressed and shared using the standard Semantic Web languages and hence be machine interpreted. In this chapter we focus on two tasks in Web information extraction. The first part focuses on mining facts from the Web, while in the second part, we present an approach to collect community-based meta-data. A search engine is used to retrieve potentially relevant texts. From these texts, instances and relations are extracted. The proposed approaches are illustrated using various case-studies, showing that we can reliably extract information from the Web using simple techniques.


2020 ◽  
Author(s):  
Guillem Subiela ◽  
Miquel Vilà ◽  
Roser Pi ◽  
Elena Sánchez

<p>Studying urban geology is a key way to identify municipal issues involved with urban development and sustainability, land resources and hazard awareness in highly populated areas. In the last decade, one of the lines of work of the Catalan Geological Survey (Institut Cartogràfic i Geològic de Catalunya) has been the development of (i) 1:5.000 scale Urban Geological Map of Catalonia project. Besides, two pilot projects have recently been started: (ii) the system of layers of geological information and (iii) the fundamental geological guides of municipalities. This communication focuses on the presentation of these projects and their utility, with the aim of finding effective ways of transferring geological knowledge and information of a territory, from a geological survey perspective.</p><p>The 1:5.000 urban geological maps of Catalonia (i) have been a great ambitious project focused on providing detailed, consistent and accurate geological, geotechnical and anthropogenic activity information of the main urban areas of Catalonia. Nevertheless, it must be taken into account that the compilation and elaboration of a large volume of geological information and also the high level of detail require a lot of time for data completeness.</p><p>In order to optimize a greater distribution of information, a system of layers of geological information (ii) covering urban areas is being developed. This pilot project consists of providing specific layers of Bedrock materials, Quaternary deposits, anthropogenic grounds, structural measures, geochemical compositions, borehole data and so on. However, as information layers are treated individually, it may not be clear the coherence between data from different layers of information and its use is currently limited to Earth-science professionals working with geological data.</p><p>Hence, as a strategy to reach a wider range of users and also provide a homogeneous and varied geological information, the development of fundamental geological guides for municipalities is also being carried out (iii). These documents include the general geological characterization of the municipality, the description of the main geological factors (related to geotechnical properties, hydrogeology, environmental concerns and geological hazards and resources) and the list of the sources of geological information to be considered. Moreover, each guide contains a 1:50.000 geological map that has cartographic continuity with the neighbouring municipalities. The municipal guides allow a synthesis of the geological environment of the different Catalan municipalities and give fundamental recommendations for the characterization of the geological environment of the municipality.</p><p>In conclusion, the three projects facilitate the characterization of geological environment of urban areas, the evaluation of geological factors in ground studies and also, in general, the management of the environment. These products differ depending on the degree of detail, the coherence of the geological information, the necessary knowledge for their execution or their purpose of use. This set of projects defines a geological urban framework, which is adjusted depending on the government’s requirements, the society’s needs and the geological survey’s available resources.</p>


2020 ◽  
Vol 10 (3) ◽  
pp. 35
Author(s):  
Ahmed Adeeb Jalal

Technology world has greatly evolved over the past decades, which led to inflated data volume. This progress of technology in the digital form generated scattered texts across millions of web pages. Unstructured texts contain a vast amount of textual data. Discover of useful and interesting relations from unstructured texts requires more processing by computers. Therefore, text mining and information extraction have become an exciting research field to get structured and valuable information. This paper focuses on text pre-processing of automotive advertisements domains to configure a structured database. The structured database was created by extract the information over unstructured automotive advertisements, which is an area of natural language processing. Information extraction deals with finding factual information in text using learning regular expressions. We manually craft rule-based specific approaches to extract structured information from unstructured web pages. Structured information will be provided by user-friendly search engine designed for topic-specific knowledge. Consequently, this information that extracted from these advertisements uses to perform a structured search over certain interesting attributes. Thus, the tuples are assigned a probability and indexed to support the efficiency of extraction and exploration via user queries.


Sign in / Sign up

Export Citation Format

Share Document