Organizational Semiotics Complements Knowledge Management

Author(s):  
Jeffrey A. Schiffel

Inserting the human element into an Information System leads to interpreting the Information System as an information field. Organizational semiotics provides a means to analyze this alternate interpretation. The semantic normal forms of organizational semiotics extract structures from natural language texts that may be stored electronically. In themselves, the SNFs are only canonic descriptions of the patterns of behavior observed in a culture. Conceptual graphs and dataflow graphs, their dynamic variety, provide means to reason over propositions in first order logics. Conceptual graphs, however, do not of themselves capture the ontological entities needed for such reasoning. The culture of an organization contains natural language entities that can be extracted for use in knowledge representation and reasoning. Together in a rigorous, two-step process, ontology charting from organizational semiotics and dataflow graphs from knowledge engineering provide a means to extract entities of interest from a subject domain such as the culture of organizations and then to represent these entities in formal logic reasoning. This chapter presents this process, and concludes with an example of how process improvement in an IT organization may be measured in this two-step process.

Author(s):  
Jeffrey A. Schiffel

The semantic normal forms of organizational semiotics extract structures from natural language texts that may be stored electronically. In themselves, the SNFs are only canonic descriptions of the patterns of behavior observed in a culture. Conceptual graphs and dataflow graphs, their dynamic variety, provide means to reason over propositions in first order logics. Conceptual graphs, however, do not of themselves capture the ontological entities needed for such reasoning. The culture of an organization contains natural language entities that can be extracted for use in knowledge representation and reasoning. Together in a rigorous, two-step process, ontology charting from organizational semiotics and dataflow graphs from knowledge engineering provide a means to extract entities of interest from a subject domain such as the culture of organizations and then to represent these entities in formal logic reasoning. This paper presents this process, and concludes with an example of how process improvement in an IT organization may be measured in this two-step process.


Author(s):  
TRU H. CAO

Conceptual graphs and fuzzy logic are two logical formalisms that emphasize the target of natural language, where conceptual graphs provide a structure of formulas close to that of natural language sentences while fuzzy logic provides a methodology for computing with words. This paper proposes fuzzy conceptual graphs as a knowledge representation language that combines the advantages of both the two formalisms for artificial intelligence approaching human expression and reasoning. Firstly, the conceptual graph language is extended with functional relation types for representing functional dependency, and conjunctive types for joining concepts and relations. Then fuzzy conceptual graphs are formulated as a generalization of conceptual graphs where fuzzy types and fuzzy attribute-values are used in place of crisp types and crisp attribute-values. Projection and join as basic operations for reasoning on fuzzy conceptual graphs are defined, taking into account the semantics of fuzzy set-based values.


2021 ◽  
Vol 26 (4) ◽  
Author(s):  
Alvaro Veizaga ◽  
Mauricio Alferez ◽  
Damiano Torre ◽  
Mehrdad Sabetzadeh ◽  
Lionel Briand

AbstractNatural language (NL) is pervasive in software requirements specifications (SRSs). However, despite its popularity and widespread use, NL is highly prone to quality issues such as vagueness, ambiguity, and incompleteness. Controlled natural languages (CNLs) have been proposed as a way to prevent quality problems in requirements documents, while maintaining the flexibility to write and communicate requirements in an intuitive and universally understood manner. In collaboration with an industrial partner from the financial domain, we systematically develop and evaluate a CNL, named Rimay, intended at helping analysts write functional requirements. We rely on Grounded Theory for building Rimay and follow well-known guidelines for conducting and reporting industrial case study research. Our main contributions are: (1) a qualitative methodology to systematically define a CNL for functional requirements; this methodology is intended to be general for use across information-system domains, (2) a CNL grammar to represent functional requirements; this grammar is derived from our experience in the financial domain, but should be applicable, possibly with adaptations, to other information-system domains, and (3) an empirical evaluation of our CNL (Rimay) through an industrial case study. Our contributions draw on 15 representative SRSs, collectively containing 3215 NL requirements statements from the financial domain. Our evaluation shows that Rimay is expressive enough to capture, on average, 88% (405 out of 460) of the NL requirements statements in four previously unseen SRSs from the financial domain.


Author(s):  
Clifford Nangle ◽  
Stuart McTaggart ◽  
Margaret MacLeod ◽  
Jackie Caldwell ◽  
Marion Bennie

ABSTRACT ObjectivesThe Prescribing Information System (PIS) datamart, hosted by NHS National Services Scotland receives around 90 million electronic prescription messages per year from GP practices across Scotland. Prescription messages contain information including drug name, quantity and strength stored as coded, machine readable, data while prescription dose instructions are unstructured free text and difficult to interpret and analyse in volume. The aim, using Natural Language Processing (NLP), was to extract drug dose amount, unit and frequency metadata from freely typed text in dose instructions to support calculating the intended number of days’ treatment. This then allows comparison with actual prescription frequency, treatment adherence and the impact upon prescribing safety and effectiveness. ApproachAn NLP algorithm was developed using the Ciao implementation of Prolog to extract dose amount, unit and frequency metadata from dose instructions held in the PIS datamart for drugs used in the treatment of gastrointestinal, cardiovascular and respiratory disease. Accuracy estimates were obtained by randomly sampling 0.1% of the distinct dose instructions from source records, comparing these with metadata extracted by the algorithm and an iterative approach was used to modify the algorithm to increase accuracy and coverage. ResultsThe NLP algorithm was applied to 39,943,465 prescription instructions issued in 2014, consisting of 575,340 distinct dose instructions. For drugs used in the gastrointestinal, cardiovascular and respiratory systems (i.e. chapters 1, 2 and 3 of the British National Formulary (BNF)) the NLP algorithm successfully extracted drug dose amount, unit and frequency metadata from 95.1%, 98.5% and 97.4% of prescriptions respectively. However, instructions containing terms such as ‘as directed’ or ‘as required’ reduce the usability of the metadata by making it difficult to calculate the total dose intended for a specific time period as 7.9%, 0.9% and 27.9% of dose instructions contained terms meaning ‘as required’ while 3.2%, 3.7% and 4.0% contained terms meaning ‘as directed’, for drugs used in BNF chapters 1, 2 and 3 respectively. ConclusionThe NLP algorithm developed can extract dose, unit and frequency metadata from text found in prescriptions issued to treat a wide range of conditions and this information may be used to support calculating treatment durations, medicines adherence and cumulative drug exposure. The presence of terms such as ‘as required’ and ‘as directed’ has a negative impact on the usability of the metadata and further work is required to determine the level of impact this has on calculating treatment durations and cumulative drug exposure.


2014 ◽  
pp. 439-472
Author(s):  
John F. Sowa

Existential graphs (EGs) are a simple, readable, and expressive graphic notation for logic. Conceptual graphs (CGs) combine a logical foundation based on EGs with features of the semantic networks used in artificial intelligence and computational linguistics. CG design principles address logical, linguistic, and cognitive requirements: a formal semantics defined by the ISO standard for Common Logic; the flexibility to support the expressiveness, context dependencies, and metalevel commentary of natural language; and cognitively realistic operations for reasoning by induction, deduction, abduction, and analogy. To accommodate the vagueness and ambiguities of natural language, informal heuristics can supplement the formal semantics. With sufficient background knowledge and a clarifying dialog, informal graphs can be refined to any degree of precision. Peirce claimed that the rules for reasoning with EGs generate “a moving picture of the action of the mind in thought.” Some philosophers and psychologists agree: Peirce's diagrams and rules are a good candidate for a natural logic that reflects the neural processes that support thought and language. They are psychologically realistic and computationally efficient.


1988 ◽  
Vol 9 (4) ◽  
pp. 19-33 ◽  
Author(s):  
Bipin C. Desai ◽  
Richard J. Pollock ◽  
Philip J. Vincent

2003 ◽  
Vol 6 (3) ◽  
pp. 167-180 ◽  
Author(s):  
HELMUT BERGER ◽  
MICHAEL DITTENBACH ◽  
DIETER MERKL

Sign in / Sign up

Export Citation Format

Share Document