Refine
Year of publication
- 2016 (21) (remove)
Document Type
- Article (21) (remove)
Language
- English (21) (remove)
Has Fulltext
- no (21)
Keywords
- Abstract interpretation (1)
- Block codes (1)
- Business plan (1)
- Bytecode (1)
- Code Generation (1)
- Concatenated codes (1)
- Content analysis (1)
- Decoding (1)
- Diversity (1)
- Domain-Specific Language (DSL) (3)
- Domain-Specific-Language (DSL) (1)
- EMF (1)
- ERP systems (1)
- Eclipse Modeling Framework (EMF) (1)
- Enterprise systems (1)
- Entry-wise perturbation (1)
- Extended Perron complement (1)
- Future tools and methods (1)
- Graphical Online Editor (1)
- Hankel matrix (1)
- Hurwitz matrix (1)
- Immutability (1)
- Joint building venture (1)
- Lake constance (1)
- Matrix interval (1)
- Measurement methods (1)
- Metamodel (1)
- Metamodel Definition (1)
- Metamodel Model-Driven Architecture (MDA) (1)
- Military land redevelopment (1)
- Model-Driven Architecture (MDA) (2)
- Model-Driven Software Development (MDSD) (3)
- Model-Driven-Development (MDD) (1)
- Movement patterns (1)
- Perron complement (1)
- Prediction (1)
- R-function (1)
- Relation (1)
- Scala (1)
- Schur complement (1)
- Shadow IT (1)
- Shadow systems (1)
- Totally nonnegative matrix (2)
- Totally positive matrix (1)
- Transaction relations (1)
- Triangulation (1)
- Urban development pattern (1)
- Urbanity (1)
- Venture emergence (1)
- Vertex matrix (1)
- Xtext (2)
- algebraic codes (1)
- modulation coding (1)
Sleep is an important aspect in life of every human being. The average sleep duration for an adult is approximately 7 h per day. Sleep is necessary to regenerate physical and psychological state of a human. A bad sleep quality has a major impact on the health status and can lead to different diseases. In this paper an approach will be presented, which uses a long-term monitoring of vital data gathered by a body sensor during the day and the night supported by mobile application connected to an analyzing system, to estimate sleep quality of its user as well as give recommendations to improve it in real-time. Actimetry and historical data will be used to improve the individual recommendations, based on common techniques used in the area of machine learning and big data analysis.
The corrosion resistance of stainless steels is massively influenced by the condition of their surface. The surface quality includes the topography of the surface, the structure and composition of the passive layer, and the surface near structure of the base material. These factors are influenced by final physical/chemical surface treatments. The presented work shows significantly lower corrosion resistance for mechanical machined specimens than for etched specimens. It also turns out that the rougher the surface, the lower the corrosion resistance gets. However, there is no general finding which shows if blasted or grinded surfaces are more appropriate, but a dependency on process parameters and the characteristics on corrosive exposure in terms of corrosion behavior. The results show that not only the surface roughness Ra has an influence on corrosion behavior but also the shape of peaks and valleys which are evolved by surface treatments. Imperfections in the base material, like sulfidic inclusions lead to a weaker passive layer, respectively, to a decrease of the corrosion resistance. By using special passivating techniques the corrosion resistance of stainless steels can be increased to a higher level in comparison to common passivation.
Even though immutability is a desirable property, especially in a multi-threaded environment, implementing immutable Java classes is surprisingly hard because of a lack of language support. We present a static analysis tool using abstract bytecode interpretation that checks Java classes for compliance with a set of rules that together constitute state-based immutability. Being realized as a Find Bugs plug in, the tool can easily be integrated into most IDEs and hence the software development process. Our evaluation on a large, real world codebase shows that the average run-time effort for a single class is in the range of a few milliseconds, with only a very few statistical spikes.
To learn from the past, we analyse 1,088 "computer as a target" judgements for evidential reasoning by extracting four case elements: decision, intent, fact, and evidence. Analysing the decision element is essential for studying the scale of sentence severity for cross-jurisdictional comparisons. Examining the intent element can facilitate future risk assessment. Analysing the fact element can enhance an organization's capability of analysing criminal activities for future offender profiling. Examining the evidence used against a defendant from previous judgements can facilitate the preparation of evidence for upcoming legal disclosure. Follow the concepts of argumentation diagrams, we develop an automatic judgement summarizing system to enhance the accessibility of judgements and avoid repeating past mistakes. Inspired by the feasibility of extracting legal knowledge for argument construction and employing grounds of inadmissibility for probability assessment, we conduct evidential reasoning of kernel traces for forensic readiness. We integrate the narrative methods from attack graphs/languages for preventing confirmation bias, the argumentative methods from argumentation diagrams for constructing legal arguments, and the probabilistic methods from Bayesian networks for comparing hypotheses.
ERP systems integrate a major part of all business processes and organizations include them in their IT service management. Besides these formal systems, there are additional systems that are rather stand-alone and not included in the IT management tasks. These so-called ‘shadow systems’ also support business processes but hinder a high enterprise integration. Shadow systems appear during their explicit detection or during software maintenance projects such as enhancements or release changes of enterprise systems. Organizations then have to decide if and to what extent they integrate the identified shadow systems into their ERP systems. For this decision, organizations have to compare the capabilities of each identified shadow system with their ERP systems. Based on multiple-case studies, we provide a dependency approach to enable their comparison. We derive categories for different stages of the dependency and base insights into integration possibilities on these stages. Our results show that 64% of the shadow systems in our case studies are related to ERP systems. This means that they share parts or all of their data and/or functionality with the ERP system. Our research contributes to the field of integration as well as to the discussion about shadow systems.
This paper proposes a soft input decoding algorithm and a decoder architecture for generalized concatenated (GC) codes. The GC codes are constructed from inner nested binary Bose-Chaudhuri-Hocquenghem (BCH) codes and outer Reed-Solomon codes. In order to enable soft input decoding for the inner BCH block codes, a sequential stack decoding algorithm is used. Ordinary stack decoding of binary block codes requires the complete trellis of the code. In this paper, a representation of the block codes based on the trellises of supercodes is proposed in order to reduce the memory requirements for the representation of the BCH codes. This enables an efficient hardware implementation. The results for the decoding performance of the overall GC code are presented. Furthermore, a hardware architecture of the GC decoder is proposed. The proposed decoder is well suited for applications that require very low residual error rates.
Domain-Specific modelling is increasingly adopted in the software development industry. While textual domain specific languages (DSLs) already have a wide impact, graphical DSLs still need to live up to their full potential. In this paper we describe an approach that reduces the time to create a graphical DSL to hours instead of months. The paper describes a generative approach to the creation of graphical editors for the Eclipse platform. A set of carefully designed textual DSLs together with an EMF meta-model are the input for the generator. The output is an Eclipse plugin for a graphical editor for the intended graphical language. The entire project is made available as open source under the name Spray and is being developed by an active community. This paper focuses on the description of the workflow and provides an introduction into the possibilities through this approach of a graphical modelling environment.
Domain-specific modeling is more and more understood as a comparable solution compared to classical software development. Textual domain-specific languages (DSLs) already have a massive impact in contrast tographical DSLs, they still have to show their full potential. The established textual DSLs are normally generated from a domain specific grammar or maybe other specific textual descriptions. And advantage of textual DSLs is that they can be development cost-efficient.
In this paper, we describe asimilar approach for the creation of graphical DSLs from textual descriptions. We present a set of specially developed textual DSLs to fully describe graphical DSLs based on node and edge diagrams. These are, together with an EMF meta-model, the input for a generator that produces an eclipse-based graphical Editor. The entire project is available as open source under the name MoDiGen.
Domain-specific modeling is increasingly adopted by the software development industry. While textual domain-specific languages (DSLs) already have a wide impact, graphical DSLs still need to live up to their full potential. Textual DSLs are usually generated from a grammar or other short textual notations; their development is often cost-efficient. In this paper, we describe an approach to similarly create graphical DSLs from textual notations. The paper describes an approach to generate a graphical node and edge online editor, using a set of carefully designed textual DSLs to fully describe graphical DSLs. Combined with an adequate metamodel, these textual definitions represent the input for a generator that produces a graphical Editor for the web with features such as collaboration, online storage and being always available. The entire project is made available as open source under the name Zeta. This paper focuses on the overall approach and the description of the textual DSLs that can be used to develop graphical modeling languages and editors.