Open Science Research Excellence

Open Science Index

Commenced in January 2007 Frequency: Monthly Edition: International Abstract Count: 66741

The Normal-Generalized Hyperbolic Secant Distribution: Properties and Applications
In this paper, a new four-parameter univariate continuous distribution called the Normal-Generalized Hyperbolic Secant Distribution (NGHS) is defined and studied. Some general and structural distributional properties are investigated and discussed, including: central and non-central n-th moments and incomplete moments, quantile and generating functions, hazard function, Rényi and Shannon entropies, shapes: skewed right, skewed left, and symmetric, modality regions: unimodal and bimodal, maximum likelihood (MLE) estimators for the parameters. Finally, two real data sets are used to demonstrate empirically its flexibility and prove the strength of the new distribution.
Closed-Form Sharma-Mittal Entropy Rate for Gaussian Processes
The entropy rate of a stochastic process is a fundamental concept in information theory. It provides a limit to the amount of information that can be transmitted reliably over a communication channel, as stated by Shannon's coding theorems. Recently, researchers have focused on developing new measures of information that generalize Shannon's classical theory. The aim is to design more efficient information encoding and transmission schemes. This paper continues the study of generalized entropy rates, by deriving a closed-form solution to the Sharma-Mittal entropy rate for Gaussian processes. Using the squeeze theorem, we solve the limit in the definition of the entropy rate, for different values of alpha and beta, which are the parameters of the Sharma-Mittal entropy. In the end, we compare it with Shannon and Rényi's entropy rates for Gaussian processes.
Converse to the Sherman Inequality with Applications in Information Theory
We proved a converse to Sherman's inequality. Using the concept of f-divergence we obtained some inequalities for the well-known entropies, such as Shannon entropies that have many applications in many applied sciences, for example, in information theory, biology and economics Zipf-Mandelbrot law gave improvement in account for the low-rankwords in corpus. Applications of Zipf-Mandelbrot law can be found in linguistics, information sciences and also mostly applicable in ecological eld studies. We also introduced an entropy by applying the Zipf-Mandelbrot law and derived some related inequalities.
On the Fractional Integration of Generalized Mittag-Leffler Type Functions
In this paper, the generalized fractional integral operators of two generalized Mittag-Leffler type functions are investigated. The special cases of interest involve the generalized M-series and K-function, both introduced by Sharma. The two pairs of theorems established herein generalize recent results about left- and right-sided generalized fractional integration operators applied here to the M-series and the K-function. The note also results in important applications in physics and mathematical engineering.
The Impact of Socialization Preferences on Perceptions of Generalized Social Trust in China
Generalized social trust among Chinese has been declining in the past few decades, making the search for its causes necessary. Drawing on the symbolic interaction theory and the 2012 Chinese General Social Survey data, this research investigated the impact of people’s socialization preferences and frequencies on their perceptions of generalized social trust in China. This research also took a preliminary step towards understanding the spatial differences of the generalized social trust using the ArcGIS software. The results show that respondents who interacted with their neighbors more frequently were more likely to have higher levels of perceptions of generalized social trust. Several demographics were also significantly related to perception of generalized social trust. Elderly and better educated Chinese and people with higher self-perceived social status were associated with greater levels of generalized social trust perception, while urban dwellers and religious respondents expressed lower levels of such perception. Implications for future research and policy are discussed.
Nano Generalized Topology
Rough set theory is a recent approach for reasoning about data. It has achieved a large amount of applications in various real-life fields. The main idea of rough sets corresponds to the lower and upper set approximations. These two approximations are exactly the interior and the closure of the set with respect to a certain topology on a collection U of imprecise data acquired from any real-life field. The base of the topology is formed by equivalence classes of an equivalence relation E defined on U using the available information about data. The theory of generalized topology was studied by Cs´asz´ar. It is well known that generalized topology in the sense of Cs´asz´ar is a generalization of the topology on a set. On the other hand, many important collections of sets related with the topology on a set form a generalized topology. The notion of Nano topology was introduced by Lellis Thivagar, which was defined in terms of approximations and boundary region of a subset of an universe using an equivalence relation on it. The purpose of this paper is to introduce a new generalized topology in terms of rough set called nano generalized topology
Generalized Central Paths for Convex Programming
The central path has played the key role in the interior point method. However, the convergence of the central path may not be true even in some convex programming problems with linear constraints. In this paper, the generalized central paths are introduced for convex programming. One advantage of the generalized central paths is that the paths will always converge to some optimal solutions of the convex programming problem for any initial interior point. Some additional theoretical properties for the generalized central paths will be also reported.
Upper Bound of the Generalized P-Value for the Difference between Two Future Population Means
This paper presents the generalized p-values for testing the difference between two future population means when the variances are unknown, in both cases for when the variances are equal and unequal. We also derive a closed form expression of the upper bound of the proposed generalized p-value.
Generalized Chaplygin Gas and Varying Bulk Viscosity in Lyra Geometry
In this paper, we have considered Friedmann-Robertson-Walker (FRW) metric with generalized Chaplygin gas which has viscosity in the context of Lyra geometry. The viscosity is considered in two different ways (i.e. zero viscosity, non-constant r (rho)-dependent bulk viscosity) using constant deceleration parameter which concluded that, for a special case, the viscous generalized Chaplygin gas reduces to modified Chaplygin gas. The represented model indicates on the presence of Chaplygin gas in the Universe. Observational constraints are applied and discussed on the physical and geometrical nature of the Universe.
Weyl Type Theorem and the Fuglede Property
Given H a Hilbert space and B(H) the algebra of bounded linear operator in H, let δAB denote the generalized derivation defined by A and B. The main objective of this article is to study Weyl type theorems for generalized derivation for (A,B) satisfying a couple of Fuglede.
Local Homology Modules
In this paper, we give several ways for computing generalized local homology modules by using Gorenstein flat resolutions. Also, we find some bounds for vanishing of generalized local homology modules.
On the Analysis of Strategies of Buechi Games
In this paper, we present some results of simultaneous infinite games. We mainly work with generalized reachability games and Buechi games. These games are two-player concurrent games where each player chooses simultaneously their moves at each step. Our goal is to give simple expressions of values for each game. Moreover, we are interested in the question of what type of optimal (ε-optimal) strategy exists for both players depending on the type of games. We first show the determinacy (optimal value) and optimal (ε-optimal) strategies in generalized reachability games. We provide a simple expressions of value of this game and prove the existence of memoryless randomized ε-optimal strategy for Player I in any generalized reachability games. We then observe games with more complex objectives, games with Buechi objectives. We present how to compute an ε-optimal strategies and approximate a value of game in some way. Specifically, the results of generalized reachability games are used to show the value of Buechi games can be approximated as values of some generalized reachability games.
On Fourier Type Integral Transform for a Class of Generalized Quotients
In this paper, we investigate certain spaces of generalized functions for the Fourier and Fourier type integral transforms. We discuss convolution theorems and establish certain spaces of distributions for the considered integrals. The new Fourier type integral is well-defined, linear, one-to-one and continuous with respect to certain types of convergences. Many properties and an inverse problem are also discussed in some details.
Durrmeyer Type Modification of q-Generalized Bernstein Operators
The purpose of this paper to introduce the Durrmeyer type modification of q-generalized-Bernstein operators which include the Bernstein polynomials in the particular α = 0. We investigate the rate of convergence by means of the Lipschitz class and the Peetre’s K-functional. Also, we define the bivariate case of Durrmeyer type modification of q-generalized-Bernstein operators and study the degree of approximation with the aid of the partial modulus of continuity and the Peetre’s K-functional. Finally, we introduce the GBS (Generalized Boolean Sum) of the Durrmeyer type modification of q- generalized-Bernstein operators and investigate the approximation of the Bögel continuous and Bögel differentiable functions with the aid of the Lipschitz class and the mixed modulus of smoothness.
Spectral Clustering from the Discrepancy View and Generalized Quasirandomness
The aim of this paper is to compare spectral, discrepancy, and degree properties of expanding graph sequences. As we can prove equivalences and implications between them and the definition of the generalized (multiclass) quasirandomness of Lovasz–Sos (2008), they can be regarded as generalized quasirandom properties akin to the equivalent quasirandom properties of the seminal Chung-Graham-Wilson paper (1989) in the one-class scenario. Since these properties are valid for deterministic graph sequences, irrespective of stochastic models, the partial implications also justify for low-dimensional embedding of large-scale graphs and for discrepancy minimizing spectral clustering.
An Extension of the Generalized Extreme Value Distribution
A q-analogue of the generalized extreme value distribution which includes the Gumbel distribution is introduced. The additional parameter q allows for increased modeling flexibility. The resulting distribution can have a finite, semi-infinite or infinite support. It can also produce several types of hazard rate functions. The model parameters are determined by making use of the method of maximum likelihood. It will be shown that it compares favourably to three related distributions in connection with the modeling of a certain hydrological data set.
Method of Estimating Absolute Entropy of Municipal Solid Waste
Entropy, as an outcome of the second law of thermodynamics, measures the level of irreversibility associated with any process. The identification and reduction of irreversibility in the energy conversion process helps to improve the efficiency of the system. The entropy of pure substances known as absolute entropy is determined at an absolute reference point and is useful in the thermodynamic analysis of chemical reactions; however, municipal solid waste (MSW) is a structurally complicated material with unknown absolute entropy. In this work, an empirical model to calculate the absolute entropy of MSW based on the content of carbon, hydrogen, oxygen, nitrogen, sulphur, and chlorine on a dry ash free basis (daf) is presented. The proposed model was derived from 117 relevant organic substances which represent the main constituents in MSW with known standard entropies using statistical analysis. The substances were divided into different waste fractions; namely, food, wood/paper, textiles/rubber and plastics waste and the standard entropies of each waste fraction and for the complete mixture were calculated. The correlation of the standard entropy of the complete waste mixture derived was found to be somsw= 0.0101C + 0.0630H + 0.0106O + 0.0108N + 0.0155S + 0.0084Cl ( and the present correlation can be used for estimating the absolute entropy of MSW by using the elemental compositions of the fuel within the range of 10.3% ≤ C ≤ 95.1%, 0.0% ≤ H ≤ 14.3%, 0.0% ≤ O ≤ 71.1%, 0.0 ≤ N ≤ 66.7%, 0.0% ≤ S ≤ 42.1%, 0.0% ≤ Cl ≤ 89.7%. The model is also applicable for the efficient modelling of a combustion system in a waste-to-energy plant.
Convergence of Generalized Jacobi, Gauss-Seidel and Successive Overrelaxation Methods for Various Classes of Matrices
Generalized Jacobi (GJ) and Generalized Gauss-Seidel (GGS) methods are most effective than conventional Jacobi and Gauss-Seidel methods for solving linear system of equations. It is known that GJ and GGS methods converge for strictly diagonally dominant (SDD) and for M-matrices. In this paper, we study the convergence of GJ and GGS converge for symmetric positive definite (SPD) matrices, L-matrices and H-matrices. We introduce a generalization of successive overrelaxation (SOR) method for solving linear systems and discuss its convergence for the classes of SDD matrices, SPD matrices, M-matrices, L-matrices and for H-matrices. Advantages of generalized SOR method are established through numerical experiments over GJ, GGS, and SOR methods.
Characterization of Probability Distributions through Conditional Expectation of Pair of Generalized Order Statistics
In this article, first a relation for conditional expectation is developed and then is used to characterize a general class of distributions F(x) = 1-e^(-ah(x)) through conditional expectation of difference of pair of generalized order statistics. Some results are reduced for particular cases. In the end, a list of distributions is presented in the form of table that are compatible with the given general class.
Generalized Dirac oscillators Associated to Non-Hermitian Quantum Mechanical Systems
In recent years, non Hermitian interaction in non relativistic as well as relativistic quantum mechanics have been examined from various aspect. We can observe interesting fact that for such systems a class of potentials, namely the PT symmetric and η-pseudo Hermitian admit real eigenvalues despite being non Hermitian and analogues of those system have been experimentally verified. Point to be noted that relativistic non Hermitian (PT symmetric) interactions can be realized in optical structures and also there exists photonic realization of the (1 + 1) dimensional Dirac oscillator. We have thoroughly studied generalized Dirac oscillators with non Hermitian interactions in (1 + 1) dimensions. To be more specific, we have examined η pseudo Hermitian interactions within the framework of generalized Dirac oscillator in (1 + 1) dimensions. In particular, we have obtained a class of interactions which are η-pseudo Hermitian and the metric operator η could have been also found explicitly. It is possible to have exact solutions of the generalized Dirac oscillator for some choices of the interactions. Subsequently we have employed the mapping between the generalized Dirac oscillator and the Jaynes Cummings (JC) model by spin flip to obtain a class of exactly solvable non Hermitian JC as well as anti Jaynes Cummings (AJC) type models.
Modeling of Maximum Rainfall Using Poisson-Generalized Pareto Distribution in Kigali, Rwanda
Extreme rainfall events have caused significant damage to agriculture, ecology, and infrastructure, disruption of human activities, injury, and loss of life. They also have significant social, economic, and environmental consequences because they considerably damage urban as well as rural areas. Early detection of extreme maximum rainfall helps to implement strategies and measures, before they occur, hence mitigating the consequences. Extreme value theory has been used widely in modeling extreme rainfall and in various disciplines, such as financial markets, the insurance industry, failure cases. Climatic extremes have been analyzed by using either generalized extreme value (GEV) or generalized Pareto (GP) distributions, which provides evidence of the importance of modeling extreme rainfall from different regions of the world. In this paper, we focused on Peak Over Thresholds approach, where the Poisson-generalized Pareto distribution is considered as the proper distribution for the study of the exceedances. This research also considers the use of the generalized Pareto (GP) distribution with a Poisson model for arrivals to describe peaks over a threshold. The research used statistical techniques to fit models that used to predict extreme rainfall in Kigali. The results indicate that the proposed Poisson-GP distribution provides a better fit to maximum monthly rainfall data. Further, the Poisson-GP models are able to estimate various return levels. The research also found a slow increase in return levels for maximum monthly rainfall for higher return periods, and further, the intervals are increasingly wider as the return period is increasing.
A Fast Version of the Generalized Multi-Directional Radon Transform
This paper presents a new fast version of the generalized Multi-Directional Radon Transform method. The new method uses the inverse Fast Fourier Transform to lead to a faster Generalized Radon projections. We prove in this paper that the fast algorithm leads to almost the same results of the eldest one but with a considerable lower time computation cost. The projection end result of the fast method is a parameterized Radon space where a high valued pixel allows the detection of a curve from the original image. The proposed fast inversion algorithm leads to an exact reconstruction of the initial image from the Radon space. We show examples of the impact of this algorithm on the pattern recognition domain.
The Generalized Pareto Distribution as a Model for Sequential Order Statistics
‎In this article‎, ‎sequential order statistics (SOS) censoring type II samples coming from the generalized Pareto distribution are considered‎. ‎Maximum likelihood (ML) estimators of the unknown parameters are derived on the basis of the available multiple SOS data‎. ‎Necessary conditions for existence and uniqueness of the derived ML estimates are given‎. Due to complexity in the proposed likelihood function‎, ‎a useful re-parametrization is suggested‎. ‎For illustrative purposes‎, ‎a Monte Carlo simulation study is conducted and an illustrative example is analysed‎.
A New Fixed Point Theorem for Almost θ-Contraction
In this work, we introduce a new type of contractive maps and we establish a new fixed point theorem for the class of almost θ-contractions (more general than the class of almost contractions) in a complete generalized metric space. The major novelty of our work is to prove a new fixed point result by weakening some hypotheses imposed on the function θ which will change completely the classical technique used in the literature review to prove fixed point theorems for almost θ-contractions in a complete generalized metric space.
A Local Invariant Generalized Hough Transform Method for Integrated Circuit Visual Positioning
In this study, an local invariant generalized Houghtransform (LI-GHT) method is proposed for integrated circuit (IC) visual positioning. The original generalized Hough transform (GHT) is robust to external noise; however, it is not suitable for visual positioning of IC chips due to the four-dimensionality (4D) of parameter space which leads to the substantial storage requirement and high computational complexity. The proposed LI-GHT method can reduce the dimensionality of parameter space to 2D thanks to the rotational invariance of local invariant geometric feature and it can estimate the accuracy position and rotation angle of IC chips in real-time under noise and blur influence. The experiment results show that the proposed LI-GHT can estimate position and rotation angle of IC chips with high accuracy and fast speed. The proposed LI-GHT algorithm was implemented in IC visual positioning system of radio frequency identification (RFID) packaging equipment.
A Generalized Model for Performance Analysis of Airborne Radar in Clutter Scenario
Performance prediction of airborne radar is a challenging and cumbersome task in clutter scenario for different types of targets. A generalized model requires to predict the performance of Radar for air targets as well as ground moving targets. In this paper, we propose a generalized model to bring out the performance of airborne radar for different Pulsed Repetition Frequency (PRF) as well as different type of targets. The model provides a platform to bring out different subsystem parameters for different applications and performance requirements under different types of clutter terrain.
Generalized Extreme Value Regression with Binary Dependent Variable: An Application for Predicting Meteorological Drought Probabilities
Logistic regression model is the most used regression model to predict meteorological drought probabilities. When the dependent variable is extreme, the logistic model fails to adequately capture drought probabilities. In order to adequately predict drought probabilities, we use the generalized linear model (GLM) with the quantile function of the generalized extreme value distribution (GEVD) as the link function. The method maximum likelihood estimation is used to estimate the parameters of the generalized extreme value (GEV) regression model. We compare the performance of the logistic and the GEV regression models in predicting drought probabilities for Zimbabwe. The performance of the regression models are assessed using the goodness-of-fit tests, namely; relative root mean square error (RRMSE) and relative mean absolute error (RMAE). Results show that the GEV regression model performs better than the logistic model, thereby providing a good alternative candidate for predicting drought probabilities. This paper provides the first application of GLM derived from extreme value theory to predict drought probabilities for a drought-prone country such as Zimbabwe.
Parameter Estimation for the Mixture of Generalized Gamma Model
Mixture generalized gamma distribution is a combination of two distributions: generalized gamma distribution and length biased generalized gamma distribution. These two distributions were presented by Suksaengrakcharoen and Bodhisuwan in 2014. The findings showed that probability density function (pdf) had fairly complexities, so it made problems in estimating parameters. The problem occurred in parameter estimation was that we were unable to calculate estimators in the form of critical expression. Thus, we will use numerical estimation to find the estimators. In this study, we presented a new method of the parameter estimation by using the expectation – maximization algorithm (EM), the conjugate gradient method, and the quasi-Newton method. The data was generated by acceptance-rejection method which is used for estimating α, β, λ and p. λ is the scale parameter, p is the weight parameter, α and β are the shape parameters. We will use Monte Carlo technique to find the estimator's performance. Determining the size of sample equals 10, 30, 100; the simulations were repeated 20 times in each case. We evaluated the effectiveness of the estimators which was introduced by considering values of the mean squared errors and the bias. The findings revealed that the EM-algorithm had proximity to the actual values determined. Also, the maximum likelihood estimators via the conjugate gradient and the quasi-Newton method are less precision than the maximum likelihood estimators via the EM-algorithm.
Some Generalized Multivariate Estimators for Population Mean under Multi Phase Stratified Systematic Sampling
The generalized multivariate ratio and regression type estimators for population mean are suggested under multi-phase stratified systematic sampling (MPSSS) using multi auxiliary information. Estimators are developed under the two different situations of availability of auxiliary information. The expressions of bias and mean square error (MSE) are developed. Special cases of suggested estimators are also discussed and simulation study is conducted to observe the performance of estimators.
Application of a Generalized Additive Model to Reveal the Relations between the Density of Zooplankton with Other Variables in the West Daya Bay, China
Zooplankton are a central issue in the ecology which makes a great contribution to maintaining the balance of an ecosystem. It is critical in promoting the material cycle and energy flow within the ecosystems. A generalized additive model (GAM) was applied to analyze the relationships between the density (individuals per m³) of zooplankton and other variables in West Daya Bay. All data used in this analysis (the survey month, survey station (longitude and latitude), the depth of the water column, the superficial concentration of chlorophyll a, the benthonic concentration of chlorophyll a, the number of zooplankton species and the number of zooplankton species) were collected through monthly scientific surveys during January to December 2016. GLM model (generalized linear model) was used to choose the significant variables’ impact on the density of zooplankton, and the GAM was employed to analyze the relationship between the density of zooplankton and the significant variables. The results showed that the density of zooplankton increased with an increase of the benthonic concentration of chlorophyll a, but decreased with a decrease in the depth of the water column. Both high numbers of zooplankton species and the overall total number of zooplankton individuals led to a higher density of zooplankton.
A New Approach for Generalized First Derivative of Nonsmooth Functions Using Optimization
In this paper, we define an optimization problem corresponding to smooth and nonsmooth functions which its optimal solution is the first derivative of these functions in a domain. For this purpose, a linear programming problem corresponding to optimization problem is obtained. The optimal solution of this linear programming problem is the approximate generalized first derivative. In fact, we approximate generalized first derivative of nonsmooth functions as tailor series. We show the efficiency of our approach by some smooth and nonsmooth functions in some examples.
Generalized Mean-Field Theory of Phase Unwrapping via Multiple Interferograms
On the basis of Bayesian inference using the maximizer of the posterior marginal estimate, we carry out phase unwrapping using multiple interferograms via generalized mean-field theory. Numerical calculations for a typical wave-front in remote sensing using the synthetic aperture radar interferometry, phase diagram in hyper-parameter space clarifies that the present method succeeds in phase unwrapping perfectly under the constraint of surface- consistency condition, if the interferograms are not corrupted by any noises. Also, we find that prior is useful for extending a phase in which phase unwrapping under the constraint of the surface-consistency condition. These results are quantitatively confirmed by the Monte Carlo simulation.
Explicit Iterative Scheme for Approximating a Common Solution of Generalized Mixed Equilibrium Problem and Fixed Point Problem for a Nonexpansive Semigroup in Hilbert Space
In this paper, we introduce and study an explicit iterative method based on hybrid extragradient method to approximate a common solution of generalized mixed equilibrium problem and fixed point problem for a nonexpansive semigroup in Hilbert space. Further, we prove that the sequence generated by the proposed iterative scheme converge strongly to the common solution of generalized mixed equilibrium problem and fixed point problem for a nonexpansive semigroup. This common solution is the unique solution of a variational inequality problem and is the optimality condition for a minimization problem. The results presented in this paper are the supplement, extension and generalization of the previously known results in this area.
Investigation on Machine Tools Energy Consumptions
Several researches have been conducted to study consumption of energy in cutting process. Most of these researches are focusing to measure the consumption and propose consumption reduction methods. In this work, the relation between the cutting parameters and the consumption is investigated in order to establish a generalized energy consumption model that can be used for process and production planning in real production lines. Using the generalized model, the process planning will be carried out by taking into account the energy as a function of the selected process parameters. Similarly, the generalized model can be used in production planning to select the right operational parameters like batch sizes, routing, buffer size, etc. in a production line. The description and derivation of the model as well as a case study are given in this paper to illustrate the applicability and validity of the model.
A Comparison of Transdiagnostic Components in Generalized Anxiety Disorder, Unipolar Mood Disorder and Nonclinical Population
Background: Dimensional and transdiagnostic approaches as a result of high comorbidity among mental disorders have captured researchers and clinicians interests for exploring the latent factors of development and maintenance of some psychological disorders. The goal of present study is to compare some of these common factors between generalized anxiety disorder and unipolar mood disorder. Methods: 27 patients with generalized anxiety disorder, 29 patients with depression disorder were recruited using SCID-I and 69 non-clinical population were selected using GHQ cut off point. MANCOVA was used for analyzing data. Results: The results show that worry, rumination, intolerance of uncertainty, maladaptive metacognitive beliefs, and experiential avoidance were all significantly different between GAD and unipolar mood disorder groups. However, there were not any significant differences in difficulties in emotion regulation and neuroticism between GAD and unipolar mood disorder groups. Discussion: Results indicate that although there are some transdiagnostic and common factors in GAD and unipolar mood disorder, there may be some specific vulnerability factors for each disorder. Further study is needed for answering these questions.
Object Recognition Approach Based on Generalized Hough Transform and Color Distribution Serving in Generating Arabic Sentences
The recognition of the objects contained in images has always presented a challenge in the field of research because of several difficulties that the researcher can envisage because of the variability of shape, position, contrast of objects, etc. In this paper, we will be interested in the recognition of objects. The classical Hough Transform (HT) presented a tool for detecting straight line segments in images. The technique of HT has been generalized (GHT) for the detection of arbitrary forms. With GHT, the forms sought are not necessarily defined analytically but rather by a particular silhouette. For more precision, we proposed to combine the results from the GHT with the results from a calculation of similarity between the histograms and the spatiograms of the images. The main purpose of our work is to use the concepts from recognition to generate sentences in Arabic that summarize the content of the image.
Pressure Distribution, Load Capacity, and Thermal Effect with Generalized Maxwell Model in Journal Bearing Lubrication
This numerical investigation aims to evaluate how a viscoelastic lubricant described by a generalized Maxwell model, affects the pressure distribution, the load capacity and thermal effect in a journal bearing lubrication. We use for the purpose the CFD package software completed by adapted user define functions (UDFs) to solve the coupled equations of momentum, of energy and of the viscoelastic model (generalized Maxwell model). Two parameters, viscosity and relaxation time are involved to show how viscoelasticity substantially affect the pressure distribution, the load capacity and the thermal transfer by comparison to Newtonian lubricant. These results were also compared with the available published results.
Population Size Estimation Based on the GPD
The purpose of the study is to estimate the elusive target population size under a truncated count model that accounts for heterogeneity. The purposed estimator is based on the generalized Poisson distribution (GPD), which extends the Poisson distribution by adding a dispersion parameter. Thus, it becomes an useful model for capture-recapture data where concurrent events are not homogeneous. In addition, it can account for over-dispersion and under-dispersion. The ratios of neighboring frequency counts are used as a tool for investigating the validity of whether generalized Poisson or Poisson distribution. Since capture-recapture approaches do not provide the zero counts, the estimated parameters can be achieved by modifying the EM-algorithm technique for the zero-truncated generalized Poisson distribution. The properties and the comparative performance of proposed estimator were investigated through simulation studies. Furthermore, some empirical examples are represented insights on the behavior of the estimators.
On a Single Server Queue with Arrivals in Batches of Variable Size, Generalized Coxian-2 Service and Compulsory Server Vacations
We study the steady state behaviour of a batch arrival single server queue in which the first service with general service times is compulsory and the second service with general service times is optional. We term such a two phase service as generalized Coxian-2 service. Just after completion of a service the server must take a vacation of random length of time with general vacation times. We obtain steady state probability generating functions for the queue size as well as the steady state mean queue size at a random epoch of time in explicit and closed forms. Some particular cases of interest including some known results have been derived.
A Fundamental Functional Equation for Lie Algebras
Inspired by the so called Jacobi Identity (x y) z + (y z) x + (z x) y = 0, the following class of functional equations EQ I: F [F (x, y), z] + F [F (y, z), x] + F [F (z, x), y] = 0 is proposed, researched and generalized. Research methodologies begin with classical methods for functional equations, then evolve into discovering of any implicit algebraic structures. One of this paper’s major findings is that EQ I, under two additional conditions F (x, x) = 0 and F (x, y) + F (y, x) = 0, proves to be a fundamental functional equation for Lie Algebras. Existence of non-trivial solutions for EQ I can be proven by defining F (p, q) = [p q] = pq –qp, where p and q are quaternions, and pq is the quaternion product of p and q. EQ I can be generalized to the following class of functional equations EQ II: F [G (x, y), z] + F [G (y, z), x] + F [G (z, x), y] = 0. Concluding Statement: With a major finding proven, and non-trivial solutions derived, this research paper illustrates and provides a new functional equation scheme for studies in two major areas: (1) What underlying algebraic structures can be defined and/or derived from EQ I or EQ II? (2) What conditions can be imposed so that conditional general solutions to EQ I and EQ II can be found, investigated and applied?
Analysis of Risk Factors Affecting the Motor Insurance Pricing with Generalized Linear Models
Casualty insurance business, the optimal premium pricing and adequate cost for an insurance company are important in risk management. Normally, the insurance pure premium can be determined by multiplying the claim frequency with the claim cost. The aim of this research was to study in the application of generalized linear models to select the risk factor for model of claim frequency and claim cost for estimating a pure premium. In this study, the data set was the claim of comprehensive motor insurance, which was provided by one of the insurance company in Thailand. The results of this study found that the risk factors significantly related to pure premium at the 0.05 level consisted of no claim bonus (NCB) and used of the car (Car code).
Impact of Weather Conditions on Generalized Frequency Division Multiplexing over Gamma Gamma Channel
The technique called as Generalized frequency division multiplexing (GFDM) used in the free space optical channel can be a good option for implementation free space optical communication systems. This technique has several strengths e.g. good spectral efficiency, low peak-to-average power ratio (PAPR), adaptability and low co-channel interference. In this paper, the impact of weather conditions such as haze, rain and fog on GFDM over the gamma-gamma channel model is discussed. A Trade off between link distance and system performance under intense weather conditions is also analysed. The symbol error probability (SEP) of GFDM over the gamma-gamma turbulence channel is derived and verified with the computer simulations.
Maximum Entropy Based Image Segmentation of Human Skin Lesion
Image segmentation plays an important role in medical imaging applications. Therefore, accurate methods are needed for the successful segmentation of medical images for diagnosis and detection of various diseases. In this paper, we have used maximum entropy to achieve image segmentation. Maximum entropy has been calculated using Shannon, Renyi, and Tsallis entropies. This work has novelty based on the detection of skin lesion caused by the bite of a parasite called Sand Fly causing the disease is called Cutaneous Leishmaniasis.
Development of Generalized Correlation for Liquid Thermal Conductivity of N-Alkane and Olefin
The objective of this research is to develop a generalized correlation for the prediction of thermal conductivity of n-Alkanes and Alkenes. There is a minority of research and lack of correlation for thermal conductivity of liquids in the open literature. The available experimental data are collected covering the groups of n-Alkanes and Alkenes.The data were assumed to correlate to temperature using Filippov correlation. Nonparametric regression of Grace Algorithm was used to develop the generalized correlation model. A spread sheet program based on Microsoft Excel was used to plot and calculate the value of the coefficients. The results obtained were compared with the data that found in Perry's Chemical Engineering Hand Book. The experimental data correlated to the temperature ranged "between" 273.15 to 673.15 K, with R2 = 0.99.The developed correlation reproduced experimental data that which were not included in regression with absolute average percent deviation (AAPD) of less than 7 %. Thus the spread sheet was quite accurate which produces reliable data.
A Generalization of Planar Pascal’s Triangle to Polynomial Expansion and Connection with Sierpinski Patterns
The very well-known stacked sets of numbers referred to as Pascal’s triangle present the coefficients of the binomial expansion of the form (x+y)n. This paper presents an approach (the Staircase Horizontal Vertical, SHV-method) to the generalization of planar Pascal’s triangle for polynomial expansion of the form (x+y+z+w+r+⋯)n. The presented generalization of Pascal’s triangle is different from other generalizations of Pascal’s triangles given in the literature. The coefficients of the generalized Pascal’s triangles, presented in this work, are generated by inspection, using embedded Pascal’s triangles. The coefficients of I-variables expansion are generated by horizontally laying out the Pascal’s elements of (I-1) variables expansion, in a staircase manner, and multiplying them with the relevant columns of vertically laid out classical Pascal’s elements, hence avoiding factorial calculations for generating the coefficients of the polynomial expansion. Furthermore, the classical Pascal’s triangle has some pattern built into it regarding its odd and even numbers. Such pattern is known as the Sierpinski’s triangle. In this study, a presentation of Sierpinski-like patterns of the generalized Pascal’s triangles is given. Applications related to those coefficients of the binomial expansion (Pascal’s triangle), or polynomial expansion (generalized Pascal’s triangles) can be in areas of combinatorics, and probabilities.
An Efficient Algorithm of Time Step Control for Error Correction Method
The aim of this paper is to construct an algorithm of time step control for the error correction method most recently developed by one of the authors for solving stiff initial value problems. It is achieved with the generalized Chebyshev polynomial and the corresponding error correction method. The main idea of the proposed scheme is in the usage of the duplicated node points in the generalized Chebyshev polynomials of two different degrees by adding necessary sample points instead of re-sampling all points. At each integration step, the proposed method is comprised of two equations for the solution and the error, respectively. The constructed algorithm controls both the error and the time step size simultaneously and possesses a good performance in the computational cost compared to the original method. Two stiff problems are numerically solved to assess the effectiveness of the proposed scheme.
Generalized Uncertainty Principle Modified Hawking Radiation in Bumblebee Gravity
The effect of Lorentz symmetry breaking (LSB) on the Hawking radiation of Schwarzschild-like black hole found in the bumblebee gravity model (SBHBGM) is studied in the framework of quantum gravity. To this end, we consider Hawking radiation spin-0 (bosons) and spin-12particles (fermions), which go in and out through the event horizon of the SBHBGM. We use the modified Klein-Gordon and Dirac equations, which are obtained from the generalized uncertainty principle (GUP) to show how Hawking radiation is affected by the GUP and LSB. In particular, we reveal that independent of the spin of the emitted particles, GUP causes a change in the Hawking temperature of the SBHBGM. Furthermore, we compute the semi-analytic greybody factors (for both bosons and fermions) of the SBHBGM. Thus, we reveal that LSB is effective on the greybody factor of the SBHBGM such that its redundancy decreases the value of the greybody factor. Our findings are graphically depicted.
A Flexible Pareto Distribution Using α-Power Transformation
In Statistical Distribution Theory, considering an additional parameter to classical distributions is a usual practice. In this study, a new distribution referred to as α-Power Pareto distribution is introduced by including an extra parameter. Several properties of the proposed distribution including explicit expressions for the moment generating function, mode, quantiles, entropies and order statistics are obtained. Unknown parameters have been estimated by using maximum likelihood estimation technique. Two real datasets have been considered to examine the usefulness of the proposed distribution. It has been observed that α-Power Pareto distribution outperforms while compared to different variants of Pareto distribution on the basis of model selection criteria.
A Model for Operating Rooms Scheduling
This paper presents a mathematical model in binary variables 0/1 to make the assignment of surgical procedures to the operating rooms in a hospital. The proposed mathematical model is based on the generalized assignment problem, which maximizes the sum of preferences for the use of the operating rooms by doctors, respecting the time available in each room. The corresponding program was written in Visual Basic of Microsoft Excel, and tested to schedule surgeries at St. Lydia Hospital in Ribeirao Preto, Brazil.
Extreme Temperature Forecast in Mbonge, Cameroon Through Return Level Analysis of the Generalized Extreme Value (GEV) Distribution
In this paper, temperature extremes are forecast by employing the block maxima method of the generalized extreme value (GEV) distribution to analyse temperature data from the Cameroon Development Corporation (CDC). By considering two sets of data (raw data and simulated data) and two (stationary and non-stationary) models of the GEV distribution, return levels analysis is carried out and it was found that in the stationary model, the return values are constant over time with the raw data, while in the simulated data the return values show an increasing trend with an upper bound. In the non-stationary model, the return levels of both the raw data and simulated data show an increasing trend with an upper bound. This clearly shows that although temperatures in the tropics show a sign of increase in the future, there is a maximum temperature at which there is no exceedance. The results of this paper are very vital in agricultural and environmental research.
Leverage Effect for Volatility with Generalized Laplace Error
We propose a new model that accounts for the asymmetric response of volatility to positive ('good news') and negative ('bad news') shocks in economic time series the so-called leverage effect. In the past, asymmetric powers of errors in the conditionally heteroskedastic models have been used to capture this effect. Our model is using the gamma difference representation of the generalized Laplace distributions that efficiently models the asymmetry. It has one additional natural parameter, the shape, that is used instead of power in the asymmetric power models to capture the strength of a long-lasting effect of shocks. Some fundamental properties of the model are provided including the formula for covariances and an explicit form for the conditional distribution of 'bad' and 'good' news processes given the past the property that is important for the statistical fitting of the model. Relevant features of volatility models are illustrated using S&P 500 historical data.
Numerical Solutions of Generalized Burger-Fisher Equation by Modified Variational Iteration Method
Numerical solutions of the generalized Burger-Fisher are obtained using a Modified Variational Iteration Method (MVIM) with minimal computational efforts. The computed results with this technique have been compared with other results. The present method is seen to be a very reliable alternative method to some existing techniques for such nonlinear problems.
Risk Factors for Defective Autoparts Products Using Bayesian Method in Poisson Generalized Linear Mixed Model
This research investigates risk factors for defective products in autoparts factories. Under a Bayesian framework, a generalized linear mixed model (GLMM) in which the dependent variable, the number of defective products, has a Poisson distribution is adopted. Its performance is compared with the Poisson GLM under a Bayesian framework. The factors considered are production process, machines, and workers. The products coded RT50 are observed. The study found that the Poisson GLMM is more appropriate than the Poisson GLM. For the production Process factor, the highest risk of producing defective products is Process 1, for the Machine factor, the highest risk is Machine 5, and for the Worker factor, the highest risk is Worker 6.
Numerical Solutions of Boundary Layer Flow over an Exponentially Stretching/Shrinking Sheet with Generalized Slip Velocity
In this paper, the problem of steady laminar boundary layer flow and heat transfer over a permeable exponentially stretching/shrinking sheet with generalized slip velocity is considered. The similarity transformations are used to transform the governing nonlinear partial differential equations to a system of nonlinear ordinary differential equations. The transformed equations are then solved numerically using the bvp4c function in MATLAB. Dual solutions are found for a certain range of the suction and stretching/shrinking parameters. The effects of the suction parameter, stretching/shrinking parameter, velocity slip parameter, critical shear rate, and Prandtl number on the skin friction and heat transfer coefficients as well as the velocity and temperature profiles are presented and discussed.
A Dual Generalized Long Memory Modeling for Forecasting Electricity Spot Price: Neural Network and Wavelet Estimate
In this paper, a dual generalized long memory modeling has been proposed to predict the electricity spot price. First, we focus on the modeling of the conditional mean of the series so we adopt a generalized fractional model with k factor of Gegenbauer (k-factor GARMA). Secondly, the residual from the k-factor GARMA model has been used as a proxy for the conditional variance; these residuals were predicted using two different approaches. In the first approach, a local linear wavelet neural network model (LLWNN) has been developed to predict the conditional variance using two different learning algorithms, so we estimate the hybrid k- factor GARMA-LLWNN based Back Propagation (BP) algorithm and based Particle Swarm Optimization (PSO) algorithm. In the second approach, the Gegenbauer generalized autoregressive conditional heteroscedasticity process (G-GARCH) has been adopted, and the parameters of GARMA-G-GARCH model estimated using the wavelet methodology based on the discrete wavelet packet transform (DWPT). To illustrate the usefulness of our methodology, we carry out an empirical application using the hourly returns of electricity price from the Nord Pool market. The empirical results showed that the k-factor GARMA-G-GARCH model has the best prediction accuracy in terms of forecasting criteria, and find that this is more appropriate for forecasts.
A Dual Generalized Long Memory Modeling for Forecasting Electricity Spot Price: Neural Network and Wavelet Estimate
In this paper, a dual generalized long memory modeling has been proposed to predict the electricity spot price. First, we focus on the modeling of the conditional mean of the series so we adopt a generalized fractional model with k factor of Gegenbauer (k-factor GARMA). Secondly, the residual from the k-factor GARMA model has been used as a proxy for the conditional variance; these residuals were predicted using two different approaches. In the first approach, a local linear wavelet neural network model (LLWNN) has been developed to predict the conditional variance using two different learning algorithms, so we estimate the hybrid k- factor GARMA-LLWNN based Back Propagation (BP) algorithm and based Particle Swarm Optimization (PSO) algorithm. In the second approach, the Gegenbauer generalized autoregressive conditional heteroscedasticity process (G-GARCH) has been adopted, and the parameters of GARMA-G-GARCH model estimated using the wavelet methodology based on the discrete wavelet packet transform (DWPT). To illustrate the usefulness of our methodology, we carry out an empirical application using the hourly returns of electricity price from the Nord Pool market. The empirical results showed that the k-factor GARMA-G-GARCH model has the best prediction accuracy in terms of forecasting criteria, and find that this is more appropriate for forecasts.
A Dual Generalized Long Memory Modeling for Forecasting Electricity Spot Price: Neural Network and Wavelet Estimate
In this paper, a dual generalized long memory modeling has been proposed to predict the electricity spot price. First, we focus on the modeling of the conditional mean of the series so we adopt a generalized fractional model with k factor of Gegenbauer (k-factor GARMA). Secondly, the residual from the k-factor GARMA model has been used as a proxy for the conditional variance; these residuals were predicted using two different approaches. In the first approach, a local linear wavelet neural network model (LLWNN) has been developed to predict the conditional variance using two different learning algorithms, so we estimate the hybrid k- factor GARMA-LLWNN based Back Propagation (BP) algorithm and based Particle Swarm Optimization (PSO) algorithm. In the second approach, the Gegenbauer generalized autoregressive conditional heteroscedasticity process (G-GARCH) has been adopted, and the parameters of GARMA-G-GARCH model estimated using the wavelet methodology based on the discrete wavelet packet transform (DWPT). To illustrate the usefulness of our methodology, we carry out an empirical application using the hourly returns of electricity price from the Nord Pool market. The empirical results showed that the k-factor GARMA-G-GARCH model has the best prediction accuracy in terms of forecasting criteria, and find that this is more appropriate for forecasts.
Forecasting Electricity Spot Price with Generalized Long Memory Modeling: Wavelet and Neural Network
This aims of this paper is to forecast the electricity spot prices. First, we focus on modeling the conditional mean of the series so we adopt a generalized fractional -factor Gegenbauer process (k-factor GARMA). Secondly, the residual from the -factor GARMA model has used as a proxy for the conditional variance; these residuals were predicted using two different approaches. In the first approach, a local linear wavelet neural network model (LLWNN) has developed to predict the conditional variance using the Back Propagation learning algorithms. In the second approach, the Gegenbauer generalized autoregressive conditional heteroscedasticity process (G-GARCH) has adopted, and the parameters of the k-factor GARMA-G-GARCH model has estimated using the wavelet methodology based on the discrete wavelet packet transform (DWPT) approach. The empirical results have shown that the k-factor GARMA-G-GARCH model outperform the hybrid k-factor GARMA-LLWNN model, and find it is more appropriate for forecasts.
Finding the Elastic Field in an Arbitrary Anisotropic Media by Implementing Accurate Generalized Gaussian Quadrature Solution
In the current study, the elastic field in an anisotropic elastic media is determined by implementing a general semi-analytical method. In this specific methodology, the displacement field is computed as a sum of finite functions with unknown coefficients. These aforementioned functions satisfy exactly both the homogeneous and inhomogeneous boundary conditions in the proposed media. It is worth mentioning that the unknown coefficients are determined by implementing the principle of minimum potential energy. The numerical integration is implemented by employing the Generalized Gaussian Quadrature solution. Furthermore, with the aid of the calculated unknown coefficients, the displacement field, as well as the other parameters of the elastic field, are obtainable as well. Finally, the comparison of the previous analytical method with the current semi-analytical method proposes the efficacy of the present methodology.
Backtesting a One-Step Ahead Density Predictions of Value-at-Risk
This study estimates a method for backtesting density forecasts that are based on a weighted threshold of a continuously ranked probability score. The weighting emphasizes regions of interest, such as the tails or the center of a variable’s range, while retaining propriety, as opposed to a recently developed weighted likelihood ratio test, which can be hedged, is emphasized in this study. Threshold-based decompositions of a continuously ranked probability score is illustrated and prompt insights into strengths and deficiencies of a forecasting method. A time-varying model with combined extreme values that is based on the score function of a predictive density at time is estimated. The mechanism to update the parameters of this model overtime is based on the scaled score of a likelihood function. The developed backtesting procedure revealed that the estimated generalized autoregressive score generalized extreme value distribution (GAS-GEVD) model with a skewed student-t distribution has the best prediction performance in forecasting the value-at-risk (VaR). Extension of this non-stationary distribution in literature is quite complicated since it requires specifications not only on how the usual Bayesian parameters change over time but also those with bulk distribution components. This implies that the combination of a stochastic econometric model with extreme value theory (EVT) procedures provides a robust basis necessary for the statistical backtesting density predictions for Value-at-risk (VaR).
Duality in Multiobjective Nonlinear Programming under Generalized Second Order (F, b, φ, ρ, θ)− Univex Functions
In the present paper, second order duality for multiobjective nonlinear programming are investigated under the second order generalized (F, b, φ, ρ, θ)− univex functions. The weak, strong and converse duality theorems are proved. Further, we also illustrated an example of (F, b, φ, ρ, θ)− univex functions. Results obtained in this paper extend some previously known results of multiobjective nonlinear programming in the literature.
Approximate Solution of Some Mixed Boundary Value Problems of the Generalized Theory of Couple-Stress Thermo-Elasticity
We have considered the harmonic oscillations and general dynamic (pseudo oscillations) systems of theory generalized Green-Lindsay of couple-stress thermo-elasticity for isotropic, homogeneous elastic media. Approximate solution of some mixed boundary value problems for finite domain, bounded by the some closed surface are constructed.
On Projective Invariants of Spherically Symmetric Finsler Spaces in Rn
In this paper we study projective invariants of spherically symmetric Finsler metrics in Rn. We find the necessary and sufficient conditions for the metrics to be Douglas and Generalized Douglas-Weyl (GDW) types. Also we show that two classes of GDW and Douglas spherically symmetric Finsler metrics coincide.
An Application of Sinc Function to Approximate Quadrature Integrals in Generalized Linear Mixed Models
This paper discusses a novel approach to approximate quadrature integrals that arise in the estimation of likelihood parameters for the generalized linear mixed models (GLMM) as well as Bayesian methodology also requires computation of multidimensional integrals with respect to the posterior distributions in which computation are not only tedious and cumbersome rather in some situations impossible to find solutions because of singularities, irregular domains, etc. An attempt has been made in this work to apply Sinc function based quadrature rules to approximate intractable integrals, as there are several advantages of using Sinc based methods, for example: order of convergence is exponential, works very well in the neighborhood of singularities, in general quite stable and provide high accurate and double precisions estimates. The Sinc function based approach seems to be utilized first time in statistical domain to our knowledge, and it's viability and future scopes have been discussed to apply in the estimation of parameters for GLMM models as well as some other statistical areas.
Balancing a Rotary Inverted Pendulum System Using Robust Generalized Dynamic Inverse: Design and Experiment
This paper presents a methodology for balancing a rotary inverted pendulum system using Robust Generalized Dynamic Inversion (RGDI) under influence of parametric variations and external disturbances. In GDI control, dynamic constraints are formulated in the form of asymptotically stable differential equation which encapsulates the control objectives. The constraint differential equations are based on the deviation function of the angular position and its rates from their reference values. The constraint dynamics are inverted using Moore-Penrose Generalized Inverse (MPGI) to realize the control expression. The GDI singularity problem is addressed by augmenting a dynamic scale factor in the interpretation of MPGI which guarantee asymptotically stable position tracking. An additional term based on Sliding Mode Control is appended within GDI control to make it robust against parametric variations, disturbances and tracking performance deterioration due to generalized inversion scaling. The stability of the closed loop system is ensured by using positive definite Lyapunov energy function that guarantees semi-global practically stable position tracking. Numerical simulations are conducted on the dynamic model of rotary inverted pendulum system to analyze the efficiency of proposed RGDI control law. The comparative study is also presented, in which the performance of RGDI control is compared with Linear Quadratic Regulator (LQR) and is verified through experiments. Numerical simulations and real-time experiments demonstrate better tracking performance abilities and robustness features of RGDI control in the presence of parametric uncertainties and disturbances.
A Study of Closed Sets and Maps with Ideals
The purpose of this paper is to study a class of closed sets, called generalized pre-closed sets with respect to an ideal (briefly Igp-closed sets), which is an extension of generalized pre-closed sets in general topology. Then, by using these sets, the concepts of Igp- compact spaces along with some classes of maps like continuous and closed maps via ideals have been introduced and analogues of some known results for compact spaces, continuous maps and closed maps in general topology have been obtained.
Machine Learning Invariants to Detect Anomalies in Secure Water Treatment
A strategic model that does not trigger any false alarms to detect anomalies in Secure Water Treatment (SWaT) test bed is presented. This model uses machine learning invariants formulated from streamlining the general form of Auto-Regressive models with eXogenous input. A creative generalized CUSUM algorithm to integrate the invariants and the detection strategy technique is successfully developed and tested in the SWaT Programmable Logic Controllers (PLCs). Three steps to fine-tune parameters, b and τ in the generalized algorithm are stated and an example used to demonstrate the tuning process is discussed. This approach can swiftly and effectively detect various scopes of cyber-attacks such as multiple points single stage and multiple points multiple stages in SWaT. This technique can be applied in water treatment plants and other cyber physical systems like power and gas plants too.
Confidence Intervals for Quantiles in the Two-Parameter Exponential Distributions with Type II Censored Data
Based on type II censored data, we consider interval estimation of the quantiles of the two-parameter exponential distribution and the difference between the quantiles of two independent two-parameter exponential distributions. We derive asymptotic intervals, Bayesian, as well as intervals based on the generalized pivot variable. We also include some bootstrap intervals in our comparisons. The performance of these intervals is investigated in terms of their coverage probabilities and expected lengths.
A Bivariate Inverse Generalized Exponential Distribution and Its Applications in Dependent Competing Risks Model
The aim of this paper is to introduce a bivariate inverse generalized exponential distribution which has a singular component. The proposed bivariate distribution can be used when the marginals have heavy-tailed distributions, and they have non-monotone hazard functions. Due to the presence of the singular component, it can be used quite effectively when there are ties in the data. Since it has four parameters, it is a very flexible bivariate distribution, and it can be used quite effectively for analyzing various bivariate data sets. Several dependency properties and dependency measures have been obtained. The maximum likelihood estimators cannot be obtained in closed form, and it involves solving a four-dimensional optimization problem. To avoid that, we have proposed to use an EM algorithm, and it involves solving only one non-linear equation at each `E'-step. Hence, the implementation of the proposed EM algorithm is very straight forward in practice. Extensive simulation experiments and the analysis of one data set have been performed. We have observed that the proposed bivariate inverse generalized exponential distribution can be used for modeling dependent competing risks data. One data set has been analyzed to show the effectiveness of the proposed model.
Finite Element Method for Solving the Generalized RLW Equation
The General Regularized Long Wave (GRLW) equation is solved numerically by giving a new algorithm based on collocation method using quartic B-splines at the mid-knot points as element shape. Also, we use the Fourth Runge-Kutta method for solving the system of first order ordinary differential equations instead of finite difference method. Our test problems, including the migration and interaction of solitary waves, are used to validate the algorithm which is found to be accurate and efficient. The three invariants of the motion are evaluated to determine the conservation properties of the algorithm.
Nonlinear Static Analysis of Laminated Composite Hollow Beams with Super-Elliptic Cross-Sections
In this paper geometrically nonlinear static behavior of laminated composite hollow super-elliptic beams is investigated using generalized differential quadrature method. Super-elliptic beam can have both oval and elliptic cross-sections by adjusting parameters in super-ellipse formulation (also known as Lamé curves). Equilibrium equations of super-elliptic beam are obtained using the virtual work principle. Geometric nonlinearity is taken into account using von-Kármán nonlinear strain-displacement relations. Spatial derivatives in strains are expressed with the generalized differential quadrature method. Transverse shear effect is considered through the first-order shear deformation theory. Static equilibrium equations are solved using Newton-Raphson method. Several composite super-elliptic beam problems are solved with the proposed method. Effects of layer orientations of composite material, boundary conditions, ovality and ellipticity on bending behavior are investigated.
An Efficient Discrete Chaos in Generalized Logistic Maps with Applications in Image Encryption
In the last few decades, the discrete chaos of difference equations has gained a massive attention of academicians and scholars due to its tremendous applications in each and every branch of science, such as cryptography, traffic control models, secure communications, weather forecasting, and engineering. In this article, a generalized logistic discrete map is established and discrete chaos is reported through period doubling bifurcation, period three orbit and Lyapunov exponent. It is interesting to see that the generalized logistic map exhibits superior chaos due to the presence of an extra degree of freedom of an ordered parameter. The period doubling bifurcation and Lyapunov exponent are demonstrated for some particular values of parameter and the discrete chaos is determined in the sense of Devaney's definition of chaos theoretically as well as numerically. Moreover, the study discusses an extended chaos based image encryption and decryption scheme in cryptography using this novel system. Surprisingly, a larger key space for coding and more sensitive dependence on initial conditions are examined for encryption and decryption of text messages, images and videos which secure the system strongly from external cyber attacks, coding attacks, statistic attacks and differential attacks.
EEG and ABER Abnormalities in Children with Speech and Language Delay
Speech and language delay (SLD) is seen commonly as a co-morbidity in children having severe resistant focal and generalized, syndromic and symptomatic epilepsies. It is however not clear whether epilepsy contributes to or is a mere association in the pathogenesis of SLD. Also, it is acknowledged that Auditory Brainstem Evoked Responses (ABER), besides used for evaluating hearing threshold, also aid in prognostication of neurological disorders and abnormalities in the hearing pathway in the brainstem. There is no circumscribed or surrogate neurophysiologic laboratory marker to adjudge the extent of SLD. The current study was designed to evaluate the abnormalities in Electroencephalography (EEG) and ABER in children with SLD who do not have an overt hearing deficit or autism. 94 children of age group 2-8 years with predominant SLD and without any gross motor developmental delay, head injury, gross hearing disorder, cleft lip/palate and autism were selected. Standard video Electroencephalography using the 10:20 international system and ABER after click stimulus with intensities 110 db until 40 db was performed in all children. EEG was abnormal in 47.9% (n= 45; 36 boys and 9 girls) children. In the children with abnormal EEG, 64.5% (n=29) had an abnormal background, 57.8% (n=27) had presence of generalized interictal epileptiform discharges (IEDs), 20% (n=9) had focal epileptiform discharges exclusively from left side and 33.3% (n=15) had multifocal IEDs occurring both in isolation or associated with generalised abnormalities. In ABER, surprisingly, the peak latencies for waves I, III & V, inter-peak latencies I-III & I-V, III-V and wave amplitude ratio V/I, were found within normal limits in both ears of all the children. Thus in the current study it is certain that presence of generalized IEDs in EEG are seen in higher frequency with SLD and focal IEDs are seen exclusively in left hemisphere in these children. It may be possible that even with generalized EEG abnormalities present in these children, left hemispheric abnormalities as a part of this generalized dysfunction may be responsible for the speech and language dysfunction. The current study also emphasizes that ABER may not be routinely recommended as diagnostic or prognostic tool in children with SLD without frank hearing deficit or autism, thus reducing the burden on electro physiologists, laboratories and saving time and financial resources.
Common Fixed Point Results and Stability of a Modified Jungck Iterative Scheme
In this study, we introduce a modified Jungck (Dual Jungck) iterative scheme and use the scheme to approximate the unique common fixed point of a pair of generalized contractive-like operators in a Banach space. The iterative scheme is also shown to be stable with respect to the maps (S,T). An example is taken to justify the convergence of the scheme. Our result is a generalization and improvement of several results in the literature on single map T.
Comparative Study of Estimators of Population Means in Two Phase Sampling in the Presence of Non-Response
A comparative study of estimators of population means in two phase sampling in the presence of non-response when Unknown population means of the auxiliary variable(s) and incomplete information of study variable y as well as of auxiliary variable(s) is made. Three real data sets of University students, hospital and unemployment are used for comparison of all the available techniques in two phase sampling in the presence of non-response with the newly generalized ratio estimators.
Solution of Some Boundary Value Problems of the Generalized Theory of Thermo-Piezoelectricity
We have considered a non-classical model of dynamical problems for a conjugated system of differential equations arising in thermo-piezoelectricity, which was formulated by Toupin – Mindlin. The basic concepts and the general theory of solvability for isotropic homogeneous elastic media is considered. They are worked by using the methods the Laplace integral transform, potential method and singular integral equations. Approximate solutions of mixed boundary value problems for finite domain, bounded by the some closed surface are constructed. They are solved in explicitly by using the generalized Fourier's series method.
Transverse Vibration of Non-Homogeneous Rectangular Plates of Variable Thickness Using GDQ
The effect of non-homogeneity on the free transverse vibration of thin rectangular plates of bilinearly varying thickness has been analyzed using generalized differential quadrature (GDQ) method. The non-homogeneity of the plate material is assumed to arise due to linear variations in Young’s modulus and density of the plate material with the in-plane coordinates x and y. Numerical results have been computed for fully clamped and fully simply supported boundary conditions. The solution procedure by means of GDQ method has been implemented in a MATLAB code. The effect of various plate parameters has been investigated for the first three modes of vibration. A comparison of results with those available in literature has been presented.
Three-Dimensional Generalized Thermoelasticity with Variable Thermal Conductivity
In this paper, a three-dimensional model of the generalized thermoelasticity with one relaxation time and variable thermal conductivity has been constructed. The resulting non-dimensional governing equations together with the Laplace and double Fourier transforms techniques have been applied to a three-dimensional half-space subjected to thermal loading with rectangular pulse and traction free in the directions of the principle co-ordinates. The inverses of double Fourier transforms, and Laplace transforms have been obtained numerically. Numerical results for the temperature increment, the invariant stress, the invariant strain, and the displacement are represented graphically. The variability of the thermal conductivity has significant effects on the thermal and the mechanical waves.
Kinetic Model to Interpret Whistler Waves in Multicomponent Non-Maxwellian Space Plasmas
Whistler waves are right handed circularly polarized waves and are frequently observed in space plasmas. The Low frequency branch of the Whistler waves having frequencies nearly around 100 Hz, known as Lion roars, are frequently observed in magnetosheath. Another feature of the magnetosheath is the observations of flat top electron distributions with single as well as two electron populations. In the past, lion roars were studied by employing kinetic model using classical bi-Maxwellian distribution function, however, could not be justified both on quantitatively as well as qualitatively grounds. We studied Whistler waves by employing kinetic model using non-Maxwellian distribution function such as the generalized (r,q) distribution function which is the generalized form of kappa and Maxwellian distribution functions by employing kinetic theory with single or two electron populations. We compare our results with the Cluster observations and found good quantitative and qualitative agreement between them. At times when lion roars are observed (not observed) in the data and bi-Maxwellian could not provide the sufficient growth (damping) rates, we showed that when generalized (r,q) distribution function is employed, the resulted growth (damping) rates exactly match the observations.
Effective Charge Coupling in Low Dimensional Doped Quantum Antiferromagnets
The interaction between the charge degrees of freedom for itinerant antiferromagnets is investigated in terms of generalized charge stiffness constant corresponding to nearest neighbour t-J model and t1-t2-t3-J model. The low dimensional hole doped antiferromagnets are the well known systems that can be described by the t-J-like models. Accordingly, we have used these models to investigate the fermionic pairing possibilities and the coupling between the itinerant charge degrees of freedom. A detailed comparison between spin and charge couplings highlights that the charge and spin couplings show very similar behaviour in the over-doped region, whereas, they show completely different trends in the lower doping regimes. Moreover, a qualitative equivalence between generalized charge stiffness and effective Coulomb interaction is also established based on the comparisons with other theoretical and experimental results. Thus it is obvious that the enhanced possibility of fermionic pairing is inherent in the reduction of Coulomb repulsion with increase in doping concentration. However, the increased possibility can not give rise to pairing without the presence of any other pair producing mechanism outside the t-J model. Therefore, one can conclude that the t-J-like models themselves solely are not capable of producing conventional momentum-based superconducting pairing on their own.
Modeling Bessel Beams and Their Discrete Superpositions from the Generalized Lorenz-Mie Theory to Calculate Optical Forces over Spherical Dielectric Particles
In this work, we propose an algorithm developed under Python language for the modeling of ordinary scalar Bessel beams and their discrete superpositions and subsequent calculation of optical forces exerted over dielectric spherical particles. The mathematical formalism, based on the generalized Lorenz-Mie theory, is implemented in Python for its large number of free mathematical (as SciPy and NumPy), data visualization (Matplotlib and PyJamas) and multiprocessing libraries. We also propose an approach, provided by a synchronized Software as Service (SaaS) in cloud computing, to develop a user interface embedded on a mobile application, thus providing users with the necessary means to easily introduce desired unknowns and parameters and see the graphical outcomes of the simulations right at their mobile devices. Initially proposed as a free Android-based application, such an App enables data post-processing in cloud-based architectures and visualization of results, figures and numerical tables.
Econophysics: The Use of Entropy Measures in Finance
Concepts of econophysics are usually used to solve problems related to uncertainty and nonlinear dynamics. In the theory of option pricing the risk neutral probabilities play very important role. The application of entropy in finance can be regarded as the extension of both information entropy and the probability entropy. It can be an important tool in various financial methods such as measure of risk, portfolio selection, option pricing and asset pricing. Gulko applied Entropy Pricing Theory (EPT) for pricing stock options and introduced an alternative framework of Black-Scholes model for pricing European stock option. In this article, we present solutions to maximum entropy problems based on Tsallis, Weighted-Tsallis, Kaniadakis, Weighted-Kaniadakies entropies, to obtain risk-neutral densities. We have also obtained the value of European call and put in this framework.
Functionally Graded of Thermoelastic Materials with Power Law by Adomian's Decomposition Method
This paper introduced an iteration method for the numerical solutions of a one-dimensional problem of generalized thermoelasticity with one relaxation time under given initial and boundary conditions. The thermoelastic material with variable properties as a power functional graded has been considered. Adomian’s decomposition techniques have been applied to the governing equations. The numerical results have been calculated by using the iterations method with a certain algorithm. The numerical results have been represented in figures, and the figures affirm that Adomian’s decomposition method is a successful method for modeling thermoelastic problems. Moreover, the empirical parameter of the functional graded, and the lattice design parameter have significant effects on the temperature increment, the strain, the stress, the displacement.
Generalized Vortex Lattice Method for Predicting Characteristics of Wings with Flap and Aileron Deflection
A generalized vortex lattice method for complex lifting surfaces with flap and aileron deflection is formulated. The method is not restricted by the linearized theory assumption and accounts for all standard geometric lifting surface parameters: camber, taper, sweep, washout, dihedral, in addition to flap and aileron deflection. Thickness is not accounted for since the physical lifting body is replaced by a lattice of panels located on the mean camber surface. This panel lattice setup and the treatment of different wake geometries is what distinguish the present work form the overwhelming majority of previous solutions based on the vortex lattice method. A MATLAB code implementing the proposed formulation is developed and validated by comparing our results to existing experimental and numerical ones and good agreement is demonstrated. It is then used to study the accuracy of the widely used classical vortex-lattice method. It is shown that the classical approach gives good agreement in the clean configuration but is off by as much as 30% when a flap or aileron deflection of 30° is imposed. This discrepancy is mainly due the linearized theory assumption associated with the conventional method. A comparison of the effect of four different wake geometries on the values of aerodynamic coefficients was also carried out and it is found that the choice of the wake shape had very little effect on the results.
Entropy Generation of Unsteady Reactive Hydromagnetic Generalized Couette Fluid Flow of a Two-Step Exothermic Chemical Reaction Through a Channel
In this study, analysis of the entropy generation of an unsteady reactive hydromagnetic generalized couette fluid flow of a two-step exothermic chemical reaction through a channel with isothermal wall temperature under the influence of different chemical kinetics namely: Sensitized, Arrhenius and Bimolecular kinetics was investigated. The modelled nonlinear dimensionless equations governing the fluid flow were simplified and solved using the combined Laplace Differential Transform Method (LDTM). The effects of fluid parameters associated with the problem on the fluid temperature, entropy generation rate and Bejan number were discussed and presented through graphs.
Adomian’s Decomposition Method to Functionally Graded Thermoelastic Materials with Power Law
This paper presents an iteration method for the numerical solutions of a one-dimensional problem of generalized thermoelasticity with one relaxation time under given initial and boundary conditions. The thermoelastic material with variable properties as a power functional graded has been considered. Adomian’s decomposition techniques have been applied to the governing equations. The numerical results have been calculated by using the iterations method with a certain algorithm. The numerical results have been represented in figures, and the figures affirm that Adomian’s decomposition method is a successful method for modeling thermoelastic problems. Moreover, the empirical parameter of the functional graded, and the lattice design parameter have significant effects on the temperature increment, the strain, the stress, the displacement.
Stress and Marital Satisfaction of Parents to Children Diagnosed with Autism
The current investigation expended on research among parents caring for a child who is diagnosed with an autism spectrum disorder (ASD). An online web survey was used to collect data from 253 parents caring for a child with a diagnosis of ASD. Both parents reported on elevated levels of parental stress associated with caring for the child on the spectrum. In addition, lower levels of marital satisfaction were found in both parents. About 13% of the parents in the sample met the diagnostic criteria for Major Depressive Disorder and About 15% of the parents met the diagnostic criteria for Generalized Anxiety Disorder. Although the majority of the sample was females (94%) significant differences were found between males and females in relation to meeting the diagnostic criteria for Major Depressive Disorder and for Generalized Anxiety Disorder. Higher levels of stress were associated with higher number of Generalized Anxiety Disorder symptoms and higher number of Major Depressive Disorder symptoms. Findings from this study indicate how vulnerable parents and especially females are in relation to caring to a child diagnosed with ASD. Educational Objectives: At the conclusion of the paper, the readers should be able to: -Identify levels of stress and marital satisfaction among parents caring for a child diagnosed with autism spectrum disorder, -Recognize the impact of stress on the development of mental health issues, -Name the two most common mood and anxiety related disorders associated with caring for a child diagnosed with an autism spectrum disorder.
On Confidence Intervals for the Difference between Inverse of Normal Means with Known Coefficients of Variation
In this paper, we propose two new confidence intervals for the difference between the inverse of normal means with known coefficients of variation. One of these two confidence intervals for this problem is constructed based on the generalized confidence interval and the other confidence interval is constructed based on the closed form method of variance estimation. We examine the performance of these confidence intervals in terms of coverage probabilities and expected lengths via Monte Carlo simulation.
A Survey on Quasi-Likelihood Estimation Approaches for Longitudinal Set-ups
The Com-Poisson (CMP) model is one of the most popular discrete generalized linear models (GLMS) that handles both equi-, over- and under-dispersed data. In longitudinal context, an integer-valued autoregressive (INAR(1)) process that incorporates covariate specification has been developed to model longitudinal CMP counts. However, the joint likelihood CMP function is difficult to specify and thus restricts the likelihood based estimating methodology. The joint generalized quasilikelihood approach (GQL-I) was instead considered but is rather computationally intensive and may not even estimate the regression effects due to a complex and frequently ill conditioned covariance structure. This paper proposes a new GQL approach for estimating the regression parameters (GQLIII) that are based on a single score vector representation. The performance of GQL-III is compared with GQL-I and separate marginal GQLs (GQL-II) through some simulation experiments and is proved to yield equally efficient estimates as GQL-I and is far more computationally stable.
Generalized Model Estimating Strength of Bauxite Residue-Lime Mix
The present work investigates the effect of multiple parameters on the unconfined compressive strength of the bauxite residue-lime mix. A number of unconfined compressive strength tests considering various curing time, lime content, dry density and moisture content were carried out. The results show that an empirical correlation may be successfully developed using volumetric lime content, porosity, moisture content, curing time unconfined compressive strength for the range of the bauxite residue-lime mix studied. The proposed empirical correlations efficiently predict the strength of bauxite residue-lime mix, and it can be used as a generalized empirical equation to estimate unconfined compressive strength.
Quantum Modelling of AgHMoO4, CsHMoO4 and AgCsMoO4 Chemistry in the Field of Nuclear Power Plant Safety
In a major nuclear accident, the released fission products (FPs) and the structural materials are likely to influence the transport of iodine in the reactor coolant system (RCS) of a pressurized water reactor (PWR). So far, the thermodynamic data on cesium and silver species used to estimate the magnitude of FP release show some discrepancies, data are scarce and not reliable. For this reason, it is crucial to review the thermodynamic values related to cesium and silver materials. To this end, we have used state-of-the-art quantum chemical methods to compute the formation enthalpies and entropies of AgHMoO₄, CsHMoO₄, and AgCsMoO₄ in the gas phase. Different quantum chemical methods have been investigated (DFT and CCSD(T)) in order to predict the geometrical parameters and the energetics including the correlation energy. The geometries were optimized with TPSSh-5%HF method, followed by a single point calculation of the total electronic energies using the CCSD(T) wave function method. We thus propose with a final uncertainty of about 2 kJmol⁻¹ standard enthalpies of formation of AgHMoO₄, CsHMoO₄, and AgCsMoO₄.
Bivariate Generalization of q-α-Bernstein Polynomials
We propose to define the q-analogue of the α-Bernstein Kantorovich operators and then introduce the q-bivariate generalization of these operators to study the approximation of functions of two variables. We obtain the rate of convergence of these bivariate operators by means of the total modulus of continuity, partial modulus of continuity and the Peetre’s K-functional for continuous functions. Further, in order to study the approximation of functions of two variables in a space bigger than the space of continuous functions, i.e. Bögel space; the GBS (Generalized Boolean Sum) of the q-bivariate operators is considered and degree of approximation is discussed for the Bögel continuous and Bögel differentiable functions with the aid of the Lipschitz class and the mixed modulus of smoothness.
A Problem on Homogeneous Isotropic Microstretch Thermoelastic Half Space with Mass Diffusion Medium under Different Theories
The present investigation deals with generalized model of the equations for a homogeneous isotropic microstretch thermoelastic half space with mass diffusion medium. Theories of generalized thermoelasticity Lord-Shulman (LS) Green-Lindsay (GL) and Coupled Theory (CT) theories are applied to investigate the problem. The stresses in the considered medium have been studied due to normal force and tangential force. The normal mode analysis technique is used to calculate the normal stress, shear stress, couple stresses and microstress. A numerical computation has been performed on the resulting quantity. The computed numerical results are shown graphically.
Estimation of Rare and Clustered Population Mean Using Two Auxiliary Variables in Adaptive Cluster Sampling
Adaptive cluster sampling (ACS) is specifically developed for the estimation of highly clumped populations and applied to a wide range of situations like animals of rare and endangered species, uneven minerals, HIV patients and drug users. In this paper, we proposed a generalized semi-exponential estimator with two auxiliary variables under the framework of ACS design. The expressions of approximate bias and mean square error (MSE) of the proposed estimator are derived. Theoretical comparisons of the proposed estimator have been made with existing estimators. A numerical study is conducted on real and artificial populations to demonstrate and compare the efficiencies of the proposed estimator. The results indicate that the proposed generalized semi-exponential estimator performed considerably better than all the adaptive and non-adaptive estimators considered in this paper.
Stochastic Matrices and Lp Norms for Ill-Conditioned Linear Systems
In quite diverse application areas such as astronomy, medical imaging, geophysics or nondestructive evaluation, many problems related to calibration, fitting or estimation of a large number of input parameters of a model from a small amount of output noisy data, can be cast as inverse problems. Due to noisy data corruption, insufficient data and model errors, most inverse problems are ill-posed in a Hadamard sense, i.e. existence, uniqueness and stability of the solution are not guaranteed. A wide class of inverse problems in physics relates to the Fredholm equation of the first kind. The ill-posedness of such inverse problem results, after discretization, in a very ill-conditioned linear system of equations, the condition number of the associated matrix can typically range from 109 to 1018. This condition number plays the role of an amplifier of uncertainties on data during inversion and then, renders the inverse problem difficult to handle numerically. Similar problems appear in other areas such as numerical optimization when using interior points algorithms for solving linear programs leads to face ill-conditioned systems of linear equations. Devising efficient solution approaches for such system of equations is therefore of great practical interest. Efficient iterative algorithms are proposed for solving a system of linear equations. The approach is based on a preconditioning of the initial matrix of the system with an approximation of a generalized inverse leading to a stochastic preconditioned matrix. This approach, valid for non-negative matrices, is first extended to hermitian, semi-definite positive matrices and then generalized to any complex rectangular matrices. The main results obtained are as follows: 1) We are able to build a generalized inverse of any complex rectangular matrix which satisfies the convergence condition requested in iterative algorithms for solving a system of linear equations. This completes the (short) list of generalized inverse having this property, after Kaczmarz and Cimmino matrices. Theoretical results on both the characterization of the type of generalized inverse obtained and the convergence are derived. 2) Thanks to its properties, this matrix can be efficiently used in different solving schemes as Richardson-Tanabe or preconditioned conjugate gradients. 3) By using Lp norms, we propose generalized Kaczmarz’s type matrices. We also show how Cimmino's matrix can be considered as a particular case consisting in choosing the Euclidian norm in an asymmetrical structure. 4) Regarding numerical results obtained on some pathological well-known test-cases (Hilbert, Nakasaka, …), some of the proposed algorithms are empirically shown to be more efficient on ill-conditioned problems and more robust to error propagation than the known classical techniques we have tested (Gauss, Moore-Penrose inverse, minimum residue, conjugate gradients, Kaczmarz, Cimmino). We end on a very early prospective application of our approach based on stochastic matrices aiming at computing some parameters (such as the extreme values, the mean, the variance, …) of the solution of a linear system prior to its resolution. Such an approach, if it were to be efficient, would be a source of information on the solution of a system of linear equations.
Evaluating Traffic Congestion Using the Bayesian Dirichlet Process Mixture of Generalized Linear Models
This study applied traffic speed and occupancy to develop clustering models that identify different traffic conditions. Particularly, these models are based on the Dirichlet Process Mixture of Generalized Linear regression (DML) and change-point regression (CR). The model frameworks were implemented using 2015 historical traffic data aggregated at a 15-minute interval from an Interstate 295 freeway in Jacksonville, Florida. Using the deviance information criterion (DIC) to identify the appropriate number of mixture components, three traffic states were identified as free-flow, transitional, and congested condition. Results of the DML revealed that traffic occupancy is statistically significant in influencing the reduction of traffic speed in each of the identified states. Influence on the free-flow and the congested state was estimated to be higher than the transitional flow condition in both evening and morning peak periods. Estimation of the critical speed threshold using CR revealed that 47 mph and 48 mph are speed thresholds for congested and transitional traffic condition during the morning peak hours and evening peak hours, respectively. Free-flow speed thresholds for morning and evening peak hours were estimated at 64 mph and 66 mph, respectively. The proposed approaches will facilitate accurate detection and prediction of traffic congestion for developing effective countermeasures.
A Boundary Backstepping Control Design for 2-D, 3-D and N-D Heat Equation
We consider the problem of stabilization of an unstable heat equation in a 2-D, 3-D and generally n-D domain by deriving a generalized backstepping boundary control design methodology. To stabilize the systems, we design boundary backstepping controllers inspired by the 1-D unstable heat equation stabilization procedure. We assume that one side of the boundary is hinged and the other side is controlled for each direction of the domain. Thus, controllers act on two boundaries for 2-D domain, three boundaries for 3-D domain and ”n” boundaries for n-D domain. The main idea of the design is to derive ”n” controllers for each of the dimensions by using ”n” kernel functions. Thus, we obtain ”n” controllers for the ”n” dimensional case. We use a transformation to change the system into an exponentially stable ”n” dimensional heat equation. The transformation used in this paper is a generalized Volterra/Fredholm type with ”n” kernel functions for n-D domain instead of the one kernel function of 1-D design.
Thermodynamic Approach of Lanthanide-Iron Double Oxides Formation
Standard Gibbs energy of formation ΔGfor(298.15) of lanthanide-iron double oxides of garnet-type crystal structure R3Fe5O12 - RIG (R – are rare earth ions) from initial oxides are evaluated. The calculation is based on the data of standard entropies S298.15 and standard enthalpies ΔH298.15 of formation of compounds which are involved in the process of garnets synthesis. Gibbs energy of formation is presented as temperature function ΔGfor(T) for the range 300-1600K. The necessary starting thermodynamic data were obtained from calorimetric study of heat capacity – temperature functions and by using the semi-empirical method for calculation of ΔH298.15 of formation. Thermodynamic functions for standard temperature – enthalpy, entropy and Gibbs energy - are recommended as reference data for technological evaluations. Through the isostructural series of rare earth-iron garnets the correlation between thermodynamic properties and characteristics of lanthanide ions are elucidated.
Standard Gibbs Energy of Formation and Entropy of Lanthanide-Iron Oxides of Garnet Crystal Structure
Standard Gibbs energy of formation ΔGfor(298.15) of lanthanide-iron double oxides of garnet-type crystal structure R3Fe5O12 - RIG (R – are rare earth ions) from initial oxides are evaluated. The calculation is based on the data of standard entropies S298.15 and standard enthalpies ΔH298.15 of formation of compounds which are involved in the process of garnets synthesis. Gibbs energy of formation is presented as temperature function ΔGfor(T) for the range 300-1600K. The necessary starting thermodynamic data were obtained from calorimetric study of heat capacity and by using the semi-empirical method for calculation of ΔH298.15 (formation). Thermodynamic functions for standard temperature – enthalpy, entropy and Gibbs energy - are recommended as reference data for technological evaluations. Through the isostructural series of rare earth-iron garnets the correlation between thermodynamic properties and characteristics of lanthanide ions are elucidated.
On the Performance of Improvised Generalized M-Estimator in the Presence of High Leverage Collinearity Enhancing Observations
Multicollinearity occurs when two or more independent variables in a multiple linear regression model are highly correlated. The ridge regression is the commonly used method to rectify this problem. However, the ridge regression cannot handle the problem of multicollinearity which is caused by high leverage collinearity enhancing observation (HLCEO). Since high leverage points (HLPs) are responsible for inducing multicollinearity, the effect of HLPs needs to be reduced by using Generalized M estimator. The existing GM6 estimator is based on the Minimum Volume Ellipsoid (MVE) which tends to swamp some low leverage points. Hence an improvised GM (MGM) estimator is presented to improve the precision of the GM6 estimator. Numerical example and simulation study are presented to show how HLPs can cause multicollinearity. The numerical results show that our MGM estimator is the most efficient method compared to some existing methods.
A Generalized Family of Estimators for Estimation of Unknown Population Variance in Simple Random Sampling
This paper is addressing the estimation method of the unknown population variance of the variable of interest. A new generalized class of estimators of the finite population variance has been suggested using the auxiliary information. To improve the precision of the proposed class, known population variance of the auxiliary variable has been used. Mathematical expressions for the biases and the asymptotic variances of the suggested class are derived under large sample approximation. Theoretical and numerical comparisons are made to investigate the performances of the proposed class of estimators. The empirical study reveals that the suggested class of estimators performs better than the usual estimator, classical ratio estimator, classical product estimator and classical linear regression estimator. It has also been found that the suggested class of estimators is also more efficient than some recently published estimators.
Generalized Synchronization in Systems with a Complex Topology of Attractor
Generalized synchronization is one of the most intricate phenomena in nonlinear science. It can be observed both in systems with a unidirectional and mutual type of coupling including the complex networks. Such a phenomenon has a number of practical applications, for example, for the secure information transmission through the communication channel with a high level of noise. Known methods for the secure information transmission needs in the increase of the privacy of data transmission that arises a question about the observation of such phenomenon in systems with a complex topology of chaotic attractor possessing two or more positive Lyapunov exponents. The present report is devoted to the study of such phenomenon in two unidirectionally and mutually coupled dynamical systems being in chaotic (with one positive Lyapunov exponent) and hyperchaotic (with two or more positive Lyapunov exponents) regimes, respectively. As the systems under study, we have used two mutually coupled modified Lorenz oscillators and two unidirectionally coupled time-delayed generators. We have shown that in both cases the generalized synchronization regime can be detected by means of the calculation of Lyapunov exponents and phase tube approach whereas due to the complex topology of attractor the nearest neighbor method is misleading. Moreover, the auxiliary system approaches being the standard method for the synchronous regime observation, for the mutual type of coupling results in incorrect results. To calculate the Lyapunov exponents in time-delayed systems we have proposed an approach based on the modification of Gram-Schmidt orthogonalization procedure in the context of the time-delayed system. We have studied in detail the mechanisms resulting in the generalized synchronization regime onset paying a great attention to the field where one positive Lyapunov exponent has already been become negative whereas the second one is a positive yet. We have found the intermittency here and studied its characteristics. To detect the laminar phase lengths the method based on a calculation of local Lyapunov exponents has been proposed. The efficiency of the method has been verified using the example of two unidirectionally coupled Rössler systems being in the band chaos regime. We have revealed the main characteristics of intermittency, i.e. the distribution of the laminar phase lengths and dependence of the mean length of the laminar phases on the criticality parameter, for all systems studied in the report. This work has been supported by the Russian President's Council grant for the state support of young Russian scientists (project MK-531.2018.2).
On Generalized Cumulative Past Inaccuracy Measure for Marginal and Conditional Lifetimes
Recently, the notion of past cumulative inaccuracy (CPI) measure has been proposed in the literature as a generalization of cumulative past entropy (CPE) in univariate as well as bivariate setup. In this paper, we introduce the notion of CPI of order α (alpha) and study the proposed measure for conditionally specified models of two components failed at different time instants called generalized conditional CPI (GCCPI). We provide some bounds using usual stochastic order and investigate several properties of GCCPI. The effect of monotone transformation on this proposed measure has also been examined. Furthermore, we characterize some bivariate distributions under the assumption of conditional proportional reversed hazard rate model. Moreover, the role of GCCPI in reliability modeling has also been investigated for a real-life problem.
Generalized Rough Sets Applied to Graphs Related to Urban Problems
Branch of modern mathematics, graphs represent instruments for optimization and solving practical applications in various fields such as economic networks, engineering, network optimization, the geometry of social action, generally, complex systems including contemporary urban problems (path or transport efficiencies, biourbanism, & c.). In this paper is studied the interconnection of some urban network, which can lead to a simulation problem of a digraph through another digraph. The simulation is made univoc or more general multivoc. The concepts of fragment and atom are very useful in the study of connectivity in the digraph that is simulation - including an alternative evaluation of k- connectivity. Rough set approach in (bi)digraph which is proposed in premier in this paper contribute to improved significantly the evaluation of k-connectivity. This rough set approach is based on generalized rough sets - basic facts are presented in this paper.
Statistical Modeling of Mobile Fading Channels Based on Triply Stochastic Filtered Marked Poisson Point Processes
Understanding the statistics of non-isotropic scattering multipath channels that fade randomly with respect to time, frequency, and space in a mobile environment is very crucial for the accurate detection of received signals in wireless and cellular communication systems. In this paper, we derive stochastic models for the probability density function (PDF) of the shift in the carrier frequency caused by the Doppler Effect on the received illuminating signal in the presence of a dominant line of sight. Our derivation is based on a generalized Clarke’s and a two-wave partially developed scattering models, where the statistical distribution of the frequency shift is shown to be consistent with the power spectral density of the Doppler shifted signal.
A Guide for Using Viscoelasticity in ANSYS
Theory of viscoelasticity is used by many researchers to represent the behavior of many materials such as pavements on roads or bridges. Several researches used analytical methods and rheology to predict the material behaviors of simple models. Today, more complex engineering structures are analyzed using Finite Element Method, in which material behavior is embedded by means of three dimensional viscoelastic material laws. As a result, structures of unordinary geometry and domain can be analyzed by means of Finite Element Method and three dimensional viscoelastic equations. In the scope of this study, rheological models embedded in ANSYS, namely, generalized Maxwell model and Prony series, which are two methods used by ANSYS to represent viscoelastic material behavior, are presented explicitly. Afterwards, a guide is illustrated to ease using of viscoelasticity tool in ANSYS.
Generalized Additive Model Approach for the Chilean Hake Population in a Bio-Economic Context
The traditional bio-economic method for fisheries modeling uses some estimate of the growth parameters and the system carrying capacity from a biological model for the population dynamics (usually a logistic population growth model) which is then analyzed as a traditional production function. The stock dynamic is transformed into a revenue function and then compared with the extraction costs to estimate the maximum economic yield. In this paper, the logistic population growth model for the population is combined with a forecast of the abundance and location of the stock by using a generalized additive model approach. The paper focuses on the Chilean hake population. This method allows for the incorporation of climatic variables and the interaction with other marine species, which in turn will increase the reliability of the estimates and generate better extraction paths for different conservation objectives, such as the maximum biological yield or the maximum economic yield.
Robust Pattern Recognition via Correntropy Generalized Orthogonal Matching Pursuit
This paper presents a novel sparse representation method for robust pattern classification. Generalized orthogonal matching pursuit (GOMP) is a recently proposed efficient sparse representation technique. However, GOMP adopts the mean square error (MSE) criterion and assign the same weights to all measurements, including both severely and slightly corrupted ones. To reduce the limitation, we propose an information-theoretic GOMP (ITGOMP) method by exploiting the correntropy induced metric. The results show that ITGOMP can adaptively assign small weights on severely contaminated measurements and large weights on clean ones, respectively. An ITGOMP based classifier is further developed for robust pattern classification. The experiments on public real datasets demonstrate the efficacy of the proposed approach.
Analysis of Factors Affecting the Number of Infant and Maternal Mortality in East Java with Geographically Weighted Bivariate Generalized Poisson Regression Method
Poisson regression is a non-linear regression model with response variable in the form of count data that follows Poisson distribution. Modeling for a pair of count data that show high correlation can be analyzed by Poisson Bivariate Regression. Data, the number of infant mortality and maternal mortality, are count data that can be analyzed by Poisson Bivariate Regression. The Poisson regression assumption is an equidispersion where the mean and variance values are equal. However, the actual count data has a variance value which can be greater or less than the mean value (overdispersion and underdispersion). Violations of this assumption can be overcome by applying Generalized Poisson Regression. Characteristics of each regency can affect the number of cases occurred. This issue can be overcome by spatial analysis called geographically weighted regression. This study analyzes the number of infant mortality and maternal mortality based on conditions in East Java in 2016 using Geographically Weighted Bivariate Generalized Poisson Regression (GWBGPR) method. Modeling is done with adaptive bisquare Kernel weighting which produces 3 regency groups based on infant mortality rate and 5 regency groups based on maternal mortality rate. Variables that significantly influence the number of infant and maternal mortality are the percentages of pregnant women visit health workers at least 4 times during pregnancy, pregnant women get Fe3 tablets, obstetric complication handled, clean household and healthy behavior, and married women with the first marriage age under 18 years.
Application of Generalized Autoregressive Score Model to Stock Returns
The current study investigates the behaviour of time-varying parameters that are based on the score function of the predictive model density at time t. The mechanism to update the parameters over time is the scaled score of the likelihood function. The results revealed that there is high persistence of time-varying, as the location parameter is higher and the skewness parameter implied the departure of scale parameter from the normality with the unconditional parameter as 1.5. The results also revealed that there is a perseverance of the leptokurtic behaviour in stock returns which implies the returns are heavily tailed. Prior to model estimation, the White Neural Network test exposed that the stock price can be modelled by a GAS model. Finally, we proposed further researches specifically to model the existence of time-varying parameters with a more detailed model that encounters the heavy tail distribution of the series and computes the risk measure associated with the returns.
Generalized Hyperbolic Functions: Exponential-Type Quantum Interactions
In the search of potential models applied in the theoretical treatment of diatomic molecules, some of them have been constructed by using standard hyperbolic functions as well as from the so-called q-deformed hyperbolic functions (sc q-dhf) for displacing and modifying the shape of the potential under study. In order to transcend the scope of hyperbolic functions, in this work, a kind of generalized q-deformed hyperbolic functions (g q-dhf) is presented. By a suitable transformation, through the q deformation parameter, it is shown that these g q-dhf can be expressed in terms of their corresponding standard ones besides they can be reduced to the sc q-dhf. As a useful application of the proposed approach, and considering a class of exactly solvable multi-parameter exponential-type potentials, some new q-deformed quantum interactions models that can be used as interesting alternative in quantum physics and quantum states are presented. Furthermore, due that quantum potential models are conditioned on the q-dependence of the parameters that characterize to the exponential-type potentials, it is shown that many specific cases of q-deformed potentials are obtained as particular cases from the proposal.
Chemometric Estimation of Phytochemicals Affecting the Antioxidant Potential of Lettuce
In this paper, the influence of six different phytochemical content (phenols, carotenoids, chlorophyll a, chlorophyll b, chlorophyll a + b and vitamin C) on antioxidant potential of Murai and Levistro lettuce varieties was evaluated. Variable selection was made by generalized pair correlation method (GPCM) as a novel ranking method. This method is used for the discrimination between two variables that almost equal correlate to a dependent variable. Fisher’s conditional exact and McNemar’s test were carried out. Established multiple linear (MLR) models were statistically evaluated. As the best phytochemicals for the antioxidant potential prediction, chlorophyll a, chlorophyll a + b and total carotenoids content stand out. This was confirmed through both GPCM and MLR, predictive ability of obtained MLR can be used for antioxidant potential estimation for similar lettuce samples. This article is based upon work from the project of the Provincial Secretariat for Science and Technological Development of Vojvodina (No. 114-451-347/2015-02).
Forecast of the Small Wind Turbines Sales with Replacement Purchases and with or without Account of Price Changes
The purpose of the paper is to estimate the US small wind turbines market potential and forecast the small wind turbines sales in the US. The forecasting method is based on the application of the Bass model and the generalized Bass model of innovations diffusion under replacement purchases. In the work an exponential distribution is used for modeling of replacement purchases. Only one parameter of such distribution is determined by average lifetime of small wind turbines. The identification of the model parameters is based on nonlinear regression analysis on the basis of the annual sales statistics which has been published by the American Wind Energy Association (AWEA) since 2001 up to 2012. The estimation of the US average market potential of small wind turbines (for adoption purchases) without account of price changes is 57080 (confidence interval from 49294 to 64866 at P = 0.95) under average lifetime of wind turbines 15 years, and 62402 (confidence interval from 54154 to 70648 at P = 0.95) under average lifetime of wind turbines 20 years. In the first case the explained variance is 90,7%, while in the second - 91,8%. The effect of the wind turbines price changes on their sales was estimated using generalized Bass model. This required a price forecast. To do this, the polynomial regression function, which is based on the Berkeley Lab statistics, was used. The estimation of the US average market potential of small wind turbines (for adoption purchases) in that case is 42542 (confidence interval from 32863 to 52221 at P = 0.95) under average lifetime of wind turbines 15 years, and 47426 (confidence interval from 36092 to 58760 at P = 0.95) under average lifetime of wind turbines 20 years. In the first case the explained variance is 95,3%, while in the second –95,3%.
A Generalized Sparse Bayesian Learning Algorithm for Near-Field Synthetic Aperture Radar Imaging: By Exploiting Impropriety and Noncircularity
The near-field synthetic aperture radar (SAR) imaging is an advanced nondestructive testing and evaluation (NDT&E) technique. This paper investigates the complex-valued signal processing related to the near-field SAR imaging system, where the measurement data turns out to be noncircular and improper, meaning that the complex-valued data is correlated to its complex conjugate. Furthermore, we discover that the degree of impropriety of the measurement data and that of the target image can be highly correlated in near-field SAR imaging. Based on these observations, A modified generalized sparse Bayesian learning algorithm is proposed, taking impropriety and noncircularity into account. Numerical results show that the proposed algorithm provides performance gain, with the help of noncircular assumption on the signals.
Generalized Additive Model for Estimating Propensity Score
Propensity Score Matching (PSM) technique has been widely used for estimating causal effect of treatment in observational studies. One major step of implementing PSM is estimating the propensity score (PS). Logistic regression model with additive linear terms of covariates is most used technique in many studies. Logistics regression model is also used with cubic splines for retaining flexibility in the model. However, choosing the functional form of the logistic regression model has been a question since the effectiveness of PSM depends on how accurately the PS been estimated. In many situations, the linearity assumption of linear logistic regression may not hold and non-linear relation between the logit and the covariates may be appropriate. One can estimate PS using machine learning techniques such as random forest, neural network etc for more accuracy in non-linear situation. In this study, an attempt has been made to compare the efficacy of Generalized Additive Model (GAM) in various linear and non-linear settings and compare its performance with usual logistic regression. GAM is a non-parametric technique where functional form of the covariates can be unspecified and a flexible regression model can be fitted. In this study various simple and complex models have been considered for treatment under several situations (small/large sample, low/high number of treatment units) and examined which method leads to more covariate balance in the matched dataset. It is found that logistic regression model is impressively robust against inclusion quadratic and interaction terms and reduces mean difference in treatment and control set equally efficiently as GAM does. GAM provided no significantly better covariate balance than logistic regression in both simple and complex models. The analysis also suggests that larger proportion of controls than treatment units leads to better balance for both of the methods.
Learning Algorithms for Fuzzy Inference Systems Composed of Double- and Single-Input Rule Modules
Most of self-tuning fuzzy systems, which are automatically constructed from learning data, are based on the steepest descent method (SDM). However, this approach often requires a large convergence time and gets stuck into a shallow local minimum. One of its solutions is to use fuzzy rule modules with a small number of inputs such as DIRMs (Double-Input Rule Modules) and SIRMs (Single-Input Rule Modules). In this paper, we consider a (generalized) DIRMs model composed of double and single-input rule modules. Further, in order to reduce the redundant modules for the (generalized) DIRMs model, pruning and generative learning algorithms for the model are suggested. In order to show the effectiveness of them, numerical simulations for function approximation, Box-Jenkins and obstacle avoidance problems are performed.
Artificial Neural Network Modeling of a Closed Loop Pulsating Heat Pipe
Technological innovations in electronic world demand novel, compact, simple in design, less costly and effective heat transfer devices. Closed Loop Pulsating Heat Pipe (CLPHP) is a passive phase change heat transfer device and has potential to transfer heat quickly and efficiently from source to sink. Thermal performance of a CLPHP is governed by various parameters such as number of U-turns, orientations, input heat, working fluids and filling ratio. The present paper is an attempt to predict the thermal performance of a CLPHP using Artificial Neural Network (ANN). Filling ratio and heat input are considered as input parameters while thermal resistance is set as target parameter. Types of neural networks considered in the present paper are radial basis, generalized regression, linear layer, cascade forward back propagation, feed forward back propagation; feed forward distributed time delay, layer recurrent and Elman back propagation. Linear, logistic sigmoid, tangent sigmoid and Radial Basis Gaussian Function are used as transfer functions. Prediction accuracy is measured based on the experimental data reported by the researchers in open literature as a function of Mean Absolute Relative Deviation (MARD). The prediction of a generalized regression ANN model with spread constant of 4.8 is found in agreement with the experimental data for MARD in the range of ±1.81%.
Relationship between Matrilin-3 (MATN-3) Gene Single Nucleotide Six Polymorphism, Transforming Growth Factor Beta 2 and Radiographic Grading in Primary Osteoarthritis
Objective: Assess serum level of Transforming growth factor beta 2 (TGF-β2) and Matrilin-3 (MATN3) SNP6 polymorphism in osteoarthritic patients Background: Osteoarthritis (OA) is a musculoskeletal disease characterized by pain and joint stiffness. TGF-β 2 is involved in chondrogenesis and osteogenesis, It has found that MATN3 gene and protein expression was correlated with the extent of tissue damage in OA. Findings suggest that regulation of MATN3 expression is essential for maintenance of the cartilage extracellular matrix microenvironment Subjects and Methods: 72 cases of primary OA (56 with knee OA and 16 with generalized OA were compared with that of 18 healthy controls. Radiographs were scored with the Kellgren-Lawrence scale. Serum TGF-β2 was measured by using (ELISA), levels of marker were correlated to radiographic grading of disease and MATN3 SNP6 polymorphism was determined by (PCR-RFLP). Results: MATN3 SNP6 polymorphism and serum level of TGF-β2 were higher in OA compared with controls. Genotype, NN and N allele frequency were higher in patients with OA compared with controls. NN genotype and N allele frequency were higher in knee osteoarthritis than generalized OA. Significant positive correlation between level of TGFβ2 and radiographic grading in group with knee OA, but no correlation between serum level of TGFβ2 and radiographic grading in generalized OA. Conclusion: MATN3 SNP6 polymorphism and TGF-β2 implicated in the pathogenesis of osteoarthritis. Association of N/N genotype with primary osteoarthritis emphasizes on the need for prospective study include larger sample size to confirm the results of the present study.
An Equivalence between a Harmonic Form and a Closed Co-Closed Differential Form in L^Q and Non-L^Q Spaces
An equivalent relation between a harmonic form and a closed co-closed form is established on a complete non-compact manifold. This equivalence has been generalized for a differential k-form ω from Lq spaces to non-Lq spaces when q=2 in the context of p-balanced growth where p=2. Especially for a simple differential k-form on a complete non-compact manifold, the equivalent relation has been verified with the extended scope of q for from finite q-energy in Lq spaces to infinite q-energy in non-Lq spaces when with 2-balanced growth. Generalized Hadamard Theorem, Cauchy-Schwarz Inequality, and Calculus skills including Integration by Parts as well as Convergent Series have been applied as estimation techniques to evaluate growth rates for a differential form. In particular, energy growth rates as indicated by an appropriate power range in a selected test function lead to a balance between a harmonic differential form and a closed co-closed differential form. Research ideas and computational methods in this paper could provide an innovative way in the study of broadening Lq spaces to non-Lq spaces with a wide variety of infinite energy growth for a differential form.
The Role of Surgery to Remove the Primary Tumor in Patients with Metastatic Breast Cancer
Purpose. To evaluate the expediency and timeliness of performance of surgical treatment as a component of multi-therapy treatment of patients with stage IV breast cancers. Materials and Methods. This investigation comparatively analyzed the results of complex treatment with or without surgery in patients with metastatic breast cancer. We analyzed retrospectively treatment experience of 196 patients with generalized breast cancer in the department of oncology and breast reconstructive surgery of P.A. Herzen Moscow Cancer Research Institute from 2000 to 2012. The average age was (58±1,1) years. Invasive ductul carcinoma was verified in128 patients (65,3%), invasive lobular carcinoma-33 (16,8%), complex form - 19 (9,7%). Complex palliative care involving drug and radiation therapies was performed in two patient groups. The first group includes 124 patients who underwent surgical intervention as complex treatment, the second group includes 72 patients with only medical therapy. Standard systemic therapy was given to all patients. Results. Overall, 3-and 5-year survival in fist group was 43,8 and 21%, in second - 15,1 and 9,3% respectively [p=0,00002 log-rank]. Median survival in patients with surgical treatment composed 32 months, in patients with only systemic therapy-21. The factors having influencing an influence on the prognosis and the quality of life outcomes for of patients with generalized breast cancer were are also studied: hormone-dependent tumor, Her2/neu hyper-expression, reproductive function status (age, menopause existence). Conclusion.Removing primary breast tumor in patients with generalized breast cancer improve long-term outcomes. Three- and five-year survival increased by 28,7 and 16,3% respectively, and median survival–for 11 months. These patients may benefit from resection of the breast tumor. One explanation for the effect of this resection is that reducing the tumor load influences metastatic growth.
Multimodal Convolutional Neural Network for Musical Instrument Recognition
The dynamic behavior of music and video makes it difficult to evaluate musical instrument playing in a video by computer system. Any television or film video clip with music information are rich sources for analyzing musical instruments using modern machine learning technologies. In this research, we integrate the audio and video information sources using convolutional neural network (CNN) and pass network learned features through recurrent neural network (RNN) to preserve the dynamic behaviors of audio and video. We use different pre-trained CNN for music and video feature extraction and then fine tune each model. The music network use 2D convolutional network and video network use 3D convolution (C3D). Finally, we concatenate each music and video feature by preserving the time varying features. The long short term memory (LSTM) network is used for long-term dynamic feature characterization and then use late fusion with generalized mean. The proposed network performs better performance to recognize the musical instrument using audio-video multimodal neural network.
A Combined Error Control with Forward Euler Method for Dynamical Systems
Variable time-stepping algorithms for solving dynamical systems performed poorly for long time computations which pass close to a fixed point. To overcome this difficulty, several authors considered phase space error controls for numerical simulation of dynamical systems. In one generalized phase space error control, a step-size selection scheme was proposed, which allows this error control to be incorporated into the standard adaptive algorithm as an extra constraint at negligible extra computational cost. For this generalized error control, it was already analyzed the forward Euler method applied to the linear system whose coefficient matrix has real and negative eigenvalues. In this paper, this result was extended to the linear system whose coefficient matrix has complex eigenvalues with negative real parts. Some theoretical results were obtained and numerical experiments were carried out to support the theoretical results.
Generalized Approach to Linear Data Transformation
This paper presents a generalized approach for the simple linear data transformation, Y=bX, through an integration of multidimensional coordinate geometry, vector space theory and polygonal geometry. The scaling is performed by adding an additional ’Dummy Dimension’ to the n-dimensional data, which helps plot two dimensional component-wise straight lines on pairs of dimensions. The end result is a set of scaled extensions of observations in any of the 2n spatial divisions, where n is the total number of applicable dimensions/dataset variables, created by shifting the n-dimensional plane along the ’Dummy Axis’. The derived scaling factor was found to be dependent on the coordinates of the common point of origin for diverging straight lines and the plane of extension, chosen on and perpendicular to the ’Dummy Axis’, respectively. This result indicates the geometrical interpretation of a linear data transformation and hence, opportunities for a more informed choice of the factor ’b’, based on a better choice of these coordinate values. The paper follows on to identify the effect of this transformation on certain popular distance metrics, wherein for many, the distance metric retained the same scaling factor as that of the features.
Symbolic Partial Differential Equations Analysis Using Mathematica
Many symbolic computations and manipulations required in the analysis of partial differential equations (PDE) or systems of PDEs are tedious and error-prone. These computations arise when determining conservation laws, entropies or integral identities, which are essential tools for the study of PDEs. Here, we discuss a new Mathematica package for the symbolic analysis of PDEs that automate multiple tasks, saving time and effort. Methodologies: During the research, we have used concepts of linear algebra and partial differential equations. We have been working on creating algorithms based on theoretical mathematics to find results mentioned below. Major Findings: Our package provides the following functionalities; finding symmetry group of different PDE systems, generation of polynomials invariant with respect to different symmetry groups; simplification of integral quantities by integration by parts and null Lagrangian cleaning, computing general forms of expressions by integration by parts; finding equivalent forms of an integral expression that are simpler or more symmetric form; determining necessary and sufficient conditions on the coefficients for the positivity of a given symbolic expression. Conclusion: Using this package, we can simplify integral identities, find conserved and dissipated quantities of time-dependent PDE or system of PDEs. Some examples in the theory of mean-field games and semiconductor equations are discussed.
Adomian’s Decomposition Method to Generalized Magneto-Thermoelasticity
Due to many applications and problems in the fields of plasma physics, geophysics, and other many topics, the interaction between the strain field and the magnetic field has to be considered. Adomian introduced the decomposition method for solving linear and nonlinear functional equations. This method leads to accurate, computable, approximately convergent solutions of linear and nonlinear partial and ordinary differential equations even the equations with variable coefficients. This paper is dealing with a mathematical model of generalized thermoelasticity of a half-space conducting medium. A magnetic field with constant intensity acts normal to the bounding plane has been assumed. Adomian’s decomposition method has been used to solve the model when the bounding plane is taken to be traction free and thermally loaded by harmonic heating. The numerical results for the temperature increment, the stress, the strain, the displacement, the induced magnetic, and the electric fields have been represented in figures. The magnetic field, the relaxation time, and the angular thermal load have significant effects on all the studied fields.
A Comparative Study of Generalized Autoregressive Conditional Heteroskedasticity (GARCH) and Extreme Value Theory (EVT) Model in Modeling Value-at-Risk (VaR)
The paper addresses the inefficiency of the classical model in measuring the Value-at-Risk (VaR) using a normal distribution or a Student’s t distribution. Specifically, the paper focuses on the one day ahead Value-at-Risk (VaR) of major stock market’s daily returns in US, UK, China and Hong Kong in the most recent ten years under 95% confidence level. To improve the predictable power and search for the best performing model, the paper proposes using two leading alternatives, Extreme Value Theory (EVT) and a family of GARCH models, and compares the relative performance. The main contribution could be summarized in two aspects. First, the paper extends the GARCH family model by incorporating EGARCH and TGARCH to shed light on the difference between each in estimating one day ahead Value-at-Risk (VaR). Second, to account for the non-normality in the distribution of financial markets, the paper applies Generalized Error Distribution (GED), instead of the normal distribution, to govern the innovation term. A dynamic back-testing procedure is employed to assess the performance of each model, a family of GARCH and the conditional EVT. The conclusion is that Exponential GARCH yields the best estimate in out-of-sample one day ahead Value-at-Risk (VaR) forecasting. Moreover, the discrepancy of performance between the GARCH and the conditional EVT is indistinguishable.
Generalized Mathematical Description and Simulation of Grid-Tied Thyristor Converters
Thyristor rectifiers, inverters grid-tied, and AC voltage regulators are widely used in industry, and on electrified transport, they have a lot in common both in the power circuit and in the control system. They have a common mathematical structure and switching processes. At the same time, the rectifier, but the inverter units and thyristor regulators of alternating voltage are considered separately both theoretically and practically. They are written about in different books as completely different devices. The aim of this work is to combine them into one class based on the unity of the equations describing electromagnetic processes, and then, to show this unity on the mathematical model and experimental setup. Based on research from mathematics to the product, a conclusion is made about the methodology for the rapid conduct of research and experimental design work, preparation for production and serial production of converters with a unified bundle. In recent years, there has been a transition from thyristor circuits and transistor in modular design. Showing the example of thyristor rectifiers and AC voltage regulators, we can conclude that there is a unity of mathematical structures and grid-tied thyristor converters.
Biosynthesis and Metabolism of Anthraquinone Derivatives
In review the generalized data about biosynthetic routs formation anthraquinone molecules in natural cells. The basic possibilities of various ways of biosynthesis of different quinoid substances are shown.
Computational Study of Chromatographic Behavior of a Series of S-Triazine Pesticides Based on Their in Silico Biological and Lipophilicity Descriptors
In this paper, quantitative structure-retention relationships (QSRR) analysis was applied in order to correlate in silico biological and lipophilicity molecular descriptors with retention values for the set of selected s-triazine herbicides. In silico generated biological and lipophilicity descriptors were discriminated using generalized pair correlation method (GPCM). According to this method, the significant difference between independent variables can be noticed regardless almost equal correlation with dependent variable. Using established multiple linear regression (MLR) models some biological characteristics could be predicted. Established MLR models were evaluated statistically and the most suitable models were selected and ranked using sum of ranking differences (SRD) method. In this method, as reference values, average experimentally obtained values are used. Additionally, using SRD method, similarities among investigated s-triazine herbicides can be noticed. These analysis were conducted in order to characterize selected s-triazine herbicides for future investigations regarding their biodegradability. This study is financially supported by COST action TD1305.
Enhancing a Recidivism Prediction Tool with Machine Learning: Effectiveness and Algorithmic Fairness
This work studies how Machine Learning (ML) may be used to increase the effectiveness of a criminal recidivism risk assessment tool, RisCanvi. The two key dimensions of this analysis are predictive accuracy and algorithmic fairness. ML-based prediction models obtained in this study are more accurate at predicting criminal recidivism than the manually-created formula used in RisCanvi, achieving an AUC of 0.76 and 0.73 in predicting violent and general recidivism respectively. However, the improvements are small, and it is noticed that algorithmic discrimination can easily be introduced between groups such as national vs foreigner, or young vs old. It is described how effectiveness and algorithmic fairness objectives can be balanced, applying a method in which a single error disparity in terms of generalized false positive rate is minimized, while calibration is maintained across groups. Obtained results show that this bias mitigation procedure can substantially reduce generalized false positive rate disparities across multiple groups. Based on these results, it is proposed that ML-based criminal recidivism risk prediction should not be introduced without applying algorithmic bias mitigation procedures.
Generalized Correlation for the Condensation and Evaporation Heat Transfer Coefficients of Propane (R290), Butane (R600), R134a, and R407c in Porous Horizontal Tubes: Experimental Investigation
This work is an experimental study on the heat transfer characteristics and pressure drop of different refrigerants during the condensation and evaporation processes in porous media. Four different refrigerants (R134a, R407C, 600a, R290), with different porosities were used to reach a real understanding of the actual heat transfer characteristics and pressure drop when using porous material inside the condenser and evaporator. Steel balls were used as porous media with different porosities (38%, 43%, 48%). The main goal of this project is to enhance the heat transfer coefficient during the condensation and evaporation processes when using different refrigerants and different porosities. Different correlations for the heat transfer coefficient and the pressure drop of the different refrigerants were developed. Also a generalized empirical correlation was developed for the different refrigerants. The experimental and predicted heat transfer coefficients and pressure drops were compared. It was found that, the Absolute standard deviation for the heat transfer coefficient and the pressure drop not exceeded values of 15% and 20%, respectively.
Generalization of Clustering Coefficient on Lattice Networks Applied to Criminal Networks
A lattice network is a special type of network in which all nodes have the same number of links, and its boundary conditions are periodic. The most basic lattice network is the ring, a one-dimensional network with periodic border conditions. In contrast, the Cartesian product of d rings forms a d-dimensional lattice network. An analytical expression currently exists for the clustering coefficient in this type of network, but the theoretical value is valid only up to certain connectivity value; in other words, the analytical expression is incomplete. Here we obtain analytically the clustering coefficient expression in d-dimensional lattice networks for any link density. Our analytical results show that the clustering coefficient for a lattice network with density of links that tend to 1, leads to the value of the clustering coefficient of a fully connected network. We developed a model on criminology in which the generalized clustering coefficient expression is applied. The model states that delinquents learn the know-how of crime business by sharing knowledge, directly or indirectly, with their friends of the gang. This generalization shed light on the network properties, which is important to develop new models in different fields where network structure plays an important role in the system dynamic, such as criminology, evolutionary game theory, econophysics, among others.
Generalized Correlation Coefficient in Genome-Wide Association Analysis of Cognitive Ability in Twins
Cognitive impairment in the elderly is a key issue affecting the quality of life. Despite a strong genetic background in cognition, only a limited number of single nucleotide polymorphisms (SNPs) have been found. These explain a small proportion of the genetic component of cognitive function, thus leaving a large proportion unaccounted for. We hypothesize that one reason for this missing heritability is the misspecified modeling in data analysis concerning phenotype distribution as well as the relationship between SNP dosage and the phenotype of interest. In an attempt to overcome these issues, we introduced a model-free method based on the generalized correlation coefficient (GCC) in a genome-wide association study (GWAS) of cognitive function in twin samples and compared its performance with two popular linear regression models. The GCC-based GWAS identified two genome-wide significant (P-value < 5e-8) SNPs; rs2904650 near ZDHHC2 on chromosome 8 and rs111256489 near CD6 on chromosome 11. The kinship model also detected two genome-wide significant SNPs, rs112169253 on chromosome 4 and rs17417920 on chromosome 7, whereas no genome-wide significant SNPs were found by the linear mixed model (LME). Compared to the linear models, more meaningful biological pathways like GABA receptor activation, ion channel transport, neuroactive ligand-receptor interaction, and the renin-angiotensin system were found to be enriched by SNPs from GCC. The GCC model outperformed the linear regression models by identifying more genome-wide significant genetic variants and more meaningful biological pathways related to cognitive function. Moreover, GCC-based GWAS was robust in handling genetically related twin samples, which is an important feature in handling genetic confounding in association studies.
Measures of Reliability and Transportation Quality on an Urban Rail Transit Network in Case of Links’ Capacities Loss
Urban rail transit (URT) plays a significant role in dealing with traffic congestion and environmental problems in cities. However, equipment failure and obstruction of links often lead to URT links’ capacities loss in daily operation. It affects the reliability and transport service quality of URT network seriously. In order to measure the influence of links’ capacities loss on reliability and transport service quality of URT network, passengers are divided into three categories in case of links’ capacities loss. Passengers in category 1 are less affected by the loss of links’ capacities. Their travel is reliable since their travel quality is not significantly reduced. Passengers in category 2 are affected by the loss of links’ capacities heavily. Their travel is not reliable since their travel quality is reduced seriously. However, passengers in category 2 still can travel on URT. Passengers in category 3 can not travel on URT because their travel paths’ passenger flow exceeds capacities. Their travel is not reliable. Thus, the proportion of passengers in category 1 whose travel is reliable is defined as reliability indicator of URT network. The transport service quality of URT network is related to passengers’ travel time, passengers’ transfer times and whether seats are available to passengers. The generalized travel cost is a comprehensive reflection of travel time, transfer times and travel comfort. Therefore, passengers’ average generalized travel cost is used as transport service quality indicator of URT network. The impact of links’ capacities loss on transport service quality of URT network is measured with passengers’ relative average generalized travel cost with and without links’ capacities loss. The proportion of the passengers affected by links and betweenness of links are used to determine the important links in URT network. The stochastic user equilibrium distribution model based on the improved logit model is used to determine passengers’ categories and calculate passengers’ generalized travel cost in case of links’ capacities loss, which is solved with method of successive weighted averages algorithm. The reliability and transport service quality indicators of URT network are calculated with the solution result. Taking Wuhan Metro as a case, the reliability and transport service quality of Wuhan metro network is measured with indicators and method proposed in this paper. The result shows that using the proportion of the passengers affected by links can identify important links effectively which have great influence on reliability and transport service quality of URT network; The important links are mostly connected to transfer stations and the passenger flow of important links is high; With the increase of number of failure links and the proportion of capacity loss, the reliability of the network keeps decreasing, the proportion of passengers in category 3 keeps increasing and the proportion of passengers in category 2 increases at first and then decreases; When the number of failure links and the proportion of capacity loss increased to a certain level, the decline of transport service quality is weakened.
Hybrid Algorithm for Non-Negative Matrix Factorization Based on Symmetric Kullback-Leibler Divergence for Signal Dependent Noise: A Case Study
Non-negative matrix factorization approximates a high dimensional non-negative matrix V as the product of two non-negative matrices, W and H, and allows only additive linear combinations of data, enabling it to learn parts with representations in reality. It has been successfully applied in the analysis and interpretation of high dimensional data arising in neuroscience, computational biology, and natural language processing, to name a few. The objective of this paper is to assess a hybrid algorithm for non-negative matrix factorization with multiplicative updates. The method aims to minimize the symmetric version of Kullback-Leibler divergence known as intrinsic information and assumes that the noise is signal-dependent and that it originates from an arbitrary distribution from the exponential family. It is a generalization of currently available algorithms for Gaussian, Poisson, gamma and inverse Gaussian noise. We demonstrate the potential usefulness of the new generalized algorithm by comparing its performance to the baseline methods which also aim to minimize symmetric divergence measures.
Physiological Action of Anthraquinone-Containing Preparations
In review the generalized data about biological activity of anthraquinone-containing plants and specimens on their basis is presented. Data of traditional medicine, results of bioscreening and clinical researches of specimens are analyzed.
Molecular Dynamics Simulation for Buckling Analysis at Nanocomposite Beams
In the present study we have investigated axial buckling characteristics of nanocomposite beams reinforced by single-walled carbon nanotubes (SWCNTs). Various types of beam theories including Euler-Bernoulli beam theory, Timoshenko beam theory and Reddy beam theory were used to analyze the buckling behavior of carbon nanotube-reinforced composite beams. Generalized differential quadrature (GDQ) method was utilized to discretize the governing differential equations along with four commonly used boundary conditions. The material properties of the nanocomposite beams were obtained using molecular dynamic (MD) simulation corresponding to both short-(10,10) SWCNT and long-(10,10) SWCNT composites which were embedded by amorphous polyethylene matrix. Then the results obtained directly from MD simulations were matched with those calculated by the mixture rule to extract appropriate values of carbon nanotube efficiency parameters accounting for the scale-dependent material properties. The selected numerical results were presented to indicate the influences of nanotube volume fractions and end supports on the critical axial buckling loads of nanocomposite beams relevant to long- and short-nanotube composites.
Feature Selection of Personal Authentication Based on EEG Signal for K-Means Cluster Analysis Using Silhouettes Score
Personal authentication based on electroencephalography (EEG) signals is one of the important field for the biometric technology. More and more researchers have used EEG signals as data source for biometric. However, there are some disadvantages for biometrics based on EEG signals. The proposed method employs entropy measures for feature extraction from EEG signals. Four type of entropies measures, sample entropy (SE), fuzzy entropy (FE), approximate entropy (AE) and spectral entropy (PE), were deployed as feature set. In a silhouettes calculation, the distance from each data point in a cluster to all another point within the same cluster and to all other data points in the closest cluster are determined. Thus silhouettes provide a measure of how well a data point was classified when it was assigned to a cluster and the separation between them. This feature renders silhouettes potentially well suited for assessing cluster quality in personal authentication methods. In this study, “silhouettes scores” was used for assessing the cluster quality of k-means clustering algorithm is well suited for comparing the performance of each EEG dataset. The main goals of this study are: (1) to represent each target as a tuple of multiple feature sets, (2) to assign a suitable measure to each feature set, (3) to combine different feature sets, (4) to determine the optimal feature weighting. Using precision/recall evaluations, the effectiveness of feature weighting in clustering was analyzed. EEG data from 22 subjects were collected. Results showed that: (1) It is possible to use fewer electrodes (3-4) for personal authentication. (2) There was the difference between each electrode for personal authentication (p< 0.01). (3) There is no significant difference for authentication performance among feature sets (except feature PE). Conclusion: The combination of k-means clustering algorithm and silhouette approach proved to be an accurate method for personal authentication based on EEG signals.
Error Amount in Viscoelasticity Analysis Depending on Time Step Size and Method used in ANSYS
Theory of viscoelasticity is used by many researchers to represent behavior of many materials such as pavements on roads or bridges. Several researches used analytical methods and rheology to predict the material behaviors of simple models. Today, more complex engineering structures are analyzed using Finite Element Method, in which material behavior is embedded by means of three dimensional viscoelastic material laws. As a result, structures of unordinary geometry and domain like pavements of bridges can be analyzed by means of Finite Element Method and three dimensional viscoelastic equations. In the scope of this study, rheological models embedded in ANSYS, namely, generalized Maxwell elements and Prony series, which are two methods used by ANSYS to represent viscoelastic material behavior, are presented explicitly. Subsequently, a practical problem, which has an analytical solution given in literature, is used to verify the applicability of viscoelasticity tool embedded in ANSYS. Finally, amount of error in the results of ANSYS is compared with the analytical results to indicate the influence of used method and time step size.
Decision Making Approach through Generalized Fuzzy Entropy Measure
Uncertainty is found everywhere and its understanding is central to decision making. Uncertainty emerges as one has less information than the total information required describing a system and its environment. Uncertainty and information are so closely associated that the information provided by an experiment for example, is equal to the amount of uncertainty removed. It may be pertinent to point out that uncertainty manifests itself in several forms and various kinds of uncertainties may arise from random fluctuations, incomplete information, imprecise perception, vagueness etc. For instance, one encounters uncertainty due to vagueness in communication through natural language. Uncertainty in this sense is represented by fuzziness resulting from imprecision of meaning of a concept expressed by linguistic terms. Fuzzy set concept provides an appropriate mathematical framework for dealing with the vagueness. Both information theory, proposed by Shannon (1948) and fuzzy set theory given by Zadeh (1965) plays an important role in human intelligence and various practical problems such as image segmentation, medical diagnosis etc. Numerous approaches and theories dealing with inaccuracy and uncertainty have been proposed by different researcher. In the present communication, we generalize fuzzy entropy proposed by De Luca and Termini (1972) corresponding to Shannon entropy(1948). Further, some of the basic properties of the proposed measure were examined. We also applied the proposed measure to the real life decision making problem.
CFD Modeling of Insect Flight at Low Reynolds Numbers
The typical insects employ a flapping-wing mode of flight. The numerical simulations on free flight of a model fruit fly (Re=143) including hovering and are presented in this paper. Unsteady aerodynamics around a flapping insect is studied by solving the three-dimensional Newtonian dynamics of the flyer coupled with Navier-Stokes equations. A hybrid-grid scheme (Generalized Finite Difference Method) that combines great geometry flexibility and accuracy of moving boundary definition is employed for obtaining flow dynamics. The results show good points of agreement and consistency with the outcomes and analyses of other researchers, which validate the computational model and demonstrate the feasibility of this computational approach on analyzing fluid phenomena in insect flight. The present modeling approach also offers a promising route of investigation that could complement as well as overcome some of the limitations of physical experiments in the study of free flight aerodynamics of insects. The results are potentially useful for the design of biomimetic flapping-wing flyers.
Model Predictive Control with Unscented Kalman Filter for Nonlinear Implicit Systems
A class of implicit systems is known as a more generalized class of systems than a class of explicit systems. To establish a control method for such a generalized class of systems, we adopt model predictive control method which is a kind of optimal feedback control with a performance index that has a moving initial time and terminal time. However, model predictive control method is inapplicable to systems whose all state variables are not exactly known. In other words, model predictive control method is inapplicable to systems with limited measurable states. In fact, it is usual that the state variables of systems are measured through outputs, hence, only limited parts of them can be used directly. It is also usual that output signals are disturbed by process and sensor noises. Hence, it is important to establish a state estimation method for nonlinear implicit systems with taking the process noise and sensor noise into consideration. To this purpose, we apply the model predictive control method and unscented Kalman filter for solving the optimization and estimation problems of nonlinear implicit systems, respectively. The objective of this study is to establish a model predictive control with unscented Kalman filter for nonlinear implicit systems.
Stabilization Control of the Nonlinear AIDS Model Based on the Theory of Polynomial Fuzzy Control Systems
In this paper, we introduced AIDS disease at first, then proposed dynamic model illustrate its progress, after expression of a short history of nonlinear modeling by polynomial phasing systems, we considered the stability conditions of the systems, which contained a huge amount of researches in order to modeling and control of AIDS in dynamic nonlinear form, in this approach using a frame work of control any polynomial phasing modeling system which have been generalized by part of phasing model of T-S, in order to control the system in better way, the stability conditions were achieved based on polynomial functions, then we focused to design the appropriate controller, firstly we considered the equilibrium points of system and their conditions and in order to examine changes in the parameters, we presented polynomial phase model that was the generalized approach rather than previous Takagi Sugeno models, then with using case we evaluated the equations in both open loop and close loop and with helping the controlling feedback, the close loop equations of system were calculated, to simulate nonlinear model of AIDS disease, we used polynomial phasing controller output that was capable to make the parameters of a nonlinear system to follow a sustainable reference model properly.
Measuring Financial Asset Return and Volatility Spillovers, with Application to Sovereign Bond, Equity, Foreign Exchange and Commodity Markets
We provide an in-depth analysis of interdependence of asset returns and volatilities in developed and developing countries. The analysis is split into three parts. In the first part, we use multivariate GARCH model in order to provide stylized facts on cross-market volatility spillovers. In the second part, we use a generalized vector autoregressive methodology developed by Diebold and Yilmaz (2009) in order to estimate separate measures of return spillovers and volatility spillovers among sovereign bond, equity, foreign exchange and commodity markets. In particular, our analysis is focused on cross-market return, and volatility spillovers in 19 developed and developing countries. In order to estimate named spillovers, we use daily data from 2008 to 2017. In the third part of the analysis, we use a generalized vector autoregressive framework in order to estimate total and directional volatility spillovers. We use the same daily data span for one developed and one developing country in order to characterize daily volatility spillovers across stock, bond, foreign exchange and commodities markets.
Time/Temperature-Dependent Finite Element Model of Laminated Glass Beams
The polymer foil used for manufacturing of laminated glass members behaves in a viscoelastic manner with temperature dependence. This contribution aims at incorporating the time/temperature-dependent behavior of interlayer to our earlier elastic finite element model for laminated glass beams. The model is based on a refined beam theory: each layer behaves according to the finite-strain shear deformable formulation by Reissner and the adjacent layers are connected via the Lagrange multipliers ensuring the inter-layer compatibility of a laminated unit. The time/temperature-dependent behavior of the interlayer is accounted for by the generalized Maxwell model and by the time-temperature superposition principle due to the Williams, Landel, and Ferry. The resulting system is solved by the Newton method with consistent linearization and the viscoelastic response is determined incrementally by the exponential algorithm. By comparing the model predictions against available experimental data, we demonstrate that the proposed formulation is reliable and accurately reproduces the behavior of the laminated glass units.
Regional Hydrological Extremes Frequency Analysis Based on Statistical and Hydrological Models
The hydrological extremes frequency analysis is the foundation for the hydraulic engineering design, flood protection, drought management and water resources management and planning to utilize the available water resource to meet the desired objectives of different organizations and sectors in a country. This spatial variation of the statistical characteristics of the extreme flood and drought events are key practice for regional flood and drought analysis and mitigation management. For different hydro-climate of the regions, where the data set is short, scarcity, poor quality and insufficient, the regionalization methods are applied to transfer at-site data to a region. This study aims in regional high and low flow frequency analysis for Poland River Basins. Due to high frequent occurring of hydrological extremes in the region and rapid water resources development in this basin have caused serious concerns over the flood and drought magnitude and frequencies of the river in Poland. The magnitude and frequency result of high and low flows in the basin is needed for flood and drought planning, management and protection at present and future. Hydrological homogeneous high and low flow regions are formed by the cluster analysis of site characteristics, using the hierarchical and C- mean clustering and PCA method. Statistical tests for regional homogeneity are utilized, by Discordancy and Heterogeneity measure tests. In compliance with results of the tests, the region river basin has been divided into ten homogeneous regions. In this study, frequency analysis of high and low flows using AM for high flow and 7-day minimum low flow series is conducted using six statistical distributions. The use of L-moment and LL-moment method showed a homogeneous region over entire province with Generalized logistic (GLOG), Generalized extreme value (GEV), Pearson type III (P-III), Generalized Pareto (GPAR), Weibull (WEI) and Power (PR) distributions as the regional drought and flood frequency distributions. The 95% percentile and Flow duration curves of 1, 7, 10, 30 days have been plotted for 10 stations. However, the cluster analysis performed two regions in west and east of the province where L-moment and LL-moment method demonstrated the homogeneity of the regions and GLOG and Pearson Type III (PIII) distributions as regional frequency distributions for each region, respectively. The spatial variation and regional frequency distribution of flood and drought characteristics for 10 best catchment from the whole region was selected and beside the main variable (streamflow: high and low) we used variables which are more related to physiographic and drainage characteristics for identify and delineate homogeneous pools and to derive best regression models for ungauged sites. Those are mean annual rainfall, seasonal flow, average slope, NDVI, aspect, flow length, flow direction, maximum soil moisture, elevation, and drainage order. The regional high-flow or low-flow relationship among one streamflow characteristics with (AM or 7-day mean annual low flows) some basin characteristics is developed using Generalized Linear Mixed Model (GLMM) and Generalized Least Square (GLS) regression model, providing a simple and effective method for estimation of flood and drought of desired return periods for ungauged catchments.
Extreme Value Modelling of Ghana Stock Exchange Indices
Modelling of extreme events has always been of interest in fields such as hydrology and meteorology. However, after the recent global financial crises, appropriate models for modelling of such rare events leading to these crises have become quite essential in the finance and risk management fields. This paper models the extreme values of the Ghana Stock Exchange All-Shares indices (2000-2010) by applying the Extreme Value Theory to fit a model to the tails of the daily stock returns data. A conditional approach of the EVT was preferred and hence an ARMA-GARCH model was fitted to the data to correct for the effects of autocorrelation and conditional heteroscedastic terms present in the returns series, before EVT method was applied. The Peak Over Threshold (POT) approach of the EVT, which fits a Generalized Pareto Distribution (GPD) model to excesses above a certain selected threshold, was employed. Maximum likelihood estimates of the model parameters were obtained and the model’s goodness of fit was assessed graphically using Q-Q, P-P and density plots. The findings indicate that the GPD provides an adequate fit to the data of excesses. The size of the extreme daily Ghanaian stock market movements were then computed using the Value at Risk (VaR) and Expected Shortfall (ES) risk measures at some high quantiles, based on the fitted GPD model.
Detecting Earnings Management via Statistical and Neural Networks Techniques
Predicting earnings management is vital for the capital market participants, financial analysts and managers. The aim of this research is attempting to respond to this query: Is there a significant difference between the regression model and neural networks’ models in predicting earnings management, and which one leads to a superior prediction of it? In approaching this question, a Linear Regression (LR) model was compared with two neural networks including Multi-Layer Perceptron (MLP), and Generalized Regression Neural Network (GRNN). The population of this study includes 94 listed companies in Tehran Stock Exchange (TSE) market from 2003 to 2011. After the results of all models were acquired, ANOVA was exerted to test the hypotheses. In general, the summary of statistical results showed that the precision of GRNN did not exhibit a significant difference in comparison with MLP. In addition, the mean square error of the MLP and GRNN showed a significant difference with the multi variable LR model. These findings support the notion of nonlinear behavior of the earnings management. Therefore, it is more appropriate for capital market participants to analyze earnings management based upon neural networks techniques, and not to adopt linear regression models.
Bayesian Analysis of Topp-Leone Generalized Exponential Distribution
The Topp-Leone distribution was introduced by Topp- Leone in 1955. In this paper, an attempt has been made to fit Topp-Leone Generalized exponential (TPGE) distribution. A real survival data set is used for illustrations. Implementation is done using R and JAGS and appropriate illustrations are made. R and JAGS codes have been provided to implement censoring mechanism using both optimization and simulation tools. The main aim of this paper is to describe and illustrate the Bayesian modelling approach to the analysis of survival data. Emphasis is placed on the modeling of data and the interpretation of the results. Crucial to this is an understanding of the nature of the incomplete or 'censored' data encountered. Analytic approximation and simulation tools are covered here, but most of the emphasis is on Markov chain based Monte Carlo method including independent Metropolis algorithm, which is currently the most popular technique. For analytic approximation, among various optimization algorithms and trust region method is found to be the best. In this paper, TPGE model is also used to analyze the lifetime data in Bayesian paradigm. Results are evaluated from the above mentioned real survival data set. The analytic approximation and simulation methods are implemented using some software packages. It is clear from our findings that simulation tools provide better results as compared to those obtained by asymptotic approximation.
The Generalized Lemaitre-Tolman-Bondi Solutions in Modeling the Cosmological Black Holes
In spite of the numerous attempts to close the discussion about the influence of cosmological expansion on local gravitationally bounded systems, this question arises in literature again and again and remains still far from its final resolution. Here one of the main problems is the problem of obtaining a physically adequate model of strongly gravitating object immersed in non-static cosmological background. Such objects are usually called ‘cosmological’ black holes and are of great interest in wide set of cosmological and astrophysical areas. In this work the set of new exact solutions of the Einstein equations is derived for the flat space that generalizes the known Lemaitre-Tolman-Bondi solution for the case of nonzero pressure. The solutions obtained are pretending to describe the black hole immersed in nonstatic cosmological background and give a possibility to investigate the hot problems concerning the effects of the cosmological expansion in gravitationally bounded systems, the structure formation in the early universe, black hole thermodynamics and other related problems. It is shown that each of the solutions obtained contains either the Reissner-Nordstrom or the Schwarzschild black hole in the central region of the space. It is demonstrated that the approach of the mass function use in solving of the Einstein equations allows clear physical interpretation of the resulting solutions, that is of much benefit to any their concrete application.
Joint Optimal Pricing and Lot-Sizing Decisions for an Advance Sales System under Stochastic Conditions
In this paper, we investigate the effect of stochastic inputs on problem of joint optimal pricing and lot-sizing decisions where the inventory cycle is divided into advance and spot sales periods. During the advance sales period, customer can make reservations while customer with reservations can cancel their order. However, during the spot sales period customers receive the order as soon as the order is placed, but they cannot make any reservation or cancellation during that period. We assume that the inter arrival times during the advance sales and spot sales period are exponentially distributed where the arrival rate is decreasing function of price. Moreover, we assume that the number of cancelled reservations is binomially distributed. In addition, we assume that deterioration process follows an exponential distribution. We investigate two cases. First, we consider two-state case where we find the optimal price during the spot sales period and the optimal price during the advance sales period. Next, we develop a generalized case where we extend two-state case also to allow dynamic prices during the spot sales period. We apply the Markov decision theory in order to find the optimal solutions. In addition, for the generalized case, we apply the policy iteration algorithm in order to find the optimal prices, the optimal lot-size and maximum advance sales amount.
Complex Network Analysis of Seismicity and Applications to Short-Term Earthquake Forecasting
Earthquakes are complex phenomena, exhibiting complex correlations in space, time, and magnitude. Recently, the concept of complex networks has been used to shed light on the statistical and dynamical characteristics of regional seismicity. In this work, we study the relationships and interactions of seismic regions in Chile, Japan, and the Philippines through weighted and directed complex network analysis. Geographical areas are digitized into cells of fixed dimensions which in turn become the nodes of the network when an earthquake has occurred therein. Nodes are linked if a correlation exists between them as determined and measured by a correlation metric. The networks are found to be scale-free, exhibiting power-law behavior in the distributions of their different centrality measures: the in- and out-degree and the in- and out-strength. The evidence is also found of preferential interaction between seismically active regions through their degree-degree correlations suggesting that seismicity is dictated by the activity of a few active regions. The importance of a seismic region to the overall seismicity is measured using a generalized centrality metric taken to be an indicator of its activity or passivity. The spatial distribution of earthquake activity indicates the areas where strong earthquakes have occurred in the past while the passivity distribution points toward the likely locations an earthquake would occur whenever another one happens elsewhere. Finally, we propose a method that would project the location of the next possible earthquake using the generalized centralities coupled with correlations calculated between the latest earthquakes and a geographical point in the future.
Nonlocal Beam Models for Free Vibration Analysis of Double-Walled Carbon Nanotubes with Various End Supports
In the present study, the free vibration characteristics of double-walled carbon nanotubes (DWCNTs) are investigated. The small-scale effects are taken into account using the Eringen’s nonlocal elasticity theory. The nonlocal elasticity equations are implemented into the different classical beam theories namely as Euler-Bernoulli beam theory (EBT), Timoshenko beam theory (TBT), Reddy beam theory (RBT), and Levinson beam theory (LBT) to analyze the free vibrations of DWCNTs in which each wall of the nanotubes is considered as individual beam with van der Waals interaction forces. Generalized differential quadrature (GDQ) method is utilized to discretize the governing differential equations of each nonlocal beam model along with four commonly used boundary conditions. Then molecular dynamics (MD) simulation is performed for a series of armchair and zigzag DWCNTs with different aspect ratios and boundary conditions, the results of which are matched with those of nonlocal beam models to extract the appropriate values of the nonlocal parameter corresponding to each type of chirality, nonlocal beam model and boundary condition. It is found that the present nonlocal beam models with their proposed correct values of nonlocal parameter have good capability to predict the vibrational behavior of DWCNTs, especially for higher aspect ratios.
Magnetic and Optical Properties of GaFeMnN
The full-potential linearized augmented plane wave method (FP-LAPW) within the Generalized Gradient Approximation (GGA) is used to calculate the magnetic and optical properties of quaternary GaFeMnN. The results show that the compound becomes magnetic and half metallic and there is an apparition of peaks at low frequencies for the optical properties.
Determinants of the Shadow Economy with an Islamic Orientation: An Application to Organization of Islamic Cooperation and Non-Organization of Islamic Cooperation Countries
The main objective of Islamic Finance is to promote social justice thorough financial inclusion and redistribution of economic resources between rich and poor. The approach of Islamic finance is more comprehensive in nature and covers both formal and informal sectors of the economy, first, through reducing the gap between both sectors, and second by using specific Islamic values to reallocate the wealth between formal and informal sectors. Applying Generalized Method of Movements (GMM) to the annual data spanning from 1995-2015 for 141 countries, this study explores the determinants of informal business sector in Organization of Islamic Cooperation (OIC) countries and then compares with Non-OIC countries. Economic freedom and institutions variables as well as economic growth and money supply are found to reduce informal business sector in both OIC and Non-OIC nations while government expenditure are found to increase informal business sector in both group of nations. Informal Business sector remain the same in both types of countries but still the majority Muslim population in OIC economies create main difference between both groups of nations and justify the potential role of Islamic Finance in informal business sector in OIC nations. The study suggests that institutions quality should be improved and entrepreneurs’ friendly business environment must be provided. This study refines the main features of informal business sector and discuss their implications on policy designing and implementation, particularly in the context of Islamic finance fight against poverty, inequality and improving living standards of informal sector participants in OIC countries.
Molecular Dynamics Simulation for Vibration Analysis at Nanocomposite Plates
Polymer/carbon nanotube nanocomposites have a wide range of promising applications Due to their enhanced properties. In this work, free vibration analysis of single-walled carbon nanotube-reinforced composite plates is conducted in which carbon nanotubes are embedded in an amorphous polyethylene. The rule of mixture based on various types of plate model namely classical plate theory (CLPT), first-order shear deformation theory (FSDT), and higher-order shear deformation theory (HSDT) was employed to obtain fundamental frequencies of the nanocomposite plates. Generalized differential quadrature (GDQ) method was used to discretize the governing differential equations along with the simply supported and clamped boundary conditions. The material properties of the nanocomposite plates were evaluated using molecular dynamic (MD) simulation corresponding to both short-(10,10) SWCNT and long-(10,10) SWCNT composites. Then the results obtained directly from MD simulations were fitted with those calculated by the rule of mixture to extract appropriate values of carbon nanotube efficiency parameters accounting for the scale-dependent material properties. The selected numerical results are presented to address the influences of nanotube volume fraction and edge supports on the value of fundamental frequency of carbon nanotube-reinforced composite plates corresponding to both long- and short-nanotube composites.
A Generalised Propensity Score Analysis to Investigate the Influence of Agricultural Research Systems on Greenhouse Gas Emissions
Bioeconomy can give the chance to face new global challenges and can move ahead the transition from a waste economy to an economy based on renewable resources and sustainable consumption. Air pollution is a grave issue in green challenges, mainly caused by anthropogenic factors. The agriculture sector is a great contributor to global greenhouse gases (GHGs) emissions due to lacking efficient management of the resources involved and research policies. In particular, livestock sector contributes to emissions of GHGs, deforestation, and nutrient imbalances. More effective agricultural research systems and technologies are crucial in order to improve farm productivity but also to reduce the GHGs emissions. Using data from FAOSTAT statistics and concern the EU countries; the aim of this research is to evaluate the impact of ASTI R&D (Agricultural Science and Technology Indicators) on GHGs emissions for countries EU in 2015 by generalized propensity score procedures, estimating a dose-response function, also considering a set of covariates. Expected results show the existence of the influence of ASTI R&D on GHGs across EU countries. Implications are crucial: reducing GHGs emissions by means of R&D based policies and correlatively reaching eco-friendly management of required resources by means of green available practices could have a crucial role for fair intra-generational implications.
Approximation by Generalized Lupaş-Durrmeyer Operators with Two Parameter α and β
This paper deals with the Stancu type generalization of Lupaş-Durrmeyer operators. We establish some direct results in the polynomial weighted space of continuous functions defined on the interval [0, 1]. Also, Voronovskaja type theorem is studied.
Bayesian Locally Approach for Spatial Modeling of Visceral Leishmaniasis Infection in Northern and Central Tunisia
This paper develops a Local Generalized Linear Spatial Model (LGLSM) to describe the spatial variation of Visceral Leishmaniasis (VL) infection risk in northern and central Tunisia. The response from each region is a number of affected children less than five years of age recorded from 1996 through 2006 from Tunisian pediatric departments and treated as a poison county level data. The model includes climatic factors, namely averages of annual rainfall, extreme values of low temperatures in winter and high temperatures in summer to characterize the climate of each region according to each continentality index, the pluviometric quotient of Emberger (Q2) to characterize bioclimatic regions and component for residual extra-poison variation. The statistical results show the progressive increase in the number of affected children in regions with high continentality index and low mean yearly rainfull. On the other hand, an increase in pluviometric quotient of Emberger contributed to a significant increase in VL incidence rate. When compared with the original GLSM, Bayesian locally modeling is improvement and gives a better approximation of the Tunisian VL risk estimation. According to the Bayesian approach inference, we use vague priors for all parameters model and Markov Chain Monte Carlo method.
Automated Natural Hazard Zonation System with Internet-SMS Warning: Distributed GIS for Sustainable Societies Creating Schema and Interface for Mapping and Communication
The research describes the implementation of a novel and stand-alone system for dynamic hazard warning. The system uses all existing infrastructure already in place like mobile networks, a laptop/PC and the small installation software. The geospatial dataset are the maps of a region which are again frugal. Hence there is no need to invest and it reaches everyone with a mobile. A novel architecture of hazard assessment and warning introduced where major technologies in ICT interfaced to give a unique WebGIS based dynamic real time geohazard warning communication system. A never before architecture introduced for integrating WebGIS with telecommunication technology. Existing technologies interfaced in a novel architectural design to address a neglected domain in a way never done before–through dynamically updatable WebGIS based warning communication. The work publishes new architecture and novelty in addressing hazard warning techniques in sustainable way and user friendly manner. Coupling of hazard zonation and hazard warning procedures into a single system has been shown. Generalized architecture for deciphering a range of geo-hazards has been developed. Hence the developmental work presented here can be summarized as the development of internet-SMS based automated geo-hazard warning communication system; integrating a warning communication system with a hazard evaluation system; interfacing different open-source technologies towards design and development of a warning system; modularization of different technologies towards development of a warning communication system; automated data creation, transformation and dissemination over different interfaces. The architecture of the developed warning system has been functionally automated as well as generalized enough that can be used for any hazard and setup requirement has been kept to a minimum.
A Bayesian Parameter Identification Method for Thermorheological Complex Materials
Polymers increasingly gained interest in construction materials over the last years in civil engineering applications. As polymeric materials typically show time- and temperature dependent material behavior, which is accounted for in the context of the theory of linear viscoelasticity. Within the context of this paper, the authors show, that some polymeric interlayers for laminated glass can not be considered as thermorheologically simple as they do not follow a simple TTSP, thus a methodology of identifying the thermorheologically complex constitutive bahavioir is needed. ‘Dynamical-Mechanical-Thermal-Analysis’ (DMTA) in tensile and shear mode as well as ‘Differential Scanning Caliometry’ (DSC) tests are carried out on the interlayer material ‘Ethylene-vinyl acetate’ (EVA). A navoel Bayesian framework for the Master Curving Process as well as the detection and parameter identification of the TTSPs along with their associated Prony-series is derived and applied to the EVA material data. To our best knowledge, this is the first time, an uncertainty quantification of the Prony-series in a Bayesian context is shown. Within this paper, we could successfully apply the derived Bayesian methodology to the EVA material data to gather meaningful Master Curves and TTSPs. Uncertainties occurring in this process can be well quantified. We found, that EVA needs two TTSPs with two associated Generalized Maxwell Models. As the methodology is kept general, the derived framework could be also applied to other thermorheologically complex polymers for parameter identification purposes.
Evaluation of Newly Synthesized Steroid Derivatives Using In silico Molecular Descriptors and Chemometric Techniques
This study considered selection of the in silico molecular descriptors and the models for newly synthesized steroid derivatives description and their characterization using chemometric techniques. Multiple linear regression (MLR) models were established and gave the best molecular descriptors for quantitative structure-retention relationship (QSRR) modeling of the retention of the investigated molecules. MLR models were without multicollinearity among the selected molecular descriptors according to the variance inflation factor (VIF) values. Used molecular descriptors were ranked using generalized pair correlation method (GPCM). In this method, the significant difference between independent variables can be noticed regardless almost equal correlation between dependent variable. Generated MLR models were statistically and cross-validated and the best models were kept. Models were ranked using sum of ranking differences (SRD) method. According to this method, the most consistent QSRR model can be found and similarity or dissimilarity between the models could be noticed. In this study, SRD was performed using average values of experimentally observed data as a golden standard. Chemometric analysis was conducted in order to characterize newly synthesized steroid derivatives for further investigation regarding their potential biological activity and further synthesis. This article is based upon work from COST Action (CM1105), supported by COST (European Cooperation in Science and Technology).
Direct CP Violation in Baryonic B-Hadron Decays
We study direct CP-violating asymmetries (CPAs) in the baryonic B decays of B- -> p\bar{p}M and Λb decays of Λb ®pM andΛb -> J/ΨpM with M=π-, K-,ρ-,K*- based on the generalized factorization method in the standard model (SM). In particular, we show that the CPAs in the vector modes of B-®p\bar{p}K* and Λb -> p K*- can be as large as 20%. We also discuss the simplest purely baryonic decays of Λb-> p\bar{p}n, p\bar{p}Λ, Λ\bar{p}Λ, and Λ\bar{Λ}Λ. We point out that some of CPAs are promising to be measured by the current as well as future B facilities.
Magnetic and Optical Properties of Quaternary GaFeMnN
The full-potential linearized augmented plane wave method (FP-LAPW) within the Generalized Gradient Approximation (GGA) is used to calculate the magnetic and optical properties of quaternary GaFeMnN. The results show that the compound becomes magnetic and half metallic and there is an apparition of peaks at low frequencies for the optical properties.
Nonparametric Truncated Spline Regression Model on the Data of Human Development Index in Indonesia
Human Development Index (HDI) is a standard measurement for a country's human development. Several factors may have influenced it, such as life expectancy, gross domestic product (GDP) based on the province's annual expenditure, the number of poor people, and the percentage of an illiterate people. The scatter plot between HDI and the influenced factors show that the plot does not follow a specific pattern or form. Therefore, the HDI's data in Indonesia can be applied with a nonparametric regression model. The estimation of the regression curve in the nonparametric regression model is flexible because it follows the shape of the data pattern. One of the nonparametric regression's method is a truncated spline. Truncated spline regression is one of the nonparametric approach, which is a modification of the segmented polynomial functions. The estimator of a truncated spline regression model was affected by the selection of the optimal knots point. Knot points is a focus point of spline truncated functions. The optimal knots point was determined by the minimum value of generalized cross validation (GCV). In this article were applied the data of Human Development Index with a truncated spline nonparametric regression model. The results of this research were obtained the best-truncated spline regression model to the HDI's data in Indonesia with the combination of optimal knots point 5-5-5-4. Life expectancy and the percentage of an illiterate people were the significant factors depend to the HDI in Indonesia. The coefficient of determination is 94.54%. This means the regression model is good enough to applied on the data of HDI in Indonesia.
Perceptual Image Coding by Exploiting Internal Generative Mechanism
In the perceptual image coding, the objective is to shape the coding distortion such that the amplitude of distortion does not exceed the error visibility threshold, or to remove perceptually redundant signals from the image. While most researches focus on color image coding, the perceptual-based quantizer developed for luminance signals are always directly applied to chrominance signals such that the color image compression methods are inefficient. In this paper, the internal generative mechanism is integrated into the design of a color image compression method. The internal generative mechanism working model based on the structure-based spatial masking is used to assess the subjective distortion visibility thresholds that are visually consistent to human eyes better. The estimation method of structure-based distortion visibility thresholds for color components is further presented in a locally adaptive way to design quantization process in the wavelet color image compression scheme. Since the lowest subband coefficient matrix of images in the wavelet domain preserves the local property of images in the spatial domain, the error visibility threshold inherent in each coefficient of the lowest subband for each color component is estimated by using the proposed spatial error visibility threshold assessment. The threshold inherent in each coefficient of other subbands for each color component is then estimated in a local adaptive fashion based on the distortion energy allocation. By considering that the error visibility thresholds are estimated using predicting and reconstructed signals of the color image, the coding scheme incorporated with locally adaptive perceptual color quantizer does not require side information. Experimental results show that the entropies of three color components obtained by using proposed IGM-based color image compression scheme are lower than that obtained by using the existing color image compression method at perceptually lossless visual quality.
Technologies of Isolation and Separation of Anthraquinone Derivatives
In review the generalized data about different methods of extraction, separation and purification of natural and modify anthraquinones is presented. The basic regularity of an isolation process is analyzed. Action of temperature, pH, and polarity of extragent, catalysts and other factors on an isolation process is revealed.
A View of Flexible Housing in China
Beginning with the debate of concept, this essay explains the historical source and development of flexible housing in China. In the former part, the flexibility contained in traditional house is explored. While in the latter, the relevant practices in modern times are systematically analyzed as three phases–the Embryonic Period (1949 - 1980), the Systematic Practice (1981 - 2000), as well as the Integrated Trend and Prosperity (2001 - present). As a conclusion, the generalized flexibility is tentatively discussed.
Jensen's Inequality and M-Convex Functions
In this paper, we generalized the Jensen's inequality for m-convex functions and also we present a correction of Jensen's inequality which is a better than the generalization of this inequality for m-convex functions. Finally, we have found new lower and new upper bounds for Jensen's discrete inequality.
Predicting Stem Borer Density in Maize Using RapidEye Data and Generalized Linear Models
Maize (Zea mays L.) is a major staple food crop in Africa, particularly in the eastern region of the continent. The maize growing area in Africa spans over 25 million ha and 84% of rural households in Africa cultivate maize mainly as a means to generate food and income. Average maize yields in Sub Saharan Africa are 1.4 t/ha as compared to global average of 2.5–3.9 t/ha due to biotic and abiotic constraints. Amongst the biotic production constraints in Africa, stem borers are the most injurious. In East Africa, yield losses due to stem borers are currently estimated between 12% to 40% of the total production. The objective of the present study was therefore to predict stem borer larvae density in maize fields using RapidEye reflectance data and generalized linear models (GLMs). RapidEye images were captured for a test site in Kenya (Machakos) in January and in February 2015. Stem borer larva numbers were modeled using GLMs assuming Poisson (Po) and negative binomial (NB) distributions with error with log arithmetic link. Root mean square error (RMSE) and ratio prediction to deviation (RPD) statistics were employed to assess the models performance using a leave one-out cross-validation approach. Results showed that NB models outperformed Po ones in all study sites. RMSE and RPD ranged between 0.95 and 2.70, and between 2.39 and 6.81, respectively. Overall, all models performed similar when used the January and the February image data. We conclude that reflectance data from RapidEye data can be used to estimate stem borer larvae density. The developed models could to improve decision making regarding controlling maize stem borers using various integrated pest management (IPM) protocols.
Differential Approach to Technology Aided English Language Teaching: A Case Study in a Multilingual Setting
Rapid evolution of technology has changed language pedagogy as well as perspectives on language use, leading to strategic changes in discourse studies. We are now firmly embedded in a time when digital technologies have become an integral part of our daily lives. This has led to generalized approaches to English Language Teaching (ELT) which has raised two-pronged concerns in linguistically diverse settings: a) the diverse linguistic background of the learner might interfere/ intervene with the learning process and b) the differential level of already acquired knowledge of target language might make the classroom practices too easy or too difficult for the target group of learners. ELT needs a more systematic and differential pedagogical approach for greater efficiency and accuracy. The present research analyses the need of identifying learner groups based on different levels of target language proficiency based on a longitudinal study done on 150 undergraduate students. The learners were divided into five groups based on their performance on a twenty point scale in Listening Speaking Reading and Writing (LSRW). The groups were then subjected to varying durations of technology aided language learning sessions and their performance was recorded again on the same scale. Identifying groups and introducing differential teaching and learning strategies led to better results compared to generalized teaching strategies. Language teaching includes different aspects: the organizational, the technological, the sociological, the psychological, the pedagogical and the linguistic. And a facilitator must account for all these aspects in a carefully devised differential approach meeting the challenge of learner diversity. Apart from the justification of the formation of differential groups the paper attempts to devise framework to account for all these aspects in order to make ELT in multilingual setting much more effective.
Extension of Positive Linear Operator
This research consideres the extension of special functions called Positive Linear Operators. the bounded linear operator which defined from normed space to Banach space will extend to the closure of the its domain, And extend identified linear functional on a vector subspace by Hana-Banach theorem which could be generalized to the positive linear operators.
The Impact of Board Characteristics on Firm Performance: Evidence from Banking Industry in India
The Board of Directors in a firm performs the primary role of an internal control mechanism. This Study seeks to understand the relationship between internal governance and performance of banks in India. The research paper investigates the effect of board structure (proportion of nonexecutive directors, gender diversity, board size and meetings per year) on the firm performance. This paper evaluates the impact of corporate governance mechanisms on bank’s financial performance using panel data for 28 listed banks in National Stock Exchange of India for the period of 2008-2014. Returns on Asset, Return on Equity, Tobin’s Q and Net Interest Margin were used as the financial performance indicators. To estimate the relationship among governance and bank performance initially the Study uses Pooled Ordinary Least Square (OLS) Estimation and Generalized Least Square (GLS) Estimation. Then a well-developed panel Generalized Method of Moments (GMM) Estimator is developed to investigate the dynamic nature of performance and governance relationship. The Study empirically confirms that two-step system GMM approach controls the problem of unobserved heterogeneity and endogeneity as compared to the OLS and GLS approach. The result suggests that banks with small board, boards with female members, and boards that meet more frequently tend to be more efficient and subsequently have a positive impact on performance of banks. The study offers insights to policy makers interested in enhancing the quality of governance of banks in India. Also, the findings suggest that board structure plays a vital role in the improvement of corporate governance mechanism for financial institutions. There is a need to have efficient boards in banks to improve the overall health of the financial institutions and the economic development of the country.
A Variational Reformulation for the Thermomechanically Coupled Behavior of Shape Memory Alloys
Thanks to their unusual properties, shape memory alloys (SMAs) are good candidates for advanced applications in a wide range of engineering fields, such as automotive, robotics, civil, biomedical, aerospace. In the last decades, the ever-growing interest for such materials has boosted several research studies aimed at modeling their complex nonlinear behavior in an effective and robust way. Since the constitutive response of SMAs is strongly thermomechanically coupled, the investigation of the non-isothermal evolution of the material must be taken into consideration. The present study considers an existing three-dimensional phenomenological model for SMAs, able to reproduce the main SMA properties while maintaining a simple user-friendly structure, and proposes a variational reformulation of the full non-isothermal version of the model. While the considered model has been thoroughly assessed in an isothermal setting, the proposed formulation allows to take into account the full nonisothermal problem. In particular, the reformulation is inspired to the GENERIC (General Equations for Non-Equilibrium Reversible-Irreversible Coupling) formalism, and is based on a generalized gradient flow of the total entropy, related to thermal and mechanical variables. Such phrasing of the model is new and allows for a discussion of the model from both a theoretical and a numerical point of view. Moreover, it directly implies the dissipativity of the flow. A semi-implicit time-discrete scheme is also presented for the fully coupled thermomechanical system, and is proven unconditionally stable and convergent. The correspondent algorithm is then implemented, under a space-homogeneous temperature field assumption, and tested under different conditions. The core of the algorithm is composed of a mechanical subproblem and a thermal subproblem. The iterative scheme is solved by a generalized Newton method. Numerous uniaxial and biaxial tests are reported to assess the performance of the model and algorithm, including variable imposed strain, strain rate, heat exchange properties, and external temperature. In particular, the heat exchange with the environment is the only source of rate-dependency in the model. The reported curves clearly display the interdependence between phase transformation strain and material temperature. The full thermomechanical coupling allows to reproduce the exothermic and endothermic effects during respectively forward and backward phase transformation. The numerical tests have thus demonstrated that the model can appropriately reproduce the coupled SMA behavior in different loading conditions and rates. Moreover, the algorithm has proved effective and robust. Further developments are being considered, such as the extension of the formulation to the finite-strain setting and the study of the boundary value problem.
Correlations in the Ising Kagome Lattice
Using a previously developed procedure and with the aid of algebraic software, a two-dimensional generalized Ising model with a 4×2 unitary cell (UC), we obtain a Kagome Lattice with twelve different spin-spin values of interaction, in order to determine the partition function per spin L(T). From the partition function we can study the magnetic behavior of the system. Because of the competition phenomenon between spins, a very complex behavior among them in a variety of magnetic states can be observed.
Nonstationary Modeling of Extreme Precipitation in the Wei River Basin, China
Under the impact of global warming together with the intensification of human activities, the hydrological regimes may be altered, and the traditional stationary assumption was no longer satisfied. However, most of the current design standards of water infrastructures were still based on the hypothesis of stationarity, which may inevitably result in severe biases. Many critical impacts of climate on ecosystems, society, and the economy are controlled by extreme events rather than mean values. Therefore, it is of great significance to identify the non-stationarity of precipitation extremes and model the precipitation extremes in a nonstationary framework. The Wei River Basin (WRB), located in a continental monsoon climate zone in China, is selected as a case study in this study. Six extreme precipitation indices were employed to investigate the changing patterns and stationarity of precipitation extremes in the WRB. To identify if precipitation extremes are stationary, the Mann-Kendall trend test and the Pettitt test, which is used to examine the occurrence of abrupt changes are adopted in this study. Extreme precipitation indices series are fitted with non-stationary distributions that selected from six widely used distribution functions: Gumbel, lognormal, Weibull, gamma, generalized gamma and exponential distributions by means of the time-varying moments model generalized additive models for location, scale and shape (GAMLSS), where the distribution parameters are defined as a function of time. The results indicate that: (1) the trends were not significant for the whole WRB, but significant positive/negative trends were still observed in some stations, abrupt changes for consecutive wet days (CWD) mainly occurred in 1985, and the assumption of stationarity is invalid for some stations; (2) for these nonstationary extreme precipitation indices series with significant positive/negative trends, the GAMLSS models are able to capture well the temporal variations of the indices, and perform better than the stationary model. Finally, the differences between the quantiles of nonstationary and stationary models are analyzed, which highlight the importance of nonstationary modeling of precipitation extremes in the WRB.
Evidence of Half-Metallicity in Cubic PrMnO3 Perovskite
The electronic and magnetic properties of the cubic praseodymium oxides perovskites PrMnO3 were calculated using the density functional theory (DFT) with both generalized gradient approximation (GGA) and GGA+U approaches, where U is on-site Coulomb interaction correction. The results show a half-metallic ferromagnetic ground state for PrMnO3 in GGA+U approached, while semi-metallic ferromagnetic character is observed in GGA. The results obtained, make the cubic PrMnO3 a promising candidate for application in spintronics.
Estimating the Timing Interval for Malarial Indoor Residual Spraying: A Modelling Approach
Background: Indoor residual spraying (IRS) reduces vector densities and malaria transmission, however, the most effective spraying intervals for IRS have not been well established. We aim to estimate the optimal timing interval for IRS using a modeling approach. Methods: We use a generalized additive model to estimate the optimal timing interval for IRS using the predicted malaria incidence. The model is applied to post IRS cohort clinical data from children aged 0.5–10 years in selected households in Tororo, historically a high malaria transmission setting in Uganda. Six rounds of IRS were implemented in Tororo during the study period (3 rounds with bendiocarb: December 2014 to December 2015, and 3 rounds with actellic: June 2016 to July 2018). Results: Monthly incidence of malaria from October 2014 to February 2019 decreased from 3.25 to 0.0 per person-years in the children under 5 years, and 1.57 to 0.0 for 5-10 year-olds. The optimal time interval for IRS differed between bendiocarb and actellic and by IRS round. It was estimated to be 17 and 40 weeks after the first round of bendiocarb and actellic, respectively. After the third round of actellic, 36 weeks was estimated to be optimal. However, we could not estimate from the data the optimal time after the second and third rounds of bendiocarb and after the second round of actellic. Conclusion: We conclude that to sustain the effect of IRS in a high-medium transmission setting, the second rounds of bendiocarb need to be applied roughly 17 weeks and actellic 40 weeks after the first round, and the timing differs for subsequent rounds. The amount of rainfall did not influence the trend in malaria incidence after IRS, as well as the IRS timing intervals. Our results suggest that shorter intervals for the IRS application can be more effective compared to the current practice, which is about 24 weeks for bendiocarb and 48 weeks for actellic. However, when considering our findings, one should account for the cost and drug resistance associated with IRS. We also recommend that the timing and incidence should be monitored in the future to improve these estimates.
Application of MoM-GEC Method for Electromagnetic Study of Planar Microwave Structures: Shielding Application
In this paper, an electromagnetic analysis is presented for describing the influence of shielding in a rectangular waveguide. A hybridization based on the method of moments combined to the generalized equivalent circuit MoM-GEC is used to model the problem. This is validated by applying the MoM-GEC hybridization to investigate a diffraction structure. It consists of electromagnetic diffraction by an iris in a rectangular waveguide. Numerical results are shown and discussed and a comparison with FEM and Marcuvitz methods is achieved.
Matrix-Based Linear Analysis of Switched Reluctance Generator with Optimum Pole Angles Determination
In this paper, linear analysis of a Switched Reluctance Generator (SRG) model is applied on the most common configurations (4/2, 6/4 and 8/6) for both conventional short-pitched and fully-pitched designs, in order to determine the optimum stator/rotor pole angles at which the maximum output voltage is generated per unit excitation current. This study is focused on SRG analysis and design as a proposed solution for renewable energy applications, such as wind energy conversion systems. The world&rsquo;s potential to develop the renewable energy technologies through dedicated scientific researches was the motive behind this study due to its positive impact on economy and environment. In addition, the problem of rare earth metals (Permanent magnet) caused by mining limitations, banned export by top producers and environment restrictions leads to the unavailability of materials used for rotating machines manufacturing. This challenge gave authors the opportunity to study, analyze and determine the optimum design of the SRG that has the benefit to be free from permanent magnets, rotor windings, with flexible control system and compatible with any application that requires variable-speed operation. In addition, SRG has been proved to be very efficient and reliable in both low-speed or high-speed applications. Linear analysis was performed using MATLAB simulations based on the (Modified generalized matrix approach) of Switched Reluctance Machine (SRM). About 90 different pole angles combinations and excitation patterns were simulated through this study, and the optimum output results for each case were recorded and presented in detail. This procedure has been proved to be applicable for any SRG configuration, dimension and excitation pattern. The delivered results of this study provide evidence for using the 4-phase 8/6 fully pitched SRG as the main optimum configuration for the same machine dimensions at the same angular speed.
A Mathematical Equation to Calculate Stock Price of Different Growth Model
This paper presents an equation to calculate stock prices of different growth model. This equation is mathematically derived by using discounted cash flow method. It has the advantages of being very easy to use and very accurate. It can still be used even when the first stage is lengthy. This equation is more generalized because it can be used for all the three popular stock price models. It can be programmed into financial calculator or electronic spreadsheets. In addition, it can be extended to a multistage model. It is more versatile and efficient than the traditional methods.
Parameter Estimation via Metamodeling
Based on appropriate multivariate statistical methodology, we suggest a generic framework for efficient parameter estimation for ordinary differential equations and the corresponding nonlinear models. In this framework classical linear regression strategies is refined into a nonlinear regression by a locally linear modelling technique (known as metamodelling). The approach identifies those latent variables of the given model that accumulate most information about it among all approximations of the same dimension. The method is applied to several benchmark problems, in particular, to the so-called ”power-law systems”, being non-linear differential equations typically used in Biochemical System Theory.
Effects of Subsidy Reform on Consumption and Income Inequalities in Iran
In this paper, we use data on Household Income and Expenditure survey of Statistics Centre of Iran, conducted from 2005-2014, to calculate several inequality measures and to estimate the effects of Iran’s targeted subsidy reform act on consumption and income inequality. We first calculate Gini coefficients for income and consumption in order to study the relation between the two and also the effects of subsidy reform. Results show that consumption inequality has not been always mirroring changes in income inequality. However, both Gini coefficients indicate that subsidy reform caused improvement in inequality. Then we calculate Generalized Entropy Index based on consumption and income for years before and after the Subsidy Reform Act of 2010 in order to have a closer look into the changes in internal structure of inequality after subsidy reforms. We find that the improvement in income inequality is mostly caused by the decrease in inequality of lower income individuals. At the same time consumption inequality has been decreased as a result of more equal consumption in both lower and higher income groups. Moreover, the increase in Engle coefficient after the subsidy reform shows that a bigger portion of income is allocated to consumption on food which is a sign of lower living standard in general. This increase in Engle coefficient is due to rise in inflation rate and relative increase in price of food which partially is another consequence of subsidy reform. We have conducted some experiments on effect of subsidy payments and possible effects of change on distribution pattern and amount of cash subsidy payments on income inequality. Result of the effect of cash payments on income inequality shows that it leads to a definite decrease in income inequality and had a bigger share in improvement of rural areas compared to those of urban households. We also examine the possible effect of constant payments on the increasing income inequality for years after 2011. We conclude that reduction in value of payments as a result of inflation plays an important role regardless of the fact that there may be other reasons. We finally experiment with alternative allocations of transfers while keeping the total amount of cash transfers constant or make it smaller through eliminating three higher deciles from the cash payment program, the result shows that income equality would be improved significantly.
Some Results on the Generalized Higher Rank Numerical Ranges
‎In this paper, ‎the notion of ‎rank-k numerical range of rectangular complex matrix polynomials‎ ‎are introduced. ‎Some algebraic and geometrical properties are investigated. ‎Moreover, ‎for ε>0 the notion of Birkhoff-James approximate orthogonality sets for ε-higher ‎rank numerical ranges of rectangular matrix polynomials is also introduced and studied. ‎The proposed definitions yield a natural generalization of the standard higher rank numerical ranges.
Non-Differentiable Mond-Weir Type Symmetric Duality under Generalized Invexity
In the present paper, a pair of Mond-Weir type non-differentiable multiobjective second-order programming problems, involving two kernel functions, where each of the objective functions contains support function, is formulated. We prove weak, strong and converse duality theorem for the second-order symmetric dual programs under η-pseudoinvexity conditions.
Second Order Solitary Solutions to the Hodgkin-Huxley Equation
Necessary and sufficient conditions for the existence of second order solitary solutions to the Hodgkin-Huxley equation are derived in this paper. The generalized multiplicative operator of differentiation helps not only to construct closed-form solitary solutions but also automatically generates conditions of their existence in the space of the equation's parameters and initial conditions. It is demonstrated that bright, kink-type solitons and solitary solutions with singularities can exist in the Hodgkin-Huxley equation.
Parameter Estimation in Dynamical Systems Based on Latent Variables
A novel mathematical approach is suggested, which facilitates a compressed representation and efficient validation of parameter-rich ordinary differential equation models describing the dynamics of complex, especially biology-related, systems and which is based on identification of the system's latent variables. In particular, an efficient parameter estimation method for the compressed non-linear dynamical systems is developed. The method is applied to the so-called 'power-law systems' being non-linear differential equations typically used in Biochemical System Theory.
The Impact of Audit Committee Industry Expertise on Internal Audit Function
This study examines whether internal audit function is indeed greater when audit committee members have industry expertise combined with auditing expertise. Data from a survey of 64 chief internal auditors from companies registered on the Saudi Stock Exchange TADAWL, provides results that suggest that when audit committee members possess both industry expertise and auditing expertise, the committee’s role in improving the quality of internal audit is enhanced. This outcome is concluded as one that can be generalized beyond the Saudi Arabian context.
A Word-to-Vector Formulation for Word Representation
This work presents a novel word to vector representation that is based on embedding the words into a sphere, whereby the dot product of the corresponding vectors represents the similarity between any two words. Embedding the vectors into a sphere enabled us to take into consideration the antonymity between words, not only the synonymity, because of the suitability to handle the polarity nature of words. For example, a word and its antonym can be represented as a vector and its negative. Moreover, we have managed to extract an adequate vocabulary. The obtained results show that the proposed approach can capture the essence of the language, and can be generalized to estimate a correct similarity of any new pair of words.
Inverse Scattering of Two-Dimensional Objects Using an Enhancement Method
A 2D complete identification algorithm for dielectric and multiple objects immersed in air is presented. The employed technique consists of initially retrieving the shape and position of the scattering object using a linear sampling method and then determining the electric permittivity and conductivity of the scatterer using adjoint sensitivity analysis. This inversion algorithm results in high computational speed and efficiency, and it can be generalized for any scatterer structure. Also, this method is robust with respect to noise. The numerical results clearly show that this hybrid approach provides accurate reconstructions of various objects.
Structural and Electronic Properties of the Rock-salt BaxSr1−xS Alloys
Structural and electronic properties of the rock-salt BaxSr1−xS are calculated using the first-principles calculations based on the density functional theory (DFT) within the generalized gradient approximation (GGA), the local density approximation (LDA) and the virtual-crystal approximation (VCA). The calculated lattice parameters at equilibrium volume for x=0 and x=1 are in good agreement with the literature data. The BaxSr1−xS alloys are found to be an indirect band gap semiconductor. Moreoever, for the composition (x) ranging between [0-1], we think that our results are well discussed and well predicted.
The Beta-Fisher Snedecor Distribution with Applications to Cancer Remission Data
In this paper, a new four-parameter generalized version of the Fisher Snedecor distribution called Beta- F distribution is introduced. The comprehensive account of the statistical properties of the new distributions was considered. Formal expressions for the cumulative density function, moments, moment generating function and maximum likelihood estimation, as well as its Fisher information, were obtained. The flexibility of this distribution as well as its robustness using cancer remission time data was demonstrated. The new distribution can be used in most applications where the assumption underlying the use of other lifetime distributions is violated.
On Hankel Matrices Approach to Interpolation Problem in Infinite and Finite Fields
Interpolation problem, as it was initially posed in terms of polynomials, is well researched. However, further mathematical developments extended it significantly. Trigonometric interpolation is widely used in Fourier analysis, while its generalized representation as exponential interpolation is applicable to such problem of mathematical physics as modelling of Ziegler-Biersack-Littmark repulsive interatomic potentials. Formulated for finite fields, this problem arises in decoding Reed--Solomon codes. This paper shows the relation between different interpretations of the problem through the class of matrices of special structure - Hankel matrices.
Quantile Coherence Analysis: Application to Precipitation Data
The coherence analysis measures the linear time-invariant relationship between two data sets and has been studied various fields such as signal processing, engineering, and medical science. However classical coherence analysis tends to be sensitive to outliers and focuses only on mean relationship. In this paper, we generalized cross periodogram to quantile cross periodogram and provide richer inter-relationship between two data sets. This is a general version of Laplace cross periodogram. We prove its asymptotic distribution under the long range process and compare them with ordinary coherence through numerical examples. We also present real data example to confirm the usefulness of quantile coherence analysis.
Analysis of a Generalized Sharma-Tasso-Olver Equation with Variable Coefficients
Considering the inhomogeneities of media, the variable-coefficient Sharma-Tasso-Olver (STO) equation is hereby investigated with the aid of symbolic computation. A newly developed simplified bilinear method is described for the solution of considered equation. Without any constraints on the coefficient functions, multiple kink solutions are obtained. Parametric analysis is carried out in order to analyze the effects of the coefficient functions on the stabilities and propagation characteristics of the solitonic waves.
Survival Data with Incomplete Missing Categorical Covariates
The survival censored data with incomplete covariate data is a common occurrence in many studies in which the outcome is survival time. With model when the missing covariates are categorical, a useful technique for obtaining parameter estimates is the EM by the method of weights. The survival outcome for the class of generalized linear model is applied and this method requires the estimation of the parameters of the distribution of the covariates. In this paper, we propose some clinical trials with ve covariates, four of which have some missing values which clearly show that they were fully censored data.
Nullity of t-Tupple Graphs
The nullity η (G) of a graph is the occurrence of zero as an eigenvalue in its spectra. A zero-sum weighting of a graph G is real valued function, say f from vertices of G to the set of real numbers, provided that for each vertex of G the summation of the weights f (w) over all neighborhood w of v is zero for each v in G.A high zero-sum weighting of G is one that uses maximum number of non-zero independent variables. If G is graph with an end vertex, and if H is an induced sub-graph of G obtained by deleting this vertex together with the vertex adjacent to it, then, η(G)= η(H). In this paper, a high zero-sum weighting technique and the end vertex procedure are applied to evaluate the nullity of t-tupple and generalized t-tupple graphs are derived and determined for some special types of graphs. Also, we introduce and prove some important results about the t-tupple coalescence, Cartesian and Kronecker products of nut graphs.
Problems Encountered in Teaching English as a Second Language in Asia
This paper conveys some of the problems teachers of ESL face in classroom settings in Thailand. The results of this paper is achieved through close and open ended questionaires administered to a group of English language teachers of three prominent schools in Kaengkhoi, saraburi Province, Thailand.(Saengvithaya school, kaengkhoi school and Pytoon withaya school). Face to face interview of some foreign teachers and students selected randomly And general observation. The data was analysed by frequency distribution and percentage: The result of the study may be generalized so that the conference committee can suggest possible solutions or give contributing ideas on how to handle some of these problems.
Theoretical Investigation on Electronic and Magnetic Properties of Cubic PrMnO3 Perovskite
The purpose of this study was to investigate the structural,electronic and magnetic properties of the cubic praseodymium oxides perovskites PrMnO3. It includes our calculations based on the use of the density functional theory (DFT) with both generalized gradient approximation (GGA) and GGA+U approaches, The spin polarized electronic band structures and densities of states as well as the integer value of the magnetic moment of the unit cell (6 μB) illustrate that PrMnO3 is half-metallic ferromagnetic. The study prove that the compound is half-metallic ferromagnetic however the results obtained, make the cubic PrMnO3 a promising candidate for application in spintronics.
Calculated Structural and Electronic Properties of Mg and Bi
The present study shows the structural, electronic and magnetic properties of magnesium (Mg) and bismuth (Bi) in a supercell (1X1X5). For both materials were studied in five crystalline structures: rock salt (NaCl), cesium chloride (CsCl), zinc-blende (ZB), wurtzite (WZ), and nickel arsenide (NiAs), using the Density Functional Theory (DFT), the Generalized Gradient Approximation (GGA), and the Full Potential Linear Augmented Plane Wave (FP-LAPW) method. By means of fitting the Murnaghan's state equation we determine the lattice constant, the bulk modulus and it's derived with the pressure. Also we calculated the density of states (DOS) and the band structure.