- Research
- Open Access

# A new gene regulatory network model based on BP algorithm for interrogating differentially expressed genes of *Sea Urchin*

- Longlong Liu
^{1}, - Tingting Zhao
^{1}, - Meng Ma
^{1}and - Yan Wang
^{2}Email author

**Received:**25 May 2016**Accepted:**12 October 2016**Published:**3 November 2016

## Abstract

### Background

Computer science and mathematical theories are combined to analyze the complex interactions among genes, which are simplified to a network to establish a theoretical model for the analysis of the structure, module and dynamic properties. In contrast, traditional model of gene regulatory networks often lack an effective method for solving gene expression data because of high durational and spatial complexity. In this paper, we propose a new model for constructing gene regulatory networks using back propagation (BP) neural network based on predictive function and network topology.

### Results

Combined with complex nonlinear mapping and self-learning, the BP neural network was mapped into a complex network. Network characteristics were obtained from the parameters of the average path length, average clustering coefficient, average degree, modularity, and map’s density to simulate the real gene network by an artificial network. Through the statistical analysis and comparison of network parameters of *Sea Urchin* mRNA microarray data under different temperatures, the value of network parameters was observed. Differentially expressed *Sea Urchin* genes associated with temperature were determined by calculating the difference in the degree of each gene from different networks.

### Conclusion

The new model we developed is suitable to simulate gene regulatory network and has capability of determining differentially expressed genes.

## Keywords

- BP algorithm
- Gene regulatory network
- Neural network model
- Differentially expressed
*Sea Urchin*genes

## Background

Complex life phenomenon is the effect and regulatory mechanism of a large number of genes. To date, studies on complex biological systems have shifted from the local description of individual gene functions to quantitative analysis of complex gene regulatory networks (Plahte et al. 2013; Ahmad et al. 2012). Computer science and mathematical theory are combined to analyze the complex interactions among genes, which are simplified to a network to establish a theory model for the analysis of the structure, module and dynamic properties of a gene regulatory network (Smart et al. 2008; Patrik D’haeseleer SLaRS 1999; Raza and Parveen 2013).

From 2000, when the first published in nature on the topological properties of biological networks based on complex network theory to now, there is a huge development and progress achieved in the gene network investigation (Stifanelli et al. 2013; Bowers et al. 2004; Araki et al. 2013; Raza and Parveen 2013). Complex networks including the construction and simulation of a gene regulatory network are widely used in biological networks. Evidence from past data collection and statistical analysis of large-scale gene network highlights the compatibility of the structural characteristics of gene regulatory networks with other complex network system. Focusing solely on the network topology of a complicated network system is no longer sufficient in the process of constructing artificial gene networks.

The artificial network simulates the real gene network though network characteristics such as the average path length, clustering coefficient, average degree, modularity, map density and et al. (Thurner 2009; Raza 2016). It has been developed a variety of models and algorithms to simulate the gene regulatory network (GRN) mainly including the Boolean (Lähdesmäki 2003; Faure et al. 2006; Kim et al. 2007; Stolovitzky et al. 2008; Politano et al. 2014; Comar et al. 2015), Bayesian (Perrin et al. 2003; Husmeier 2003; Friedman et al. 2000; Bansal et al. 2006; Chai et al. 2014; Lo et al. 2015), linear differential equation (Chen et al. 1999; de Jong and Ropers 2006; van Someren et al. 2000), relevance (Butte and Kohane 2000; Runcie et al. 2012; Parmigiani et al. 2003) and neural network model (Vohradsky 2001; Rui et al. 2007; Raza and Alam 2016). However, traditional models of gene regulatory networks often lack an effective method of solving the gene expression profiling data because of high time and spatial complexity.

An artificial neural network (ANN) usually dubbed as “neural network” (the term we adopted and defined in this paper), is a computational model originally intended to simulate the structural and/or function of biological neural networks (Marshall 1995). And it exhibited powerful modeling ability and yielded significant results in terms of network structure, training algorithm, approximation performance, and stability (Aussem 1999; Mak et al. 1999). The use of recurrent neural network for constructing a gene regulatory network has achieved much better results than traditional models. However, the complexity of the recurrent neural network models makes it difficult and unsuitable for analysis of biologically significant gene regulatory relationships based on high-throughput microarray or sequencing data. Back propagation (BP) network as a kind of developed ANN is a multi-layered feed forward networks, in which the propagation is forward, error spreads reversely makes it faster and more powerful when used to model the high-throughput microarray or sequencing data than using the recurrent neural network algorithms.

Recently, reverse network model was developed as a suitable analysis for high-throughput data (such as microarray and high-throughput sequencing data) to mine regulatory mechanisms among the components of a system and has been extensively applied to examine various biological systems (Raza and Alam 2016; Werhli et al. 2006; Wang et al. 2010; Perkins et al. 2006). For increasing the accuracy of simulating GRN, we the first time mapped BP algorithm neural network based on sigmoid function into a common complex network with the microarray expression data. And thought the network parameters, the differential genes were determined. Rest of the paper is organized as follows. In method part, BP network was described briefly and the genes networks based on BP ANN was built. The result part is model application and comparison. Then discussions were presented and finally paper was concluded.

## Methods

Reverse network model is built based on BP network combined with complex nonlinear mapping and self-learning. An artificial network is simulated the real gene network according to the network characteristics: the average path length, average clustering coefficient, average degree, modularity, map’s density and et al.

### Structure and algorithm of the BP neural network

*w*
_{
kn
}: The weight from neuron *n* to neuron *k*; *o*
_{
n
}: the output of neuron *n* or of the nth input: *b*
_{
k
}: the neuron threshold.

*E*

^{ p }of a given the pth pattern is calculated as

*t*
_{
j
}
^{
p
}
: the pth desired output value; *o*
_{
j
}
^{
p
}
: the output of the corresponding neuron.

- a.
Initialization: small random values are taken for weights of each layer and thresholds of each neuron, the max cycle times and the min whole error are set as

*m*and*ɛ*respectively; - b.
Vector \(X^{{^{p} }} = \left[ {x_{0}^{P} ,x_{1}^{P} , \ldots ,x_{n - 1}^{P} } \right]\) is the inputted data pattern, where P means the Pth pattern, and

*n*is the number of neurons of initial layer,*x*_{ i }^{ P }is the input of the given hidden layer; - c.
The actual output of hidden layer \(O_{j}^{p} = f\left( {\sum\nolimits_{j = 0}^{n - 1} {w_{ij} x_{ji} - B_{i} } } \right)\) is calculated and regarded as an input to the next layer,

*f*is the activation function; - d.
If the layer is the last layer (output layer), the actual error

*E*^{ P }is calculated as \(E^{P} = \sum _{j} \left( {t_{j}^{P} - o_{j}^{P} } \right)^{ 2}\), otherwise the error calculation as (c); - e.
The whole error

*E*is calculated as \(E = 1/ 2 \times \sum\limits_{{_{P} }} {E^{P} }\); - f.Weights are adjusted from the last layer and going backwards (BP),$$W_{ji} ({\text{new}}) = W_{ji} ({\text{old}}) + \eta \times \delta_{j}^{p} o_{j}^{p} +\upalpha \times \text{(}W_{ji} ({\text{new}}) - W_{ji} ({\text{old}})\text{)}$$(4)
where

*η*(0 <*η*< 1) and*α*(0 <*α*< 1) are constants named learning rate and momentum, respectively;*η*measures the influence degree of the error;*α*determines the influence of the weight change.When neuron

*j*is output layer neuron and hidden layer neuron, the error term for pattern*p*is \(\delta_{j}^{p} = f^{\prime }\left( {{\text{o}}_{j}^{p} } \right)\left( {{\text{t}}_{j}^{p} - {\text{o}}_{j}^{p} } \right)\) and \(\delta_{j}^{p} = f^{\prime }\left( {{\text{o}}_{j}^{p} } \right)\left( {\sum\nolimits_{k} {W_{kj} \times \delta_{k}^{p} } } \right)\) respectively; - g.
If the cycle time is

*m*or the whole error is less than*ɛ*, train is over, otherwise go to (b).

The weight adjusting in the above algorithm is aimed at minimizing the whole error *E*, which is performed with the gradient descent via weight changing to make the error steepest down (Bishop 1996). The *η* term is a measure of the influence degree for updating weights in the formula, whereas the *α* term determines the influence of the past history of weight changes in the same formula. The single-layer neural network structure is a two-layer network structure with input and output layer. It facilitates more easily the mapping from the trained BP network to the gene regulatory network, so we will use the single-layer BP network in this paper.

### Establishment of genes networks based on BP ANN

*E*:

*A*:

where *a*
_{
ij
} and *a*
_{
ji
} in *A* represents the regulation between gene *i* and *j*. Assuming that the state changes of genes in a real regulatory network mainly depend on the effect of other genes, the self-regulation of a gene can be ignored. Then, the diagonal elements in *A* are 0, i.e. *a*
_{
ij
} = *0*. Two kinds of regulatory relationships exist between genes, activation and inhibition which indicated by the negative or positive value of *a*
_{
ij
}.

*(n*− 1

*)*− 1 structure is adopted to construct the model (Fig. 3a). The

*n*− 1 neurons in the input layer are used as the temporary storage for the input data corresponding to the

*n*− 1 genes. The neuron in the output layer is corresponding to the

*n*th genes. Self-correlation is not considered in this model which means network weight

*w*

_{ ii }doesn

^{’}t exist. So, the input samples for the network are \(X^{p} = x_{1}^{p} , \cdots ,x_{n - 1}^{p}\), and the target sample is \(T^{p} = x_{1}^{p}\). The training samples are {

*X*

^{ p }, T

^{ p }}, i.e., the input samples are the same as the target samples. The

*p*training samples are inputted to the network to train the network until the error or the operating cycle reaches the set value. Therefore, a weight vector

*W*

_{ i }can be obtained. This process is reiterated form

*x*

_{1}to

*x*

_{ n }, and then, a weight matrix

*W*is obtained.

*E*< 10

^{−2}],

*o*

_{i ≈}

*x*

_{ i }can be considered; That is, any gene among the

*n*genes can become a sigmoid function of the linear combination of the other

*n*− 1 genes.

Second, the trained BP neural network is mapped into a gene regulatory network. The trained neural network is mapped into a directed gene regulatory network (Fig. 3b). *A* = *W*, weight \(w_{ij} \left( {{\text{i}}\;{ \ne }\;{\text{j}}} \right)\) denotes the edge weight from neuron *i* to neuron *j* in the gene regulatory network, whereas \(w_{ji} \left( {{\text{i}}\,{ \ne }\,{\text{j}}} \right)\) denotes the edge weight from neuron *j* to neuron *i*.

Third, a reasonable weight threshold is selected to choose the highly relevant genes and finally determine the gene regulatory network. The process of choosing threshold is very important. The threshold can be set according to the weight ratio \(\lambda = {{\left| {w_{ij} } \right|} \mathord{\left/ {\vphantom {{\left| {w_{ij} } \right|} {\sum\nolimits_{i} {\left| {w_{ij} } \right|} }}} \right. \kern-0pt} {\sum\nolimits_{i} {\left| {w_{ij} } \right|} }}\) or by other methods. In the regulatory network, the edges with weight ratio which is less than the threshold are removed to obtain the final gene regulatory network. The weight ratios corresponding to weight *w*
_{1i} and *w*
_{21} are assumed to be less than the threshold and are deleted, and then the corresponding edges in the regulatory network are also removed (Fig. 3c).

### Structural parameters of network

*G*=

*(V, E)*is assumed to be a complex network with node set

*V*=

*{1,2,…N}*and edge set

*E*. The parameters are determined according to the network statistics as following:

- 1.
Average path length (

*L*)

*d*

_{ ij }between nodes

*i*and

*j*in

*V*is defined as the minimal number of edges connecting nodes

*i*and

*j*. The average path length of the network is defined as \(L = \sum\nolimits_{i > j} {{{d_{ij} }/ {\left( {0.5{\text{N(N}} - 1)} \right)}}}\).

- 2.
Average clustering coefficient (

*C*)

*C*

_{ i }of node

*I*is the ratio of the number of actually existing edges to that of possible existing edges among the adjacent nodes of

*i*. The clustering coefficient of the network

*C*is the average of the clustering coefficients of all nodes.

- 3.
Average degree (

*K*)

*K*of the network is the average of degrees of all nodes.

- 4.
Modularity (

*Q*)

Network *G* is supposed to contain *k* communities as G_{1}, G_{2}, …G_{k}. A symmetric matrix *H* = (h_{
ij
})_{
k × k
} is defined, where *h*
_{
ij
} denotes the ratio of the number of edges between two communities *G*
_{
i
} and *G*
_{
j
} to the number of total edges of the network.

*α*

_{ i }denotes the sum of elements in the

*i*th row of matrix

*H*, which represents the ratio of the number of edges connecting to community

*G*

_{ i }to the number of total edges.

- 5.
Density of map (

*D*)

Density of map is the ratio of the total path length to the area of the map.

## Results

In this section, we applied the network model on the microarray data to determine the differentially expressed genes and to assess how the model works. We chose the *sea urchin* (*Strongylocentrotus purpuratus*) mRNA microarray data in the GPL13644 platform from NCBI database in our analysis (http://www.ncbi.nlm.nih.gov/geo/query/acc.cgi?acc=GPL13644).

### Data sources

In this experiment, mRNA microarray data of *Sea Urchin* was used to investigate if the gene expression responses characterize molecular signatures of temperature stress, and as a result to know how stress responses alter gene expression. There were totally 191 samples divided into three groups according to the different temperature: 64 samples at 12 °C (T12), 63 samples at 15 °C (T15), and 64 samples at 18 °C (T18), respectively (Runcie et al. 2012). We applied the microarray data, 336 transcripts totally, to the network model to build gene networks and to analyze what genes responding to different growing temperature stress are significantly differential expressed.

### Data processing

- 1.
If \(x_{\rm min}\,{\ne}\,x_{\rm max}\), then \(x^\prime\,=\,\frac{x\,-\,x_{\rm min}}{{x_{\rm max}\,-\,x_{\rm min}}}\);

- 2.
If \(x_{\rm min} = x_{\rm max}\), then \(x^{\prime } = x_{\rm min}\).

where, *x* represents the element of each sample. *x*
_{min} and *x*
_{max} represent the minimum and maximum of all the samples elements, respectively, and *x*′ represents the normalized sample element.

To further reduce the noise from the different experimental conditions of the samples, the samples of each group were divided into two different blocks and the mean value of each sample was computed. So, a total of 32, 31, and 32 samples were observed in the T12, T15, and T18 groups.

### Establishment of gene networks

The model of a single-layer feed forward neural network with the structure of 335-1 was shown in Fig. 3a. The 335 neurons in the input layer (temporary storage) correspond to the data of 335 genes, whereas the neuron in the output layer corresponds to the data of another gene of the 335 genes. Therefore, the input sample of the network is \(X^{p} = x_{1}^{p} , \ldots ,x_{335}^{p}\), the target sample is \(T^{p} = x_{i}^{p}\), and the training sample is {*X*
^{
p
}, T^{
p
}}. The transfer function of neurons is sigmoid function with learning rate 0.7 and the threshold −1.

The parameters of 3 networks constructed based on samples of T12, T15 and T18 group, respectively

Samples | Average path length ( | Average clustering coefficient ( | Average degree ( | Modularity ( | Density of map( |
---|---|---|---|---|---|

T12 | 3.74 | 0.082 | 6.875 | 0.273 | 0.021 |

T15 | 3.833 | 0.049 | 5.765 | 0.287 | 0.017 |

T18 | 3.256 | 0.109 | 7.426 | 0.264 | 0.022 |

### Comparison of gene networks

Parameters of relevance network with correlation coefficient 0.85

Samples | Average path length ( | Average clustering coefficient ( | Average degree ( | Modularity ( | Density of map ( |
---|---|---|---|---|---|

T12 | 1 | 0 | 0.006 | 0 | 0 |

T15 | 1.167 | 0 | 0.03 | 0.72 | 0 |

T18 | 1 | 0 | 0.024 | 0.75 | 0 |

### Differentially expressed genes determination

*η*was introduced and defined as degree difference ratio:

where *T*
_{i} and *T*
_{0} denote the input and output degrees of a gene in the network of one group; \(T_{i}^{{\prime }}\) and \(T_{0}^{{\prime }}\) denote the input and output degrees of a gene in the network of another group.

*η*of each gene is calculated using samples from T15 and T12 group. The genes are ranked according to

*η*in descending order. And the same process is performed using samples from T18 and T15 groups. In this experiment, 13 differentially expressed genes are found (Table 3). At the same time, we calculate the significantly differential expressed genes (DEGs) using the same data after normalization and outlier removal. The significantly differential genes are defined as |log2FC| ≥ 0.05 and p value <0.05 by z-test. There are 22 DEGs in common between T12 and T18 compared with T15 (Two groups of differential genes were calculated between T12 and T15, and between T18 and T15, the same as network.). Comparison is carried on between DEGs calculated by BP algorithm and by fold change (Fig. 5f). There were 7 DEGs (

*APOBEC,FoxG,FoxO,gataC,Gsk*-

*3,OTX,SM30*-

*E*) significantly overlapped indicating that the BP network model we build has capability of finding a large part of the significantly differential genes based on the high-throughput sequencing data or microarray data determined by experimental method.

List of differentially expressed genes

ID | ORF | SEQUENCE |
---|---|---|

1776 | APOBEC | ATAAGAACCAGTGGGGCCCACCCAGTTTCACCCTCCTCTCTCAT |

5123 | Otp | ACCCGCATCGCAATCTCCTCCCGCATGAAGATATCAGGATAGT |

2789 | Gsk-3 | GTCCTAGGAACCCCAAGCCGTGACCAGATCAAGGAGATGAAC |

1078 | Nk1 | GCCATCATCACCCGACCCAACTGCAGCAGCTATTCATACAT |

6021 | gataC | TAGTTCAGCACCTCATCCCGGTCCAACAAGTTCCTACACGTTACC |

3972 | FoxG | TCATGATGGCTATTCGCTCGAGTCCAGAGAAAAGACTAACTCTAAATG |

4071 | G-cadherin | GTGCGAGGAGACCAGCCTTTCCATCGAGTTCATCACAGAGACTC |

5397 | Blimp1-Krox | ACCTATGTATGGCCTGTCACCAAACTACATCAGTACTGCAGGTGGT |

3498 | FoxO | CGATCATGACCACACATCCAGAAATCGACATGCATGACAATGAAGTC |

265 | APOBEC | ACAACAGCTCCTCCCCTCACCCCTACCAGTCAGGCTACCACC |

2208 | SM30-E | TCAACCTGGTTTTGGACAACCCGGTGTTGGTCAACCCAATAGA |

5073 | otx | GCACTTTCTGATCTTGCTAGTCGTGAAATCAAGATGGAATCACATTCT |

1987 | P16 | AAGTGATGACGACGGCAGCAGCGATGATGACGGTAGCAGTGAT |

## Discussion

To date, most gene regulatory networks are small networks for hundreds of genes. Traditional models of gene regulatory networks often lack an effective method of solving gene expression profiling data because of high time and space complexity. Based on predictive function construction and network topology, a new model for constructing gene regulatory networks using a BP neural network was tested in this paper. Combined with complex nonlinear mapping and self-learning, the BP neural network was mapped into a complex network. Since ANN can easily implement parallel processing, building a large-scale gene regulatory network model with different layers and modules is possible. Concretely, the internal characteristics and operation mechanism of the function modules of the network should be investigated. And the function and robustness under the outside interference of a sub-network should be discussed according to the classification of a regulatory network structure as well.

Mathematical theory has shown that multilayer feed forward BP networks can carry on any complex nonlinear functions, making it particularly suitable for solving problems with complex internal mechanism. BP networks have the ability of self-learning and generalization, but are also limited by slow learning speed, difficulty in determining the number of hidden layer nodes, and falling into local minima. In this study, we adopt a single-layer network structure, in which there is no hidden layer. Thus, selection of initial values of the network parameters is performed more frequently and the local minima are more likely avoided. Without changing the network structure, the data is added to the training set directly. With the training of the neural networks, network weights and the mapped gene regulatory networks are changed.

Through statistical analysis and comparison of differential genes based on the mRNA microarray data from *Sea Urchin* growing in different temperatures, parameters of diverse average degrees, average path lengths, modularity, average clustering coefficients, and map densities were obtained. Differentially expressed *Sea Urchin* genes associated with temperature were determined by calculating the difference in the degree of each gene from different networks. To check the effectiveness of BP network, comparison of the parameters with the common relevance network based on Pearson’s coefficient and significantly overlapped differential genes showed that the parameters of BP network were more efficient to build gene regulatory network. The remain un-overlapped genes reminded us that the gene regulatory network built based on BP network still need to improve maybe though improving some algorithm.

Besides, the convergence of a network is important that reducing, maintaining, or increasing the training error of the network within a specific controlled range allows the retention of newly added samples in the training set by the convergence of a network and should be ensured. If the error is large, the sample is regarded as a singular point and cannot be retained in the training set. Therefore, the dynamic property and stability of the network should be guaranteed.

## Conclusion

In this paper, we developed a new model for constructing gene regulatory networks based on back propagation neural network. The application of the new model to the mRNA microarray data and the comparison with the common reverse network and differential genes indicated that the new model is suitable to simulate gene regulatory network and has capability of determining differentially expressed genes.

## Declarations

### Authors’ contributions

LLL and TTZ developed the idea for the study. MM, LLL and YW designed the algorithm and network, did the literature review and prepared the manuscript. MM and TTZ helped to revise the manuscript. All authors read and approved the final manuscript.

### Acknowledgements

The authors would like to thank all of the researchers who made publicly available data used in this study.

### Competing interests

The authors declare that they have no competing interests.

### Availability of supporting data

*Sea Urchin* (*Strongylocentrotus purpuratus*) mRNA microarray data used in the analysis was downloaded from NCBI database (http://www.ncbi.nlm.nih.gov/geo/query/acc.cgi?acc=GPL13644).

### Funding

This work was supported in part by the National Natural Science Foundation of China (No: 61303145 and No: 61401459) and by the University Basic Research Foundation (No: 201362031).

**Open Access**This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

## Authors’ Affiliations

## References

- Ahmad FK, Deris S, Othman NH (2012) The inference of breast cancer metastasis through gene regulatory networks. J Biomed Inform 45:350–362. doi:10.1016/j.jbi.2011.11.015 View ArticlePubMedGoogle Scholar
- Araki R, Seno S, Takenaka Y, Matsuda H (2013) An estimation method for a cellular-state-specific gene regulatory network along tree-structured gene expression profiles. Gene 518:17–25. doi:10.1016/j.gene.2012.11.090 View ArticlePubMedGoogle Scholar
- Aussem A (1999) Dynamical recurrent neural networks towards prediction and modeling of dynamical systems. Neurocomputing 28:207–232. doi:10.1016/s0925-2312(98)00125-8 View ArticleGoogle Scholar
- Bansal M, Gatta GD, di Bernardo D (2006) Inference of gene regulatory networks and compound mode of action from time course gene expression profiles. Bioinformatics 22:815–822. doi:10.1093/bioinformatics/btl003 View ArticlePubMedGoogle Scholar
- Bishop CM (1996) Neural networks for pattern recognition. Clarendon Press, OxfordMATHGoogle Scholar
- Bowers PM, Cokus SJ, Eisenberg D, Yeates TO (2004) Use of logic relationships to decipher protein network organization. Science 306:2246–2249. doi:10.1126/science.1103330 ADSView ArticlePubMedGoogle Scholar
- Butte AJ, Kohane IS (2000) Mutual information relevance networks: functional genomic clustering using pairwise entropy measurements. Pac Symp Biocomput 5:418–429Google Scholar
- Cao F, Wang D, Zhu H, Wang Y (2016) An iterative learning algorithm for feedforward neural networks with random weights. Inf Sci 328:546–557. doi:10.1016/j.ins.2015.09.002 View ArticleGoogle Scholar
- Chai L, Mohamad M, Deris S, Chong C, Choon Y, Omatu S (2014) Current development and review of dynamic Bayesian network-based methods for inferring gene regulatory networks from gene expression data. Curr Bioinform 9:531–539. doi:10.2174/1574893609666140421210333 View ArticleGoogle Scholar
- Chen T, He HL, Church GM (1999) Modeling gene expression with differential equations. Pac Symp Biocomput 4:29–40Google Scholar
- Comar TD, Hegazy M, Henderson M, Hrozencik D (2015) A comparison of the Boolean and continuous dynamics of three-gene regulatory networks. Lett Biomath 1:51–65. doi:10.1080/23737867.2014.11414470 View ArticleGoogle Scholar
- D’haeseleer P, Liang S, Somogyi R (1999) Gene expression data analysis and modeling. In: Pacific symposium on biocomputing, Hawaii, 4–9 January 1999Google Scholar
- de Jong H, Ropers D (2006) Qualitative approaches to the analysis of genetic regulatory networks. Syst Model Cell Biol. doi:10.7551/mitpress/9780262195485.003.0007 Google Scholar
- Faure A, Naldi A, Chaouiya C, Thieffry D (2006) Dynamical analysis of a generic Boolean model for the control of the mammalian cell cycle. Bioinformatics 22:e124–e131. doi:10.1093/bioinformatics/btl210 View ArticlePubMedGoogle Scholar
- Friedman N, Linial M, Nachman I, Pe’er D (2000) Using Bayesian networks to analyze expression data. J Comput Biol. doi:10.1145/332306.332355 Google Scholar
- Husmeier D (2003) Reverse engineering of genetic networks with Bayesian networks. Biochem Soc Trans 31:1516–1518. doi:10.1042/bst0311516 View ArticlePubMedGoogle Scholar
- Kim H, Lee JK, Park T (2007) Boolean networks using the Chi-square test for inferring large-scale gene regulatory networks. BMC Bioinformatics 8:37. doi:10.1186/1471-2105-8-37 View ArticlePubMedPubMed CentralGoogle Scholar
- Lähdesmäki H (2003) On learning gene regulatory networks under the Boolean network model. Mach Learn 52:147–167. doi:10.1023/a:1023905711304 View ArticleMATHGoogle Scholar
- Liu J, Jin X, Dong F, He L, Liu H (2016) Fading channel modelling using single-hidden layer feedforward neural networks. Multidimens Syst Signal Process. doi:10.1007/s11045-015-0380-1 Google Scholar
- Lo L-Y, Wong M-L, Lee K-H, Leung K-S (2015) High-order dynamic Bayesian network learning with hidden common causes for causal gene regulatory network. BMC Bioinformatics. doi:10.1186/s12859-015-0823-6 Google Scholar
- Mak MW, Ku KW, Lu YL (1999) On the improvement of the real time recurrent learning algorithm for recurrent neural networks. Neurocomputing 24:13–36. doi:10.1016/s0925-2312(98)00089-7 View ArticleMATHGoogle Scholar
- Marshall JA (1995) Neural networks for pattern recognition. Neural Netw 8:493–494. doi:10.1016/0893-6080(95)90002-0 View ArticleGoogle Scholar
- Parmigiani G, Garrett ES, Irizarry RA, Zeger SL (2003) The analysis of gene expression data. InThe Anal Gene Expr Data. doi:10.1007/b97411 MathSciNetView ArticleMATHGoogle Scholar
- Perkins TJ, Jaeger J, Reinitz J, Glass L (2006) Reverse engineering the gap gene network of Drosophila melanogaster. PLoS Comput Biol 2:e51. doi:10.1371/journal.pcbi.0020051 ADSView ArticlePubMedPubMed CentralGoogle Scholar
- Perrin BE, Ralaivola L, Mazurie A, Bottani S, Mallet J, d’Alche-Buc F (2003) Gene networks inference using dynamic Bayesian networks. Bioinformatics 19:ii138–ii148. doi:10.1093/bioinformatics/btg1071 View ArticlePubMedGoogle Scholar
- Plahte E, Gjuvsland AB, Omholt SW (2013) Propagation of genetic variation in gene regulatory networks. Physica D 256–257:7–20. doi:10.1016/j.physd.2013.04.002 MathSciNetView ArticlePubMedPubMed CentralGoogle Scholar
- Politano G, Savino A, Benso A, Di Carlo S, Ur Rehman H, Vasciaveo A (2014) Using Boolean networks to model post-transcriptional regulation in gene regulatory networks. J Comput Sci 5:332–344. doi:10.1016/j.jocs.2013.10.005 View ArticleGoogle Scholar
- Rampone S, Valente A (2011) Neural network aided evaluation of landslide susceptibility in Southern Italy. Int J Mod Phys C. doi:10.1142/s0129183111016993 Google Scholar
- Raza K (2016) Reconstruction, topological and gene ontology enrichment analysis of cancerous gene regulatory network modules. Curr Bioinform 11:243–258. doi:10.2174/1574893611666160115212806 View ArticleGoogle Scholar
- Raza K, Alam M (2016) Recurrent neural network based hybrid model for reconstructing gene regulatory network. Comput Biol Chem 64:322–334. doi:10.1016/j.compbiolchem.2016.08.002 MathSciNetView ArticlePubMedGoogle Scholar
- Raza K, Parveen R (2013a) Reconstruction of gene regulatory network of colon cancer using information theoretic approach. 9.06–9.06. doi:10.1049/cp.2013.2357 Google Scholar
- Raza K, Parveen R (2013b) Soft computing approach for modeling genetic regulatory networks. Adv Comput Inform Technol 178:1–11. doi:10.1007/978-3-642-31600-5_1 View ArticleGoogle Scholar
- Rui X, Wunsch DC, Frank RL (2007) Inference of genetic regulatory networks with recurrent neural network models using particle swarm optimization. IEEE/ACM Trans Comput Biol Bioinf 4:681–692. doi:10.1109/tcbb.2007.1057 View ArticleGoogle Scholar
- Runcie DE, Garfield DA, Babbitt CC, Wygoda JA, Mukherjee S, Wray GA (2012) Genetics of gene expression responses to temperature stress in a sea urchin gene network. Mol Ecol 21:4547–4562. doi:10.1111/j.1365-294X.2012.05717.x View ArticlePubMedGoogle Scholar
- Smart AG, Amaral LA, Ottino JM (2008) Cascading failure and robustness in metabolic networks. Proc Natl Acad Sci USA 105:13223–13228. doi:10.1073/pnas.0803571105 ADSView ArticlePubMedPubMed CentralGoogle Scholar
- Stifanelli PF, Creanza TM, Anglani R, Liuzzi VC, Mukherjee S, Schena FP, Ancona N (2013) A comparative study of covariance selection models for the inference of gene regulatory networks. J Biomed Inform 46:894–904. doi:10.1016/j.jbi.2013.07.002 View ArticlePubMedGoogle Scholar
- Stolovitzky G, Davidich MI, Bornholdt S (2008) Boolean network model predicts cell cycle sequence of fission yeast. PLoS ONE 3:e1672. doi:10.1371/journal.pone.0001672 ADSView ArticleGoogle Scholar
- Sukumar N, Krein MP, Embrechts MJ (2012) Predictive cheminformatics in drug discovery: statistical modeling for analysis of micro-array and gene expression data. Meth Mol Biol 910:165–194. doi:10.1007/978-1-61779-965-5_9 View ArticleGoogle Scholar
- Thurner S (2009) Statistical mechanics of complex networks. Anal Complex Netw. doi:10.1002/9783527627981.ch2 MATHGoogle Scholar
- van Someren EP, Wessels LF, Reinders MJ (2000) Linear modeling of genetic networks from experimental data. Proc Int Conf Intell Syst Mol Biol 8:355–366PubMedGoogle Scholar
- Vohradsky J (2001) Neural network model of gene expression. FASEB J 15:846–854. doi:10.1096/fj.00-0361com View ArticlePubMedGoogle Scholar
- Wang S, Chen Y, Wang Q, Li E, Su Y, Meng D (2010) Analysis for gene networks based on logic relationships. J Syst Sci Complexity 23:999–1011. doi:10.1007/s11424-010-0205-0 MathSciNetView ArticleMATHGoogle Scholar
- Werhli AV, Grzegorczyk M, Husmeier D (2006) Comparative evaluation of reverse engineering gene regulatory networks with relevance networks, graphical gaussian models and bayesian networks. Bioinformatics 22:2523–2531. doi:10.1093/bioinformatics/btl391 View ArticlePubMedGoogle Scholar