Technika Wysokich Napi Plik Technika wysokich napiec Z. Flisowski.pdf na koncie u. Laborka Technika Wysokich Napi. Technika Wysokich Napiec Pdf File Pobierz zespoly napiec pdf. Wysokich napiec, WPL [3] S. Szpor: Technika wysokich napiec, WNT Warszawa [4] Z. Flisowski: Technika wysokich napiec, WNT Warszawa [5] Z. Gacek: Wysokonapieciowa technika izolacyjna, WPS Gliwice [6] Khajavi M., Zenger W., Desing and commissioning of a 230 kV cross linked [7] polyethylene insulated cable system, JICABLE 2003,paper.
Abstract
This paper presents an algorithm for solving N-equations of N-unknowns. This algorithm allows to determine the solution in a situation where coefficients Ai in equations are burdened with measurement errors. For some values of Ai (where i = 1,…, N), there is no inverse function of input equations. In this case, it is impossible to determine the solution of equations of classical methods.
Introduction
Mathematical models that describe electric dependencies in the receiver tested are built from discrete components. For a full description of such a model, it is required to identify the parameters xi of these elements (see Eq. 1). Most frequently, this identification is carried out indirectly through the measurements of electrical quantities Ai on the object tested [5, 8]. The parameters sought are determined from the mathematical relations (1) that describe the object.
where fi certain functions depending on the model, Ai the values measured, yi parameters that describe the model.
The classic method to solve Eq. (1) consists in determining inverse functions (2). Measurement inaccuracies that are contained in Ai are transferred to parameters yi to be determined.
In some cases, the determination of Eq. (2) may not be possible [14, 16]. This means that for adopted coefficients Ai, there are no inverse functions gi. Eq. (2) are determined for the values of environment Ai and which contain measurement errors. In this case, approximate solutions are sought which satisfy Relation (3).
The solution will be close to coefficients Ai.
An example of a model for identification
The analysis covered a single phase on an induction motor. The purpose of the analysis is to determine current–voltage dependences on the terminals of one motor phase. These relationships can be determined from the model that consists of serially connected elements: Rs, Ls i es (Fig. 1).
Electric model of the motor and the power source, where R internal resistance of sources
Coefficients RS, LS, Em, φes, α that are being sought represent many of the phenomena that occur in the motor and the system that is driven. For example, the inertia of the rotor and the system driven will affect es, and the angular velocity will exert an influence on mutual inductances, which are described with LS. When searching for the parameters of the model, the fact is also important that these factors cannot be determined with the engine being stopped. This means that the RS does not reflect the winding resistance and LS does not reflect their inductance. The parameters of the model are defined for a constant load on the machine shaft and for constant rotations. When changing the load, the parameters of the model change, as well.
In this situation, the parameters that are being determined cannot in any way be unified. They should be determined for a specific drive train (the motor and the machine driven). These parameters can vary considerably for the same engine with different mechanical properties of the system driven.
The identification of the model consists in searching for Em, φes, RS, and LS. These parameters can be determined on the receiver [6, 7, 9–11] by making measurements in the steady state (in the case of an induction motor: during operation with a constant load and a constant speed) in the system as shown below (Fig. 2):
Measuring circuit for parametric identification
According to the model adopted, we know that:
In the field of complex numbers, the following can be written:
Cygan Elektryk Wysokich Napiec
For one mesh, the voltage equation is as follows:
where
Next, by transforming (7), we determine current Ia:
Voltmeter V measures the difference in the supply voltage and in the voltage drop across internal resistance R. Thus, in the field of complex numbers, there will be the following:
From Eqs. (8) and (9), one can obtain the following:
Knowing that the forces and the current are equal, respectively:
We obtain the following equations:
Equation (11) is consistent with (1). The coefficients of the model of the receiver that are obtained from the above equations are not determinable for all the input parameters (UV, Ia, PW, QW). There are those areas that result from measurement inaccuracies where the system of Eq. (11) has no solutions.
It was found that these coefficients cannot be determined using the Newton’s interpolation algorithm [15, 16]. There are no functions that are inverse to Eq. (11), either.
Process time constant 1/α that is being sought, and which is mainly related to the inertia of the rotor and the system driven, can be determined experimentally by observing the course of voltage versus time at the motor terminals immediately after commutation.
In [13], the authors proved that amplitude ES can be equal to amplitude U. In this paper, it was also observed that frequency ES is similar to the frequency of the mains voltage. It was also noted that phase shift φes is equal to 0.
In this model, it is assumed that the frequencies of both sources are identical. This assumption does not substantially affect the results of further simulations.
Construction of an artificial neural network
Coefficients Em, RS, and LS can be determined from Eq. (11) using a neural network. The network input parameters x1 = Uv [V], x2 = Ia [A], x3 = Pw [W], x4 = Qw [VAr] contain measurement errors. Due to the nature of the adopted activation function [1–3], the output neuron of the output layer must be within range . The initial values were as follows: y1 = Rs [kΩ], y2 = Ls [H], y3 = Es [kV].
Training of the network must be for those learning vectors that do not contain any measurement errors. Learning vectors are constructed from Eqs. (1) or (11) for random values y1, y2, y3 that lie within the set of permissible changes, and which is limited with values a and b [4, 12].
The test vector is built according to Fig. 3 for values yi that are not contained within the training set.
Construction of learning vectors
The neural network was built in a VBA environment in EXCEL.
The script associated with the button in Fig. 4 determines the random values: .
The workbook that builds learning vectors
Further values Uv, Ia, Pw and Qw are determined from Eq. (11).
After tests of several neural networks, a decision was made to build a neural network with topology (Fig. 5), with one hidden layer. The weights of neurons are determined by back propagation.
Structure of the neural network
Individual neurons in the network are structured according to Fig. 6.
Diagram of neuron
In the network being built, the following indications were accepted:
- t
- iteration step, t = 1, 2, …
- ith output of the neuron
- ith output of the network
- k
- network layer, k = 1, …, L
- L
- network output layer, the number of network layers
- i
- neuron number in layer, i = 1,…, Nk
- input signal in the kth layer
- input of the network
- j
- number of the input signal in the kth layer, j = 1,…, Nk−1
- N0
- number of inputs to the network
- Nk
- number of neurons in the kth layer
- NL
- number of neurons in the last layer
- neuron membrane potential in the kth layer
- weight of the jth input of the ith neuron in the kth layer
- ith reference signal output from the learning vector
- error of the ith network output,
- η
- network learning rate
- Q(t)
- error at the output of the network for one reference vector
- Q*(t)
- error at the output of the network for the entire epoch
The output of neuron (Fig. 6) at time t is described with the following relation:
while membrane potential is equal to:
The input neuron for k = 1 layer is equal to network inputs . Each layer has one input . Other inputs are the outputs of the previous layer.
The error at the output of the network for one learning vector σ is:
The weights of the individual neuron inputs are determined from the steepest descent rule:
where is the vector gradient.
From Eq. (16), for any weight in any layer, the following is obtained:
Parameter is determined differently than for the output layer and the hidden layer:
where
Network training is carried out by an incremental updating of weights, that is, each time after the entry of a successive learning vector, responses are determined and the weights are modified. The simulation is continued until the total output error for entire epoch Q*(t) is smaller than the accepted set Qmin.
where M is the number of learning vectors in the epoch.
The neuron activation function was adopted as a continuous unipolar function of the signum type:
where β is the steepness factor.
With low values of coefficient β, the function is usually mild. By increasing β, the plot becomes steeper until the threshold course is obtained.
The derivative of the activation function is as follows:
The calculation sheet in Fig. 7 allows an observation of the characteristic values of the network tested. Starting of the network training produces a script written in VBA that executes in a loop of a neural network algorithm according to (12) ÷ (21) and the block diagram in Fig. 8.
Sheet for the visualization of the network operation
Block diagram of the network learning algorithm
The start of the algorithm is possible for the weights that are selected at random from range or the reading stored from the previous simulations (Fig. 9).
The reading window of the recorded data
Learning of the network
The set of learning vectors that form one epoch consists of 200 elements. Owing to the ability to read and write data, it is possible to pause the simulation and to change its parameters during operation [4, 12].
Reading of the stored data allows a continuation of the previously stopped simulation. The window in Fig. 9 retrieves the values from the appropriate data sheet (Fig. 10).
Sheet with the saved results of the simulation
The output values of the neurons (Fig. 8a) are determined by analyzing the neurons in layers starting from the input layer; the output layer comes last.
For all the layers, the steepness factor β (20) of activation function fx was assumed as equal to 0.1.
The determination of value (Fig. 8b) shall be in accordance with Formula (18). This determination takes place starting from the output layer; the input layer comes last.
Correction of the values of weights (Fig. 8c) is carried out according to Relation (17).
Network learning factor η from Formula (17) was adopted on the first stage of the simulation as being constant and equal to 0.1. After an analysis of ca. 70,000 epochs, the value of target function Q*(t), which was calculated in accordance with Formula (19), began to oscillate on the level of 1.42. A decrease in Q*(t) occurred only after a reduction in network learning rate η. The correct procedure for the network training should provide for an ability to change this ratio during the analysis (Fig. 11).
The method to reduce the coefficient of network learning
Oscillations around the optimal solution are manifested with a momentary increase in the value of Q*(t).
Once the required value of Q*(t) from Eq. (19) has been reached, the network test is performed (Fig. 12).
The window for testing and results analysis
Network test
The network test consists in determining the values of UV, Ia, PW, and QW from Relation (11). These values are then substituted into the neural network input, whose solution is RS, LS, and ES. The window in Fig. 12 also allows a determination of the network’s solution for a selected set of weights.
Table 1 illustrates the network test for randomly selected values of RS, LS, and ES.
Table 1
Random values | Network input | Network output | Relative error | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
x1 | x2 | x3 | x4 | y1 | y2 | y3 | δ1 | δ2 | δ3 | δ | |||
RS | LS | ES | UV | Ia | P | Q | R’S | L’S | E’S | ||||
[Ω] | [H] | [V] | [V] | [A] | [W] | [VAr] | [Ω] | [H] | [V] | – | – | – | – |
20 | 0.2 | 180 | 229.95 | 0.7576 | 53.21 | 165.88 | 19.721 | 0.210 | 168.626 | −0.014 | 0.050 | −0.063 | 0.127 |
151 | 0.315 | 193 | 229.97 | 0.2048 | 39.395 | 25.7897 | 147.083 | 0.354 | 192.849 | −0.026 | 0.124 | −0.001 | 0.151 |
150 | 0.09 | 197 | 229.96 | 0.2159 | 48.794 | 9.18694 | 152.107 | 0.093 | 197.019 | 0.014 | 0.033 | 0.0001 | 0.048 |
147 | 0.3 | 197 | 229.97 | 0.1888 | 36.563 | 23.415 | 147.413 | 0.352 | 192.976 | 0.003 | 0.173 | −0.020 | 0.197 |
47 | 0.46 | 193 | 229.99 | 0.2434 | 17.368 | 53.211 | 41.373 | 0.409 | 195.946 | −0.120 | −0.111 | 0.015 | 0.246 |
150 | 0.09 | 193 | 229.95 | 0.2421 | 54.707 | 10.301 | 151.590 | 0.092 | 196.866 | 0.011 | 0.022 | 0.020 | 0.053 |
45 | 0.448 | 190 | 229.98 | 0.2706 | 19.015 | 59.256 | 41.182 | 0.408 | 195.771 | −0.085 | −0.089 | 0.030 | 0.205 |
145 | 0.4 | 190 | 229.97 | 0.2083 | 36.218 | 31.353 | 145.773 | 0.357 | 192.810 | 0.005 | −0.108 | 0.015 | 0.128 |
50 | 0.03 | 160 | 229.73 | 1.3705 | 309.422 | 58.163 | 58.641 | 0.026 | 159.829 | 0.173 | −0.133 | −0.001 | 0.307 |
150 | 0.1 | 200 | 229.96 | 0.1955 | 44.0056 | 9.20584 | 152.394 | 0.093 | 197.103 | 0.016 | −0.070 | −0.014 | 0.100 |
140 | 0.317 | 189 | 229.96 | 0.2384 | 44.6931 | 31.755 | 146.133 | 0.354 | 192.544 | 0.044 | 0.117 | 0.019 | 0.179 |
40 | 0.317 | 195 | 229.98 | 0.3259 | 28.036 | 69.507 | 40.416 | 0.402 | 195.029 | 0.010 | 0.268 | 0.0002 | 0.279 |
135 | 0.4 | 195 | 229.97 | 0.1896 | 31.937 | 29.691 | 143.751 | 0.358 | 192.972 | 0.065 | −0.105 | −0.010 | 0.180 |
145 | 0.3 | 190 | 229.96 | 0.2311 | 44.568 | 28.936 | 146.443 | 0.353 | 192.636 | 0.010 | 0.177 | 0.014 | 0.201 |
140 | 0.4 | 190 | 229.97 | 0.2125 | 36.378 | 32.615 | 145.157 | 0.356 | 192.790 | 0.037 | −0.110 | 0.015 | 0.162 |
The relative error for all the output neurons for the randomly adopted input vectors is:
The total network error for the accepted values of RS, LS, and ES are:
The percentage error made by the network is determined from the largest error (the top bar in the chart in Fig. 13), and it is equal to 26.8 %.
Mistake made by the network
Elektryk Wysokich Napiec
Conclusions
The large error value is shown for the input values that occur least frequently in the training set. An improved performance is possible by enlarging the training set or by reducing the range of acceptable changes of the values being sought.
Owing to the method presented of the selection of the electrical model parameters from the values that are measured on the receiver, it is not required to build any complex physical and electrical dependences. The engineering method of voltage, current, and power measurement allows one to determine the parameters of the model for constant electrical and mechanical conditions in the engine. The method presented is particularly useful in situations where measurement errors make it impossible to solve Eq. (2).
Building of a network with the use of the VBA environment is relatively simple. It requires the knowledge of the language basics. An important advantage of this approach is the ability to build its own networks of any topology. The design loop iteration depends largely on how one defines those variables that describe the network.
In the present solution, the individual variables occupy adjacent bytes of the memory. A sample definition of the variable holding the weights of neurons is:
where Lweights is the number of weights of all neurons.This solution facilitates the construction of a loop program, but special attention is to be paid to assigning the weight number with the neuron number.
An alternative is to build one’s own variable (using the opportunity to build one’s own type of variables) that represents the neuron, and then group all the parameters that describe the type of the neuron in this variable. This approach will make the program more transparent, but there are problems in the construction of iterative loops. This will make the source code longer and will require more CPU load.
Open Access
This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.