Outlier Detection-aided Supervised Learning for Modeling of Thinned Cylindrical Conformal Array
Yang Hong, Wei Shao, Yan He Lv, and Zhi Ning Chen
1School of Physics
University of Electronic Science and Technology of China, Chengdu, 610054, China
yanghong@std.uestc.edu.cn, weishao@uestc.edu.cn
2Department of Electrical and Computer Engineering
National University of Singapore, Singapore, 117583, Singapore
aplyh@nus.edu.sg, eleczn@nus.edu.sg
Submitted On: February 27, 2023; Accepted On: June 13, 2023
In this paper, a scheme of outlier detection-aided supervised learning (ODASL) is proposed for analyzing the radiation pattern of a thinned cylindrical conformal array (TCCA), considering the impact of mutual coupling. The ODASL model has the advantage in speed improvement and memory consumption reduction, which enables a quick generation of the synthesis results with good generalization. The utilization of the active element pattern (AEP) technique in the model also contributes to the prediction of the array performance involving mutual coupling. The effectiveness of the ODASL model is demonstrated through a numerical example of the 12-element TCCA.
Index Terms: Active element pattern (AEP), conformal array, outlier detection-aided supervised learning (ODASL), thinned array.
Recently, conformal arrays have gained popularity in airborne and satellite applications for the sake of their adaptability and aerodynamic performance. However, their analysis and synthesis are particularly complicated due to the varying positions and axial directions of the elements [1]. In addition, the impact of mutual coupling between elements makes it difficult to analyze the far field of conformal arrays, using the directional product theorem applied to planar arrays [2].
The supervised learning method has been extensively recognized as a prediction tool with significant improvement and contribution to electromagnetic (EM) modeling [3–5]. It provides a fast synthesis process for array behaviors while maintaining high-level accuracy with a reduced number of full-wave simulations. However, a multitude of inner parameters of a supervised learning method needs to be determined for large-space and high dimensional problems [6, 7], which easily makes learning progress tardy. In addition, the dependence on sampling data may affect the credibility of the model, especially in the context of array modeling with a complex structure and EM environment. To solve this problem, outlier detection (OD) in data mining is explored as an effective decision-making tool [8]. The multivariate distance-based OD method is used to traverse the raw dataset and identify outlier objects, helping to construct the model effectively.
Considering the impact of mutual coupling and array environment on radiation patterns, a large-scale array can be transformed into the superposition of small sub-arrays by employing the active element pattern (AEP) technique [9]. The technique offers attractive benefits, including the avoidance of heavy calculation burden associated with the whole-array simulation.
To make better use of sampling data, this paper proposes an outlier detection-aided supervised learning (ODASL) model as an alternative to the costly measurement or full-wave simulation. Considering mutual coupling and EM environment, the AEP technique is employed to extract the patterns of sub-arrays, instead of the whole array. By filtering out the invalid sampling data, i.e., outlier data, the ODASL model can obtain satisfactory prediction results, with a strong generalization ability even for larger thinned cylindrical conformal array (TCCA) scales. A TCCA is taken as an example to demonstrate the effectiveness of the proposed model.
In the regression supervised learning method for EM modeling, the learning information is completely sourced from the sampling dataset. Hence, it is necessary to perform OD on the raw dataset for the analysis to accurately construct the ODASL framework. The training dataset , defined as , contains a total of labeled samples, and the EM response denoted as in each sample is obtained from full-wave simulations. Assuming that there are uncertain outlier samples in the raw training dataset, they can be recorded as . To measure the similarity in distance between a pair of data objects and , a distance function is defined as , satisfying the positive definiteness: , shown as
(1) |
where represents the number of sampling points in for each sample, and are the the sampling point values for samples and , respectively. Generally, the Euclidean distance, which is the second-order Minkowski distance with in formula (1), is adopted as the calculation method to ensure the stability of results regardless of the variation of the dataset space.
With the aid of the transfer function (TF) [10], poles and residues , which are TF coefficients to describe behaviors of samples, are extracted as the corresponding measurement values and used to identify outliers. The implementation of the specific steps in the OD stage shown in Fig. 1, is described as follows:
1. Step 1: For each sample in , the full-wave simulation of the EM response target , representing the results obtained with AEP, is completed in the range of the specified geometric parameters. The results obtained from the frequency-domain analysis are then fitted to poles/residues-based transfer functions utilizing the vector fitting (VF)technique [11].
2. Step 2: The TF coefficients extracted from all samples have the same order , and the poles of a sample are set as . Similarly, the residues of are represented by , where and are the pole and residue values of the ith order of , respectively. and in the dataset are clustered according to their corresponding order. For the calculation of , it is converted to the calculation of the accumulation of two-part distances: one is related to expressed as , and another is for .
3. Step 3: The given dataset is evaluated by ) To distinguish obvious outliers unequivocally, the distances of and with the same order are subject to a restrictive distance threshold , with a dimension of . The threshold varies with different sampling datasets. ) Intrinsically, due to the interdependence among the data points, detecting micro-clusters becomes more complex as these outliers may be neglected as data points from the dense regions of data distribution. Therefore, the outlier score mechanism (OSM) is adopted,denoted as
(2) |
where
(3) |
where and are the average values of the ith order pole and residue, respectively, and represent the standard deviation. The absolute value of the score indicates the distance between the data points and the population averages, within the scope of the standard deviation. By filtering out samples with outlier points, a collection of outlier samples is obtained, and the remained dataset is .
To sum up, the approach for analyzing and identifying outliers can be accomplished by choosing the top outliers with the largest outlier scores from the score ranking list and by selecting outliers from a cut-off threshold, depending on the distribution ofoutliers.
Taking the parametric modeling of arrays for example, the specific procedure with various stages of the ODASL model is presented in Fig. 2. The overall input consists of geometric parameters of the element and element spacing. In practice, we want to obtain the relationship between variables in and EM response that is affected by mutual coupling and array environment. For exploring the relationship of labeled data pairs of , we have
(4) |
where is the nonlinear function for the mapping of input space to output space . In the whole set, and cannot be traversed completely because of the limitation of the sampling requirement, expressed by
(5) |
In Phase I of the ODASL architecture, dataset generation is the first step to provide the samples for modeling. According to the modeling requirement, the EM response is set as the target, such as or pattern information. Then the next step is to obtain the samples with the variable geometric structure and element spacings from the full-wave simulation, and to extract the relevant TF coefficients from EM response.
In Phase II, the potential outliers in are filtered out using the previous process, and then the input samples of are obtained. For the output, the sub-array patterns based on TF coefficients are accurately captured by the AEP technique. According to the positions of elements, they are categorized into edge element, adjacent-edge element, and interior element, and their AEPs are extracted separately [12, 13].
Subsequently, in the training process of the multi-branch radial basis function neural network (RBFNN) [14], by adjusting the network parameters of multi-branch RBFNN, including the weights of hidden/output layer and center/width of basis function, the relationships between the input and the output in three branches are established. The main purpose of training is to minimize the disparity between and the predicted from the superposition of all branch results.
To test the model, its generalization ability is crucial to the stable prediction, especially for input beyond the range of the training dataset [15]. Combined with AEP, a high-degree freedom of array design is guaranteed, and the time-consuming full-wave simulation for the whole array is substituted by the superposition of the predictions from multi-branch networks for sub-arrays. Once the proposed model is well-trained, it immediately provides an accurate response for a given input.
A 12-element TCCA in Fig. 3 is taken as the example to evaluate the ODASL model, where it works at [16]. The elements are placed on a cylindrical substrate with a radius and a relative dielectric constant of . To obtain a high degree of freedom, the input is denoted as , extracted with the design of the experiment method for sampling [17]. The circumferential distance in is approximately from to during the data collection, where is equal to 2, 3, or 4 for AEP extraction, depending on the sub-array scale. Table 1 shows the sampling data for the branch of the adjacent-edge element. Data values are standardized before they are used in the network. Similarly, 81 training samples and 36 testing samples are collected for edge elements, and 100 training samples and 64 testing samples are collected for interior elements.
Structure Parameter | Training Dataset (100 Samples) | Testing Dataset (49 Samples) | ||||
---|---|---|---|---|---|---|
Min | Max | Step | Min | Max | Step | |
18 | 22.5 | 0.5 | 18.6 | 21.6 | 0.5 | |
13 | 17.5 | 0.5 | 13.8 | 16.8 | 0.5 | |
1.75 | 2.2 | 0.05 | 1.82 | 2.12 | 0.05 | |
5.6 | 6.5 | 0.1 | 5.75 | 6.35 | 0.1 | |
1.5 | 2.4 | 0.1 | 1.65 | 2.25 | 0.1 | |
-8.5 | -7.4 | 0.1 | -8.75 | -7.35 | 0.1 | |
37 | 46 | 1 | 38.5 | 44.5 | 1 |
To predict the array pattern considering mutual coupling, firstly, the AEPs of all elements are extracted from the raw data, and the corresponding and are obtained. All collected raw datasets are operated in OD processing, and the distances of and for each order are acquired, with the maximum order . Secondly, the correlated values of the OSM in formula (3) are calculated, and the scores of each sample are ranked in terms of their degree of deviation. For outlier samples, the distance exceeds the threshold, and meanwhile, the level of outliers is relatively obvious. An example of OD processing for edge elements in the training process is displayed in Table 2, where the values of and score ranking for each order are exhibited.
Further, Table 3 embodies the situations of outlier samples, with a certain order of or in over and a high degree of dispersion score, where is the th distance between and for the outliers. Similarly, performing the same operations on all element categories, the resulting filtered sets are used as the branch target outputs.
Order | 1 | 2 | 3 | 4 | 5 | 6 |
---|---|---|---|---|---|---|
0.71 | 0.87 | 0.55 | 0.68 | 0.89 | 2.87 | |
2.18 | 2.91 | 3.62 | 2.13 | 6.57 | 2.29 | |
0.35 | 0.46 | 0.69 | 1.61 | 2.42 | 3.54 | |
0.37 | 0.49 | 1.41 | -0.98 | 3.77 | 0.21 | |
0.25 | 0.35 | 0.82 | 0.17 | 0.14 | 0.09 | |
1.31 | 1.15 | 1.52 | 0.81 | 2.20 | 0.83 | |
Order | 7 | 8 | 9 | 10 | 11 | 12 |
0.29 | 0.55 | 0.48 | 1.76 | 0.87 | 0.45 | |
1.87 | 1.48 | 2.12 | 4.61 | 1.12 | 1.83 | |
-0.31 | -0.42 | 0.45 | 2.23 | 1.08 | 0.74 | |
-0.11 | -0.85 | 1.44 | 1.25 | 0.82 | 0.55 | |
0.13 | 0.06 | 0.05 | 0.08 | 0.14 | 0.06 | |
1.10 | 0.84 | 0.53 | 2.19 | 0.25 | 0.37 |
No. of Outlier | Outlier Order | |||
---|---|---|---|---|
Index | ||||
1 | 4 | 3.52 | 3.38 | |
2 | 7 | 2.91 | 1.41 | |
3 | 4 | 3.26 | 3.32 | |
4 | 6 | 3.72 | 2.30 | |
5 | 11 | 1.10 | 2.78 | |
6 | 10 | 2.49 | 2.80 | |
7 | 4 | 1.25 | 3.09 | |
8 | 7 | 2.63 | 1.26 | |
9 | 1 | 1.18 | 5.29 | |
10 | 10 | 2.27 | 2.72 | |
11 | 11 | 1.06 | 2.71 | |
12 | 7 | 2.42 | 1.15 | |
13 | 5 | 1.21 | 2.21 |
Assisted by the characteristics of RBFNN, fast learning speed and high accuracy, the ODASL model predicts the pattern promptly from the three branches, avoiding full-wave EM simulations of the whole array. Furthermore, the predicted results can be written as
(6) |
where , and represent the fields acquired from the proposed ODASL model of all the edge elements, adjacent-edge elements, and interior elements, respectively.
To describe the above formula, is shown as an example:
(7) |
where refers to the excitation amplitude of the sth edge element, is the number of edge elements, is the AEP result of the sth edge element, is the wavenumber in free space, where is the wavelength, and is the spatial phase factor. Accordingly, the far-field pattern of the TCCA is obtained by superimposing the extracted results.
Accordingly, after the OD process filters out 13 training and 6 testing outlier samples for edge elements, 15 training and 7 testing samples for adjacent-edge elements, and 16 training and 10 testing samples for interior elements, the construction of the ODASL model for the 12-element TCCA costs approximately 13.92 hours, and the average mean absolute percent errors (MAPEs) of the training and testing processes for the whole proposed ODASL model are and 4.358%, respectively. All calculations are performed on an Intel i7-6700 3.40 GHz machine with 16 GB RAM.
As an example of TCCA modeling, the results of two separate arrays are shown in Fig. 4. For Array 1, the parameters of input are within the training dataset range, while those of input in Array 2 are outside the range: The parameters of , with a MAPE value 3.317%, and , 46.3, 36.8, 46.1, 36.6], with a MAPE value 4.334%. From Fig. 4, the obtained agreement between the simulation and the ODASL model results proves the advantage of the proposed model in terms of accuracy for input parameters both within and outside the training dataset range.
To reflect the properties of the ODASL model with the TF coefficients as the outputs by extracting the AEPs from the sub-arrays, it is compared with the efficient extreme learning machine (ELM) [18, 19] and RBFNN, which directly output the whole array performance without involving the AEP technique. Table 4 indicates the network structure and computational accuracy of the three models. The error measurement standards include MAPE and root mean square error (RMSE), and the small MAPE and RMSE values of ODASL show its well-predicted performance in accuracy and stability. The ELM and RBFNN models collect 49 training samples and 25 testing samples within the parameter range in Table 1.
TF Order | No. of Hidden Neurons | MAPE | RMSE | |
---|---|---|---|---|
RBFNN | 17 | 23 | 0.0125 | |
21 | 30 | 0.0093 | ||
ELM | 17 | 10 | 0.0076 | |
21 | 15 | 0.0052 | ||
ODASL | 12 | 0.0029 | ||
14 | 0.0018 |
The ELM and RBFNN models use a single hidden layer, with 17 inputs, including the structure parameters and non-uniform element spacings. Here two cases of different TF orders are employed for TCCA modeling.
From Table 4, the proposed ODASL model shows lower errors than those of ELM and RBFNN for the two cases. In other words, more collected samples for the training of ELM and RBFNN are needed to get the same level of accuracy.
To examine the applicability of the OD process in the proposed model, Table 5 provides the error comparison for the multi-branch RBFNN model and the ODASL model, which are based on the same dataset obtained from the AEP technique. It is shown that even though the network structures for the two models are similar, the proposed model yields more accurate results.
Element | Category | ODASL Model | Multi-branch RBFNN Model | |
---|---|---|---|---|
Training Error | Testing Error | Training Error | Testing Error | |
Edge | ||||
Adjacent-edge | ||||
Interior |
In Fig. 5 (a), the predicted MAPEs of the multi-branch RBFNN and the proposed model are compared, where the parameter of is considered as a single variable of the array. Figure 5 (b) provides the results of the parameter of . The results show that within the training dataset range, the proposed model gets satisfactory results. Even if the input parameter is out of the range of the training dataset, the proposed model can obtain much more accurate results than the multi-branch RBFNN model.
For further comparison, the proposed ODASL model, full-wave simulation, multi-branch RBFNN and ELM are employed to simulate a 22-element array, a 46-element array, and a 75-element array. The CPU time is listed in Table 6. Compared with the full-wave simulation, the ODASL model is constructed at a cost of 13.9 hours. For large scale arrays, however, the well-trained ODASL can be re-called to realize the fast simulation. Because ELM does not involve the AEP technique, three ELM models corresponding to the different array scales are constructed. Thus the whole modeling time with ELM is much more than that with ODASL. Compared with the multi-branch RBFNN combined with the AEP technique, the ODASL model filters out the invalid sampling data. Therefore, ODASL needs fewer training samples than the multi-branch RBFNN, and then it shows higher modeling efficiency.
Number of Elements | ||||
---|---|---|---|---|
Full-wave Simulation | – | – | – | |
RT | ||||
Total | ||||
ELM | ||||
RT | ||||
Total | ||||
Multi-branch RBFNN | – | – | ||
RT | ||||
Total | ||||
ODASL | CT | – | – | |
RT | ||||
Total | ||||
CT/RT: Construction/Running time, h: hour, m: minute |
In this paper, a novel ODASL framework is proposed for efficient TCCA modeling, addressing the challenge posed by mutual coupling and data dependence, with the aim of meeting high-performance requirements for radiation pattern prediction. The proposed model provides a fast pattern realization process with an appreciable reduction of full-wave EM simulations. Combined with the AEP technique, the OD method employs multivariate distance-based clustering and OSM to enhance the discernment and quantification of outliers, constructing the operation for a highgeneralizable model. The valid samples are obtained by outlier elimination, and a numerical example demonstrates the effectiveness of the ODASL model. Additionally, the proposed model with the related data mining method can be further extended to other microwave applications.
This work was supported by the National Natural Science Foundation of China under Grant 62171093 and by the Sichuan Science and Technology Programs under Grants 2022NSFSC0547 and 2022ZYD0109.
[1] J. Harris and H. Shanks, “A method for synthesis of optimum directional patterns from nonplanar apertures,” IRE Trans. Antennas Propag., vol. 10, no. 3, pp. 228-236, May 1962.
[2] L. I. Vaskelainen, “Constrained least-squares optimization in conformal array antenna synthesis,” IEEE Trans. Antennas Propag., vol. 55, no. 3, pp. 859-867, Mar. 2007.
[3] C. A. Oroza, Z. Zhang, T. Watteyne, and S. D. Glaser, “A machine-learning-based connectivity model for complex terrain large-scale low-power wireless deployments,” IEEE Trans. Cognit. Commun. Netw., vol. 3, no. 4, pp. 576-584, Dec. 2017.
[4] L.Y. Xiao, W. Shao, F. L. Jin, B. Z. Wang, and Q. H. Liu, “Inverse artificial neural network for multiobjective antenna design,” IEEE Trans. Antennas Propag., vol. 69, no. 10, pp. 6651-6659, Oct. 2021.
[5] P. Liu, L. Chen, and Z. N. Chen, “Prior-knowledge-guided deep-learning-enabled synthesis for broadband and large phase shift range metacells in metalens antenna,” IEEE Trans. Antennas Propag., vol. 70, no. 7, pp. 50245034, July 2022.
[6] A. Seretis and C. D. Sarris, “An overview of machine learning techniques for radiowave propagation modeling,” IEEE Trans. Antennas Propag., vol. 70, no. 6, pp. 3970-3985, June 2022.
[7] R. Caruana, N. Karampatziakis, and A. Yessenalina, “An empirical evaluation of supervised learning in high dimensions,” Proc. 25th Int. Conf. Mach. Learn., pp. 96-103, 2008.
[8] H. Wang, M. J. Bah, and M. Hammad, “Progress in outlier detection techniques: A survey,” IEEE Access, vol. 7, pp. 107964-108000, Aug. 2019.
[9] D. F. Kelley and W. L. Stutzman, “Array antenna pattern modeling methods that include mutual coupling effects,” IEEE Trans. Antennas Propag., vol. 41, no. 12, pp. 1625-1632, Dec. 1993.
[10] F. Feng, C. Zhang, J. Ma, and Q. J. Zhang, “Parametric modeling of EM behavior of microwave components using combined neural networks and pole-residue-based transfer functions,” IEEE Trans. Microw. Theory Techn., vol. 64, no. 1, pp. 60-77, Jan. 2016.
[11] B. Gustavsen and A. Semlyen, “Rational approximation of frequency domain responses by vector fitting,” IEEE Trans. Power Del., vol. 14, no. 3, pp. 1052-1061, July 1999.
[12] Q. Q. He and B.-Z. Wang, “Design of microstrip array antenna by using active element pattern technique combining with Taylor synthesis method,” Prog. Electromagn. Res., vol. 80, pp. 63-76,2008.
[13] Y. Hong, W. Shao, Y. H. Lv, B. Z. Wang, L. Peng, and B. Jiang, “Knowledge-based neural network for thinned array modeling with active element patterns,” IEEE Trans. Antennas Propag., vol. 70, no. 11, pp. 11229-11234, July 2022.
[14] F. Feng, W. Na, J. Jin, J. Zhang, W. Zhang, and Q. J. Zhang, “Artificial neural networks for microwave computeraided design: The state of the art,” IEEE Trans. Microw. Theory Techn., vol. 70, no. 11, pp. 4597-4619, Nov. 2022.
[15] H. Kabir, Y. Cao, and Q. J. Zhang, “Advances of neural network modeling methods for RF/ microwave applications,” Applied Computational Electromagnetics Society (ACES) Journal, vol. 25, no. 5, pp. 423-432, May 2010.
[16] Y. Kimura, S. Saito, Y. Kimura, and T. Fukunaga, “Design of wideband multi-ring microstrip antennas fed by an L-probe for single-band and dual-band operations,” Proc. 2020 IEEE AP-S. Int. Symp, pp. TU-A1.4A.8, July 2020.
[17] R. Schmidt and R. G. Launsby, “Understanding Industrial Designed Experiments,” Colorado Springs, CO, USA: Air Force Academy, 1992.
[18] L.Y. Xiao, W. Shao, S. B. Shi, and B. Z. Wang, “Extreme learning machine with a modified flower Pollination algorithm for filter design,” Applied Computational Electromagnetics Society (ACES) Journal, vol. 23, no. 3, pp. 279-284,2018.
[19] B. Deng, X. Zhang, W. Gong, and D. Shang, “An overview of extreme learning machine,” Proc. 4th Int. Conf. Control Robot. Cybern. (CRC), pp. 189-195, Sep. 2019.
Yang Hong received the B.S. degree in electronic information science and technology from the University of Electronic Science and Technology of China (UESTC), Chengdu, China, in 2017. Currently, she is working toward the Ph.D. degree in physics at UESTC. In 2022, she joined the Department of Electrical and Computer Engineering, National University of Singapore, Singapore, as a visiting student.
Her research interest is neural network, antenna array, and computational electromagnetics.
Wei Shao received the B.E. degree in electrical engineering from UESTC in 1998, and received M.Sc. and Ph.D. degrees in radio physics from UESTC in 2004 and 2006, respectively.
He joined UESTC in 2007 and is now a professor there. From 2010 to 2011, he was a visiting scholar in the Electromagnetic Communication Laboratory, Pennsylvania State University, State College, PA. From 2012 to 2013, he was a visiting scholar in the Department of Electrical and Electronic Engineering, the University of Hong Kong. His research interests include computational electromagnetics and antenna design.
Yan-he Lv received the B.S. degree and Ph.D. degree in electronic information science and technology and radio physics from UESTC in 2017 and 2022, respectively.
He is currently a Research Fellow in the National University of Singapore, Singapore. His main research interests include metasurface, phased array, time-reversed electromagnetic, and computational electromagnetics.
Zhi Ning Chen received the B.Eng., M.Eng., and Ph.D. degrees in electrical engineering from the Institute of Communications Engineering (ICE), China, in 1985, 1998, and 1993, and a second Ph.D. degree from the University of Tsukuba, Tsukuba, Japan in 2003.
He joined the National University of Singapore in 2012 as a tenured full professor. From 1988 to 1995, He was a lecturer and later a professor with ICE and a post-doctoral fellow and later an associate professor with Southeast University, Nanjing, China. From 1995 to 1997, he was a research assistant and later a research fellow with the City University of Hong Kong, Hong Kong. In 2001 and 2004, he visited the University of Tsukuba twice under the JSPS Fellowship Program (senior fellow). In 2004, he joined the IBM Thomas J. Watson Research Center, Ossining, NY, USA, as an academic visitor. In 2013, he joined the “Laboratoire des SignauxetSystèmes,” UMR8506 CNRS-Supelec-University Paris Sud, Gif-sur-Yvette, France, as a senior DIGITEO guest scientist. In 2015, he joined the Center for Northeast Asian Studies, Tohoku University, Sendai, Japan, as a senior visiting professor.
He was elevated a Fellow of the IEEE for the contribution to small and broadband antennas for wireless applications in 2007 and a Fellow of the Academy of Engineering, Singapore, in 2019 for the contribution to research, development, and commercialization of wireless technology.
He is pioneering in developing small and wideband/ultrawideband antennas, wearable/implanted medical antennas, package antennas, near-field antennas/coils, 3-D integrated LTCC arrays, microwave lens antennas, microwave metamaterial-metasurface (MTS)-metaline-based antennas for communications, sensing, and imaging systems. His current research interests include the translational research of electromagnetic metamaterials and the applications of prior-knowledge-guided machine learning to antenna engineering.
ACES JOURNAL, Vol. 38, No. 9, 638–645
doi: 10.13052/2023.ACES.J.380902
© 2023 River Publishers