Abstract
While granular computing, in particular fuzzy sets, offers a comprehensive conceptual framework of managing conceptual entities-information granules with the processing mechanisms provided by fuzzy sets, this mechanism does not fully engage experimental data whose characteristics could be fully incorporated in the obtained results. Having this in mind, the fundamental constructs of relational computing and fuzzy relational computing were augmented by mechanisms originating from the area of granular computing. It is advocated that the prescriptive constructs existing in the area of relational calculus are augmented by incorporating a descriptive component, which is governed by the existing data. It is shown that the principle of justifiable granularity helps enhance the existing results by experimental evidence residing with the data and develop the results in the form of information granules.
人工智能、数据挖掘。E-mail:wpedrycz@ualberta.ca
Relational computin
In this paper, some generic definitions are presented concerning Cartesian products, projections and the following reconstruction mechanism, composition operators. In the sequel, fuzzy relational equations are covered along with a topic of a reconstruction of fuzzy relations. In addition, the principle of justifiable granularity, which is central to the ensuing granular constructs is introduced. Moreover, both a one-dimensional and multivariable case is studied. Furthermore, G-operations are presented 4 along with a variety of architectures (such as G-Cartesian products, G-relational composition operations, granular relational equations, and granular logic networks).
Let A and B be two sets or fuzzy sets defined in finite spaces X and Y, respectively, dim(X) =n, and dim(Y)=m. They are described by characteristic functions (sets) or membership functions (fuzzy sets). In what follows, the fundamental definitions and constructs forming a backbone of relational calculus is briefly recalled. Interestingly, these entities are the same for sets and fuzzy sets meaning that the definitions put forward apply to the same extent to sets and fuzzy sets.
Relations and fuzzy relations are cornerstones when describing and processing relationships existing among real-world objects.
Cartesian product
A Cartesian product of A and B, A B, is a relation (fuzzy relation) described by their characteristic or membership function
(A B)(x, y)=min (A(x), B(y))
x X, y Y | (1) |
For fuzzy sets, the minimum operation is replaced by any t-norm (A B)(x, y)= A(x)t B(y). Of course, if A and B are sets, all t-norms return the same result.
To focus attention, relations defined over two spaces are considered. However, the constructs are readily extended to multidimensional situations. A two-dimensional fuzzy relation R is defined over the Cartesian product X Y. The well-known definition of projection is presented as
Projection of fuzzy relation
The projections of R on X and Y, respectively, are defined as
projXR(x)=supyR(x,y) | (2) |
and
projYR(y)=supxR(x,y) | (3) |
The result of projection is a fuzzy set. From the computational perspective, it is worth noting that the result of projection is determined by the maximal value of the relation across one argument for the second one being fixed. In this way, this is an optimistic view: the membership values of R depend upon the extreme value of the slice of R reported along the x or y coordinate.
Reconstruction
The results of projection of R can be used to reconstruct the original fuzzy relation. This is a typical example in image processing when an original 3D object has to be reconstructed in an efficient way based on their projections.
The common way of realizing reconstruction is to take a Cartesian product of the projection results, say A and B and compute their Cartesian product. If R is an image, it is apparent that the projection operation returns a result based on the maximal value of R for a value of one argument fixed, see

Fig. 1 Projection of relation on x and y coordinates along with the reconstruction result
There is no guarantee that A B always returns R. This problem will be expounded later on.
The composition operators that commonly discussed in relational calculus involve a sup-min (max-min) and inf-max (min-max) operations. Given a fuzzy set A in X and a fuzzy relation R, these compositions return another fuzzy set B in Y with the following membership function
sup-min composition
B =A R, B(y) = supx[min(A(x), R(x,y)] | (4) |
inf-max composition
B = A R, B(y) =
infx[max(A(x), R(x,y)] | (5) |
The apparent generalizations concern the use of triangular norms and conforms in the realization of the composition operators
sup-t (max-t)
B(y)=supx[A(x)tR(x,y)] | (6) |
inf-s (min-s)
B(y)=infx[A(x)sR(x,y)] | (7) |
Where t stands for t-norm and s denotes some t-conorm. Finally, the composition operation could be further generalized into s-t and t-s composition which are read as
s-t
B(y) = Sx[A(x)tR(x,y] | (8) |
t-s
B(y) = Tx[A(x)sR(x,y] | (9) |
It is worth noting that if A is an overall space, A(x)=1, then the composition operator returns the projection of R.
t-norms and t-conorms are commonly used in the realization of logic aggregation. The or operation of fuzzy sets defined in the same space A1, A2,..., Ac is applied to the individual membership grades A1(x) A2(x)... An(x) returning A1(x) sA2(x)s... sAn(x). For the operation, one considers any t-norm and the results becomes A1(x) tA2(x)t... tAn(x). One could note that depending on the t-norm being used, the largest and smallest membership grades of the results are bounded
or: maximum < A1(x) sA2(x)s... sAn(x)< drastic sum
and
and: drastic product < A1(x) tA2(x)t... tAn(x)< minimum.
If the number of arguments increases, the result tends to zero (and operation) or 1 (or operation).
Fuzzy relational equations have been studied from the seventies
–estimation problem: Given A and B, determine R
–inverse problem: Given R and B, determine A
For the sup-min and inf-min composition operators, the solutions are well-known
The estimation problem can be generalized by considering a system of fuzzy relational equations
A1 R = B1, A2 R = B2,..., Ac R = Bc
The previous way of carrying out projection is generalized in the form of the procedure where the s-t composition is introduced and some parametric flexibility are brought in to the calculations
A(x) = Sx [w(y)tR(x,y)] | (10) |
B(y)= Sy [v(x)tR(x,y)] | (11) |
In case of finite spaces X and Y, w and v are vectors of weights assuming values in [0,1]. Their objective is to calibrate an impact of individual elements of X and Y when determining the respective contributions to the generated results.
As to the reconstruction, it is augmented by weights f and g which yield
(12) |
The overall scheme of projection and reconstruction is visualized in

Fig. 2 Realization of reconstruction of fuzzy relation
As a matter of fact, the expressions (10)–(12) are examples of logic neurons: the projection is completed in terms of the two OR neurons whereas the reconstruction is conducted with the aid of the AND neuron.
The selection of the values of the weights w, v, f, and g is completed by solving the following optimization problem,
minw,v,f,g || R – | (13) |
where ||. || is a certain distance function.
Building information granules on a basis of experimental data constitutes a pivotal item on the agenda of granular computing with far-reaching implications on the design methodology of granular models and ensuing applications. Clustering and fuzzy clustering
Formally speaking, these two intuitively appealing criteria are expressed by the criterion of coverage and the criterion of specificity. Coverage states how much data are positioned behind the constructed information granule. Rephrase it differently, coverage quantifies an extent to which information granule is supported by available experimental evidence. Specificity, on the other hand, is concerned with the semantics of information granule stressing the semantics (meaning) of the granule.
In what follows, the developments of the principle of justifiable granularity is elaborated on by starting with a generic version. The main assumptions and the features of the environment in which information granules are being formed are also summarized in a concise manner. Information granules are formalized in an interval format so that an interval [a, b] is constructed. The interval type of information granule is explored for illustrative purposes; however the method is far more general and the principle can be applied to different granular and numeric experimental data and produce information granules formalized in terms fuzzy sets, rough sets and others.
One-dimensional real data x1, x2, …, xN are considered. The bounds of the data are xmin and xmax; xmin= arg mink xk, xmax = arg maxk xk.
The coverage measure is associated with a count of the number of data embraced by A, namely
cov(A)= card | (14) |
Where card (.) denotes the cardinality of A, viz. the number (count) of elements xk belonging (covered) to A. In essence, coverage has a visible probabilistic flavor. The specificity of A, sp(A), is regarded as a decreasing function g of the size (length) of information granule. If the granule is composed of a single element, sp(A) attains the highest value and returns 1. If A is included in some other information granule B, then sp(A) > sp(B). In a limit case, if A is an entire space, sp(A) returns zero. For an interval-valued information granule A =[a, b], a simple implementation of specificity with g being a linearly decreasing function comes as
sp(A)=g(length(A))=1- | (15) |
where range is defined as range = xmax -xmin.
The criteria of coverage and specificity are in an obvious conflicting relationship. The increase in the values of coverage implies lower values of specificity and vice versa. If it is wished to maximize both criteria, a sound compromise has to be reached. Let us introduce the following product of the criteria
V = cov(A)sp(A) | (16) |
The maximization of the performance index V gives rise to information granule where some trade-offs between coverage and specificity are reached. The design of information granule is accomplished by maximizing the above product of coverage and specificity. Formally speaking, consider that an information granule is described by a vector of parameters p, V(p), the principle of justifiable granularity yields an information granule that maximizes V, popt = arg pV(p).
To maximize the index V through the adjusting the parameters of the information granule, two different strategies are encountered:
(1) A two-phase development is considered. First a numeric representative (mean, median, mode, etc.) of the available data is determined. It can be regarded as their initial representation. Next, the parameters of the information granule are optimized by maximizing V. For instance, in case of an interval, one has two bounds (a and b) to be determined. These two parameters are determined separately, viz. a and b are formed by maximizing V(a) and V(b). The data used in the maximization of V(b) involves the data larger than the numeric representative. Likewise, V(a) is optimized on a basis of the data lower than this representative.
(2) A single-phase procedure in which all parameters of information granule are determined at the same time. In the above case, the values of a and b are simultaneously optimized. In comparison with the previous method, here the location of the interval is not biased towards one of the “anchor” points such as the mean or median. This approach yields more flexibility and finally results in possibly higher values of V(p).
As an example, let us consider 500 data governed by a normal distribution N(0, 1). The range is [-3.43, 3.00].
The use of the principle of justifiable granularity where both bounds are optimized at the same time leads to the interval information granule [-0.896, 1.261]. The obtained product of the coverage and specificity criteria is 0.730.
Proceeding with the separate optimization of the bounds and assuming a numeric representative to be the mean value (0.047), the optimal interval [-0.983, 0.846] is obtained and the optimized performance index is 0.654. If the numeric representative is taken as the median (0.654), the numeric representative is practically the same as before, [-0.987, 0.847] resulting in the same value of the performance index.
Considering the 500 data generated by the Laplace distribution L(0,1), the obtained results are
–simultaneous optimization of the bounds: [-1.876 , 1.497 ], performance is 0.822
–separate optimization of the bounds, mean as the representative: [-1.493, 1.498], performance 0.774
–separate optimization of the bounds, median as the representative: [-1.480, 1.473], performance 0.768
For the Cauchy distribution (again 500 data), the obtained results are
–simultaneous optimization of the bounds: [-10.123 , 14.808], performance is 0.970
–separate optimization of the bounds, mean as the representative: [-5.639, 21.847]
performance 0.958
–separate optimization of the bounds, median as the representative: [-5.943, 14.794], performance 0.952
If fuzzy sets are considered as information granules, the definitions of coverage and specificity are reformulated to take into account the notion of partial membership. Here the fundamental representation theorem is invoked, stating that any fuzzy set can be represented as a family of its α-cuts [8,15], namely
A(x)=sup | (17) |
where
(18) |
The supremum (sup) operation is taken over all values of α. In virtue of the representation theorem, any fuzzy set represented as a collection of sets is obtained.
Having this family of -cuts in mind and considering Eqs. (
–coverage
cov(A)= / | (19) |
where X is a space over which A is defined. Moreover, one assumes that A can be integrated. The discrete version of the coverage expression comes in the form of the sum of membership degrees. If each data point is associated with some weight w(x), the calculations of the coverage involve these values
cov(A)= / | (20) |
–specificity
sp(A)= | (21) |
Let us consider multivariable data X ={x1, x2,..., xN} where xk
Around the numeric prototype v, one spans an information granule G, G=(v, ρ) whose optimal size (radius) is obtained as the result of the maximization of the already introduced criterion, namely
(22) |
where
cov(Vi)=card{xk | ||xk-V|| n ρ} | (23) |
As before, the specificity is expressed as a linearly decreasing function of ρ
sp(G) =1 - ρ | (24) |
The geometry of information granule depends upon the form of the distance function. For instance, the Tchebyshev distance implies a hyperbox shape of the granules.
When using the fuzzy clustering method, the data X come with their weights associated with the individual elements of X, say (x1,w1), (x2,w2), ..., (xN,wN), where wk [0,1] serves as a degree of membership of the kth data point. The coverage criterion is modified to reflect the existing weights. Introduce the following set
Ω= { xk | ||xk-v|| nρ} | (25) |
Then the coverage is expressed in the form
cov(G) = | (26) |
The specificity measure is defined as presented before.
Projections and composition operators are examples of multi-argument aggregation operators. Many input variables give rise to a single result. In this paper, it is advocated that the result of aggregation of numeric input arguments is an information granule which helps incorporate the diversity of input arguments and copes with the descriptive aspect of the result.
The classic definition of projection, although conceptually sound, is quite limited when it comes to the incorporation of data distribution. Both in Eq. (
In sum, it could be intuitive to anticipate that the result of projection has to accommodate the diversity of the entries of R forming any row or column and become an information granule of type-2. The result of any operation on fuzzy sets involving many arguments should reflect the existing diversity and return an information granule whose localization and specificity is impacted by the existing data. This gives rise to a slew of constructs such as projection (and the associated reconstruction process), granular composition operators, and granular relational equations. It is also shown that this leads to a new class of granular logic networks built on a basis of G-AND and G-OR neurons.
Conceptually, the results of processing individual inputs, say f(x1), f(x2),..., f(xn) where f is a certain operator, say max, min, t-norm, t-conorm, etc. are considered en block and regarded as a certain information granule. In a concise way, the process is described as T = G( f({x1, x2, ... , xN}) with G denoting a procedure of the principle of justifiable granularity returning an interval information granule T=[
G-Cartesian product
Recall that A(y)= projxR = supxR(x,y). The G-projxR returns an information granule
Likewise
Reconstruction
The reconstruction, as in case of type-1 information granules (see Section 4), is realized by taking a Cartesian product of A and B. Here, however, A and B are type-2 information granules. Thus R itself is a type-2 fuzzy relation
(27) |
and
(28) |
The coverage operator, cov(a, [b,c]) returns 1 once a is included in the interval [b,c]. The performance of reconstruction is expressed as the product , the higher the value of this index, the better the performance of the reconstructed fuzzy relation.
The composition operators discussed in Section 2.2 are examples of multi-argument aggregation operations. the operation of G-min, G-max, G-t, and G-s are introduced as analogous to the discussed compositions in the following way
G-min and G-t compositions
B(y) = G-minx(A(x), R(x,y)) | (29) |
B(y) = G-tx(A(x), R(x,y)) | (30) |
The principle of justifiable granularity is applied to sets {min(A(x1), R(x1,y)), min(A(x2), R(x2,y)),...,min(A(xn), R(xn,y))} and {A(x1)tR(x1,y), A(x2)tR(x2,y),..., A(xn)tR(xn,y)} , respectively.
For the inf-min and inf-s, one has
B(y) = G-maxx(A(x), R(x,y)) | (31) |
B(y) = G-sx(A(x), R(x,y)) | (32) |
With the principle of justifiable granularity applied to sets
{max(A(x1), R(x1,y)), max(A(x2), R(x2,y)),..., max(A(xn), R(xn,y))} | (33) |
{A(x1)sR(x1,y), A(x2)sR(x2,y),..., A(xn)sR(xn,y)}
(34)
The composition operations discussed above, both in terms of their numeric and granular versions, are considered in the generalized versions of relational equations, G-relational equations. As before, there are two types of problems
Estimation problem. Given are fuzzy sets A and B, determine R. In virtue of the G-max or G-t composition, the result of the composition of A and R is
(35) |
and
(36) |
The solution (optimal fuzzy relation) Ropt is the one that maximizes the product of the average coverage and specificity, Ropt =arg maxR (
For the system of equations where pairs of data (Ak, Bk), k =1, 2,..., N are provided, the formulation of the problem is the same as shown above with the coverage and specificity expressed over all data meaning that
(37) |
and
(38) |
In the estimation problem, the data in the form (Ak,
(39) |
where card(.) denotes the cardinality of the information granules.
Inverse problem
Here given are R and B and A has to be determined. The formulation of the optimization problem in which A is searched for where G-min or G-t applied to A and R returns
The solutions to the estimation and inverse problems cannot be obtained analytically. A sound alternative is to involve some population-based optimization methods such as PSO(Particle Swarm Optimization), GA(genetic algorithms), DE(differential evolution) or alike and regard the product of coverage and specificity as a suitable fitness function.
The composition operators form the computing setting of granular neurons. The G-t composition gives rise to a G-OR neurons and G-s composition produces a G-AND neuron; refer to

Fig. 3 Granular logic neurons
In essence, the granular neurons are realized by G-t or G-s composition operators returning a type-2 information granule Y. The weights play an important role by endowing the neurons by some parametric flexibility which is central to the realization of the learning capabilities of the neurons.
G-s composition: if all weights are equal to 1, the original data are used in the principle of justifiable granularity. If the weights are getting lower then the population of inputs is shifted towards lower values and the resulting information granule is moved towards the lower end of the unit interval. For the G-AND neuron, if all weights are set to zero, the obtained information granule is built on a basis of the original data. If the weights are getting larger, the obtained information granule migrates towards higher end of the space. Note that in both classes of neurons, the numeric inputs produce a granular output Y.
The generalization of the logic processor discussed in Refs. [

Fig. 4 Architecture of G logic processor
The outputs of G-AND neurons are information granules. Denote them by Z1, Z2,..., Zm. They are processed by the G-OR neuron. The lower and upper bounds of Zis are processed separately, which at the end gives rise to information granule of type-2 (the level of type of information granule has been elevated). For the purpose of learning, one can treat them as information granule of type-1 use in the performance index the bounds of the granule as depicted in

Fig. 5 Processing in G logic processor(Stressed is an elevated level of information granules obtained when moving along successive layers of the architecture)
To realize learning of the network, two components have to be clearly identified, namely a suitable performance index and a learning algorithm. To focus on this discussion, consider that the learning data are given as a family of input-output pairs (xk, targetk), k=1, 2,..., N with n input variables and a single output. With regard to the first one, because numeric data are confronted with the information granule, the optimization process has to be guided by a suitable performance index that takes into account the diverse nature of data and results of the model.
As we encounter type-2 information granules, two optimization performance indexes are considered. There are
The optimization problem is formulated for
(40) |
and
(41) |
V1= | (42) |
and
(43) |
(44) |
V2 = | (45) |
The solution is wopt = arg maxw V1 or wopt = arg maxw V2 where w is a collection of weights of the G-neurons forming the network.
Given the complexity of the optimized performance index whose gradient with respect to the parameters of the network is difficult to determine, a feasible optimization vehicle comes from a family of population-based optimization techniques.
This study develops a new perspective and provides algorithmic environment of data augmented constructs of relational computing. This enhanced the current consideration from a purely prescriptive ground to embrace the descriptive aspect involving data content and data characteristics. It has been shown that multivariable constructs (projections, compositions, etc.) give rise to results of elevated aspect of information granularity. In particular, the arguments that are membership grades lead to the granular result, say, an interval of possible membership grades. It can be stated that there is an interesting effect of elevation of type of information granules: an aggregation (composition) of type-p information granules produces a single type (p+1) information granule. Granular architectures have been introduced, referred to as G-constructs, say G-projection, G-composition, among others. The idea of fuzzy relational equations is generalized to G-relational equations. As a follow-up of weighted composition operations, a family of granular neurons and neural networks has been established that focuses on numeric processing producing granular results.
While the study opens up a new avenue of research in fuzzy sets, relational calculus, there are a number of promising directions worth pursuing. First, all investigations have been conducted for interval information granules. However, the framework proposed here is for more general and as such deserves more studies. The solutions to the G-relational equations are demanding given the underlying processing thus a way of their determination calls for more thorough investigations as to efficiency of optimization methods.
References
SANCHEZ E.Resolution of composite fuzzy relation equations[J].Information and Control, 1976 (30) :38. [Baidu Scholar]
NOLA DI, SESSA S, PEDRYCZ W,et al.Fuzzy relation equations and their applications to knowledge engineering [M]. Norwell: Kluwer,1989. [Baidu Scholar]
BARGIELA A, PEDRYCZ W. Granular computing: An introduction[M]. Dordrecht: Kluwer Academic Publishers, 2003. [Baidu Scholar]
BARGIELA A, PEDRYCZ W. Toward a theory of Granular Computing for human-centered information processing[J]. IEEE Transactions on Fuzzy Systems, 2008,16(2): 320. [Baidu Scholar]
PEDRYCZ W. Granular computing[M]. Boca Raton: CRC Press, 2013. [Baidu Scholar]
PEDRYCZ W. Granular computing for data analytics: A manifesto of human-centric computing[J]. IEEE /CAA Journal of Automatica Sinica, 2018, 5: 1025. [Baidu Scholar]
ZADEH L A. Towards a theory of fuzzy information granulation and its centrality in human reasoning and fuzzy logic[J]. Fuzzy Sets and Systems, 1997, 90: 111. [Baidu Scholar]
KLIR G J, YUAN B. Fuzzy Sets and fuzzy logic theory and applications[J].Upper Saddle River: Prentice Hall, 1995. [Baidu Scholar]
PEDRYCZ W, GOMIDE F. Fuzzy systems engineering: Toward human-centric computing[M]. Hoboken: John Wiley, 2007. [Baidu Scholar]
CZOGALA E, DREWNIAK J, PEDRYCZ W. Fuzzy relation equations on a finite set[J]. Fuzzy Sets and Systems, 1982, 7:89. [Baidu Scholar]
PEDRYCZ W. Numerical and applicational aspects of fuzzy relational equations[J]. Fuzzy Sets and Systems , 1983, 11:1. [Baidu Scholar]
YANG S J. An algorithm for minimizing a linear objective function subject to the fuzzy relation inequalities with addition-min composition[J]. Fuzzy Sets & Systems, 2014, 255:41. [Baidu Scholar]
YANG X. Solutions and strong solutions of min-product fuzzy relation inequalities with application in supply chain[J]. Fuzzy Sets and Systems, 2020 (384) :54. [Baidu Scholar]
BEZDEK J C. Pattern recognition with fuzzy objective function algorithms[M]. New York: Plenum Press, 1981. [Baidu Scholar]
PEDRYCZ W. An Introduction to computing with fuzzy sets - Analysis, design, and applications[M]. Springer, 2020. [Baidu Scholar]
PEDRYCZ W, HOMENDA W. Building the fundamentals of granular computing: A principle of justifiable granularity[J]. Applied Soft Computing, 2013, 13: 4209. [Baidu Scholar]
PEDRYCZ W. 从局部到全局的规则模型: 粒聚合研究[J]. 同济大学学报(自然科学版): 2021, 49(1): 142. [Baidu Scholar]
PEDRYCZ W. From local to global rule-based models: A study in granular aggregation[J]. Journal of Tongji University (Natural Science), 2021:49(1):142. [Baidu Scholar]
PEDRYCZ W. Allocation of information granularity in optimization and decision-making models: Towards building the foundations of Granular Computing[J]. European Journal of Operational Research, 2014, 232: 137. [Baidu Scholar]