ARS MATHEMATICA CONTEMPORANEA Volume 9, Number 2, Fall/Winter 2015, Pages 145-344 Covered by: Mathematical Reviews Zentralblatt MATH COBISS SCOPUS Science Citation Index-Expanded (SCIE) Web of Science ISI Alerting Service Current Contents/Physical, Chemical & Earth Sciences (CC/PC & ES) The University of Primorska The Society of Mathematicians, Physicists and Astronomers of Slovenia The Institute of Mathematics, Physics and Mechanics The publication is partially supported by the Slovenian Research Agency from the Call for co-financing of scientific periodical publications. ARS MATHEMATICA CONTEMPORANEA Coping with backlog If we consider the quality of our journal there is one weakness that we find more severe than anything else: the backlog. This is the span of time between the moment a paper is submitted up to the time the printed copy of the journal reaches the reader. Each article is assigned three dates: received (R), accepted (A), and published on-line (P). Another important date is the date of printed copy. Usually the on-line version of the paper is available very much earlier than the printed version, and because the electronic version of our journal is freely available (under so-called diamond open access), readers are not hurt by the difference in time between the publication of the electronic and paper versions of the paper. Part of this lag is legitimate and is depends on the technology: it is possible to put papers on-line separately, but one has to wait for the last article to be ready before the whole issue is sent to the printer. When we set up the journal in 2007, our primary concern was to ensure about 20 high-quality papers per year. It was not clear that we would get sufficiently many enthusiasts who would be willing to submit their good papers for publication in an unknown journal with uncertain future. At first we relied on papers arising from conference series such as the 4-yearly Slovenian Graph Theory conference, GEMS, and SIGMAP. The rigidity of special issues, however, proved to be a far greater problem than we initially envisaged. When the journal's visibility increased and its high quality became apparent, the flow of manuscripts increased, and very quickly we had more papers accepted than we needed for a single year. We adopted three strategies for reducing the backlog: (a) We started opening future issues. This moved the problem into the future, and also gave us more flexibility in numbering the papers. But this has to be done with care: once page numbers are assigned to the electronic version of an article, they have to remain the same in the printed version, and so we cannot open the second issue of a given volume until the first one is completed. (b) We increased the number of papers publishable each year (from the initial 20 to 60 or more), by producing two volumes each year, and publishing more articles per issue. (c) We also raised the standards for acceptance, and so now the rate of acceptance of papers has been reduced to 20 per cent. The increased volume of submissions has increased the workload for our editorial staff, and in turn this has increased the length of time taken for the review stage. Next year we will involve more Editors in managing papers. Hopefully this will reduce the backlog at this stage of the process. Nino Bašic, Selena Praprotnik and Gordon Williams have collected and analysed data on all papers published so far in our journal. Back in 2008, it took on average less than 12 months to process a paper for the first issue of our journal. In contrast, it took over 21 months on the average to process a paper for issue Vol. 8, no. 2 in 2015. Our goal is to reduce the average time to under 12 months, as soon as possible. The first figure (above) is a stacked plot showing the cumulative average processing time by issue: green in the bottom is the time from receipt to accepted (RA), blue at the top is the additional time from accepted to published online ( AP), and the thick black line indicates the total time from receipt to published online (RP). The second figure (above) gives plots of receipt to accepted (RA), accepted to published online ( AP) and total time from receipt to published online (RP), with linear regression lines superimposed. Unfortunately, the length of time from accepted to published on-line continues to grow as well, so our policy of slow growth and opening additional volumes is not enough. We hope that moving from 20 papers per year to the current 60 papers per year, better handling of special issues, and further management of the acceptance rate, will stabilise the backlog at an appropriate level. Dragan Marašič and Tomaž Pisanski Editors In Chief Contents Construction of planar 4-connected triangulations Gunnar Brinkmann, Craig Larson, Jasper Souffriau, Nico Van Cleemput . . 145 On minimal forbidden subgraphs for the class of EDM-graphs Gašper JakliC, Jolanda Modic.........................151 Sparse line deletion constructions for symmetric 4-configurations Leah Wrenn Berman, William H. Mitchell..................165 Bounds on the domination number of Kneser graphs Patrie Östergärd, Zehui Shao, Xiaodong Xu.................187 The expected values of Kirchhoff indices in the random polyphenyl and spiro chains Guihua Huang, Meijun Kuang, Hanyuan Deng................197 Extending patches to fullerenes Christina M. Graves, Jennifer McLoud-Mann, Kristen Stagg Rovira .... 209 Fast recognition of partial star products and quasi cartesian products Marc Hellmuth, Wilfried Imrich, Tomas Kupka...............223 Rational sums of hermitian squares of free noncommutative polynomials Kristijan Cafuta, Igor Klep, Janez Povh ...................243 On mixed discriminants of positively definite matrix Chang-Jian Zhao, Xiao-Yan Li........................261 Odd edge coloring of graphs Borut Lužar, Mirko Petruševski, Riste Škrekovski..............267 Edge looseness of plane graphs Julius Czap..................................279 Levels in bargraphs Aubrey Blecher, Charlotte Brennan, Arnold Knopfmacher .........287 What can Wikipedia and Google tell us about stock prices under diferent market regimes? Boris Cergol, Matjaž Omladic........................301 Subdivision into i-packings and S-packing chromatic number of some lattices Nicolas Gastineau, Hamamache Kheddouci, Olivier Togni.........321 Volume 9, Number 2, Fall/Winter 2015, Pages 145-344 ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 145-149 Construction of planar 4-connected triangulations Gunnar Brinkmann Applied Mathematics, Computer Science and Statistics, Ghent University, Krijgslaan 281 S9, 9000 Ghent, Belgium Craig Larson Department of Mathematics and Applied Mathematics, Virginia Commonwealth University, 4106 Grace E. Harris Hall, 1015 Floyd Avenue, Richmond, VA 23284-2014 Jasper Souffriau Applied Mathematics, Computer Science and Statistics, Ghent University, Krijgslaan 281 S9, 9000 Ghent, Belgium Nico Van Cleemput Applied Mathematics, Computer Science and Statistics, Ghent University, Krijgslaan 281 S9, 9000 Ghent, Belgium Received 20 March 2013, accepted 14 August 2013, published online 21 November 2014 In this article we describe a recursive structure for the class of 4-connected triangulations or - equivalently - cyclically 4-connected plane cubic graphs. Keywords: Planar triangulation, cubic graph, generation, recursive structure. Math. Subj. Class.: 05C10, 05C30, 05C75 Introduction A recursive structure for a class C of graphs is a base set B C C of initial graphs together with a set of operations on graphs that transform a graph in C to another graph in C so that each graph in C can be constructed from a graph in B by a sequence of these operations. E-mail addresses: Gunnar.Brinkmann@UGent.be (Gunnar Brinkmann), clarson@vcu.edu (Craig Larson), Jasper.Souffriau@UGent.be (Jasper Souffriau), Nicolas.VanCleemput@UGent.be (Nico Van Cleemput) Abstract An operation is typically the replacement of a finite substructure by another - larger -substructure. In the ideal case, the set B as well as the set of operations are finite and small. All graphs discussed in this article are simple. The two main applications for recursive structures are structure generation programs and inductive proofs, where the recursive structures describe the induction step. In this paper we discuss planar triangulations - that is plane graphs where every face is a triangle. For several classes of triangulations, recursive structures have been published: for all tri-angulations (that is: 3-connected triangulations) [6], for 5-connected triangulations [1][5], for triangulations with minimum degree 4 [2], for 3- and 4- connected triangulations with minimum degree 5 [3], and for Eulerian triangulations [2]. In the dual, these are constructions for 3-connected planar cubic graphs, cyclically 5-connected planar cubic graphs, 3-connected planar cubic graphs with girth 4, 3- resp. cyclically 4-connected planar cubic graphs with girth 5 and 3-connected bipartite planar cubic graphs. In this article we will add the missing link between 3-connected triangulations and 5-connected triangulations and give a recursive structure for 4-connected triangulations. The operations necessary to construct all 4-connected triangulations are in fact the same as the ones used in [4] to construct all triangulations with minimum degree 4 - except for the operation inducing separating triangles. While it is obvious that an operation introducing separating triangles does not lead to 4-connected triangulations, it is not obvious that all 4-connected triangulations can be obtained with the remaining two operations. □ O5 4 5 e Figure 1: Two of the operations used by Eberhard [6] to generate all triangulations. Edges and vertices outside of the bounding 4-, or 5-cycle in the figure are not drawn. Two of the operations given by Eberhard to construct all triangulations are given in Figure 1. We will show: Theorem 0.1. The class C4 of all 4-connected triangulations can be generated from the octahedron graph (depicted in Figure 2) by operations O4 and O5. O Proof. We will write C4 for the class C4 without the octahedron graph. The operations O4 and O5 are in fact similar to special cases of the edge expansion operation used by Batagelj in [2]. This can best be seen when looking at the reduction -that is the inverse of the construction operation. If one compresses the edges marked with an x (that is: removes the edge and identifies the endpoints) in Figure 1, the resulting graph is the same as after replacing the vertices and their adjacent edges by one, resp. two edges. To prove this theorem, note first that in a triangulation being 4-connected is equivalent to not having a separating - that is: non-facial - 3-cycle. We will show that for each element of the class C4 an inverse operation can be applied that does not introduce separating 3-cycles and therefore leads to an element of C4. This is the consequence of 3 observations; (a) In a 4-connected triangulation no two edges in the same facial triangle belong to the same separating 4-cycle. This follows immediately as in that case the other edges of the separating 4-cycle together with the third edge of the triangle would form a separating 3-cycle. (b) In a 4-connected triangulation that is not the octahedron graph, no two edges in the same facial triangle with a common vertex v of degree 4 belong to different separating 4-cycles C, C'. Suppose that this was the case. Then - due to (a) - the two separating 4-cycles must cross each other and there is an edge {v, y1} belonging to (w.l.o.g.) C so that the next edges {v, x1}, {v, x2} in counterclockwise, resp. clockwise direction around v belong to the separating 4-cycle C' formed by the vertices x1, v, x2, a. This situation is depicted in Figure 3. From the previous observation it follows that C cannot contain x1 or x2, so the Jordan curve theorem gives that it must contain a and that the situation is as with the dotted edges in Figure 3. This implies the presence of 8 triangles which must all be facial triangles - as no non-facial triangles exist - and implies that there are no more edges than those depicted. So the graph was the octahedron graph. (c) In a 4-connected triangulation without vertices of degree 4, for each edge {v, x1} con- taining a vertex v of degree 5 that belongs to a separating 4-cycle C, either the previous or the next edge in the cyclic order around v or both do not belong to a separating 4-cycle. a x 2 Figure 3: Two separating 4-cycles crossing in a vertex of degree 4. By choosing the neighboring edge as the one that shares a triangle with both edges of C containing v, we can follow the same line of arguments as before to get - up to symmetry -the situation in Figure 4. In this case we don't have 8 triangles, but we do have the triangles (a,yi,x2), (xi, yi,a), (v,yi,xi) and (x2,yi,v) which must all be facial. This implies that the degree of yi is 4 - contradicting the assumption. Figure 4: Two separating 4-cycles crossing in a vertex of degree 5 in a triangulation with minimum degree 5. As in a 4-connected triangulation there are always vertices with degree 4 or degree 5, (a),(b),(c) together imply that a triangulation in C4 contains an edge adjacent to a vertex of degree 4 or 5 that does not lie on a separating 4-cycle. Using this edge as the edge x in Figure 1 we can reduce such a triangulation to a smaller one without separating triangles. a □ References [1] D. Barnette, On generating planar graphs, Discrete Math. 7 (1974), 199-208. [2] V. Batagelj, An improved inductive definition of two restricted classes of triangulations of the plane, Combinatorics and Graph Theory, Banach Center Publications 25 (1989), 11-18. [3] G. Brinkmann and B. D. McKay, Construction of planar triangulations with minimum degree 5, Discrete Math. 301 (2005), 147-163. [4] G. Brinkmann and B. D. McKay, Fast generation of planar graphs, MATCH Commun. Math. Comput. Chem. 58 (2007), 323-357, see http://cs.anu.edu.au/~bdm/index.html. [5] J. W. Butler, A generation procedure for the simple 3-polytopes with cyclically 5-connected graphs, Can. J. Math. 26 (1974), 686-708. [6] V. Eberhard, Zur Morphologie der Polyeder. Teubner, 1891. /^creative ^commor Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 151-163 On minimal forbidden subgraphs for the class of EDM-graphs Gašper JakliC FMF and IMFM, University of Ljubljana, Jadranska 19, 1000 Ljubljana, Slovenia and IAM, University ofPrimorska, Slovenia Jolanda Modic * XLAB d.o.o., Pot za Brdom 100, 1000 Ljubljana, Slovenia and FMF, University of Ljubljana, Slovenia Received 2 April 2013, accepted 2 April 2014, published online 21 November 2014 Abstract In this paper, a relation between graph distance matrices and Euclidean distance matrices (EDM) is considered. Graphs, for which the distance matrix is not an EDM (NEDM-graphs), are studied. All simple connected non-isomorphic graphs on n < 8 nodes are analysed and a characterization of the smallest NEDM-graphs, i.e., the minimal forbidden subgraphs, is given. It is proven that bipartite graphs and some subdivisions of the smallest NEDM-graphs are NEDM-graphs, too. Keywords: Graph, Euclidean distance matrix, distance, eigenvalue. Math. Subj. Class.: 15A18, 05C50, 05C12 1 Introduction A matrix D e Rnxn is Euclidean distance matrix (EDM), if there exist xi, x2,..., xn € Rr, such that dj = ||xj - Xj ||2, i,j = 1,2 ,...,n. The minimal possible r is called the embedding dimension (see [2], e.g.). * Corresponding author. This research was funded in part by the European Union, European Social Fund, Operational Programme for Human Resources, Development for the Period 2007-2013. The author would like to thank XLAB d.o.o., Dr. Daniel Vladušic and Dr. Gregor Berginc for all the support. E-mail addresses: gasper.jaklic@fmf.uni-lj.si (Gasper Jaklic), jolanda.modic@gmail.com (Jolanda Modic) Euclidean distance matrices were introduced by Menger in 1928 and have received a considerable attention. They were studied by Schoenberg [13], Young and Householder [14], Gower [4], and many other authors. In recent years many new results were obtained (see [5, 7, 8, 11] and the references therein). They are used in various applications in linear algebra, graph theory, geodesy, bioinfor-matics, chemistry, e.g., where frequently a question arises, what can be said about a set of points, if only interpoint distance information is known. Some examples can be found in [2]. EDMs have many interesting properties. They are symmetric, hollow (i.e., with only zeros on the diagonal) and nonnegative. The sum of their eigenvalues is zero and they have exactly one positive eigenvalue (for a nonzero matrix). Schoenberg ([13]), Hayden, Reams and Wells ([5]) gave the following characterization of EDMs. Theorem 1.1. Let D e Rnxn be a nonzero symmetric hollow matrix and let e e Rn be the vector of ones. The following propositions are equivalent: (a) The matrix D is EDM. (b) For all x e Rn such that xT e = 0, xTDx < 0. (c) The matrix D has exactly one positive eigenvalue and there exists w e Rn such that Dw = e (1.1) and wTe > 0. Throughout the paper we will use the notation e for the vector of ones of appropriate size. Vectors ež will denote the standard basis. Let G be a graph with a vertex set V (G ) and an edge set E (G ). Let the distance d(u, v) between vertices u,v e V (G) be defined as their graph distance, i.e., the length of the shortest path between them. Let G := [d(u, v)]„jVeV(£) be the distance matrix of G. If the graph distance matrix of a graph is EDM, the graph is called an EDM-graph. Otherwise the graph is a NEDM-graph. Graph distance matrices of EDM-graphs were studied in several papers. Path and cycles were analysed in [9]. Star graphs and their generalizations were considered in [6, 10]. Some results on Cartesian products of EDM-graphs are also known (see [11]). However, the characterization of EDM-graphs in general is still an open problem. In this paper, all simple connected non-isomorphic graphs on n < 8 nodes are analysed and a characterization of the smallest NEDM-graphs, i.e., the minimal forbidden subgraphs, is given. In algebraic graph theory, a lot is known on the adjacency matrix and the Laplacian matrix of a graph. Many results on their eigenvalues exist, but not much is known on the graph distance matrix. Hopefully, this paper will provide a deeper insight into the relation between general graphs or networks and EDM theory. There are some interesting possibilities of application. Molecular conformation in bioinformatics, dimensionality reduction in statistics, 3D reconstruction in computer vision, just to name a few. The structure of the paper is as follows. In Section 2, all NEDM-graphs on n < 8 nodes are considered. Analysis of their properties enables us to find some larger NEDM-graphs, which are presented in sections 3 and 4. A proof that bipartite graphs are NEDM-graphs is given. We present two families of subdivision graphs of the smallest NEDM-graphs that are NEDM-graphs, too. There exist graphs, for which the system (1.1) has no solution. Such graphs are studied in Section 5. The paper is concluded with an example, where we show that not all subdivisions of graphs result in NEDM-graphs. 2 The smallest NEDM-graphs In this section we consider simple connected non-isomorphic graphs on n < 5 nodes and find the smallest NEDM-graphs. There is one simple connected graph on 2 nodes, the path graph P2, and there exist only two simple connected graphs on 3 nodes, the path graph P3 and the cycle graph C3. In [9] it was proven that path graphs and cycle graphs are EDM-graphs. For n = 4, there are 6 simple connected graphs (see Fig. 1). First four of them are the star graph S4, the path graph P4, the cycle graph C4 and the complete graph K4, respectively, which are EDM-graphs (see [9, 10]). Therefore we only need to consider the last two graphs, g45) and g46). Figure 1: Simple connected graphs on 4 nodes. Let us denote vertices of graphs g45) and g46) counterclockwise by 1,2, 3 and 4 starting with the upper right vertex. The characteristic polynomials of the corresponding graph distance matrices G (5) 0 1 2 2 0 1 2 1 1 0 1 1 and g46) = 1 0 1 1 2 1 0 1 2 1 0 1 2 1 1 0 1 1 1 0 are pG(5) (A) = (A + 1)(A3 - A2 - 11A - 7), G4 pG(6) (A) = (A + 1)(A + 2)(A2 - 3A - 2). Thus matrices g45) and g46) have eigenvalues o„(5) = {4.1, -0.7, -1, -2.4} and a„(6) = GG 3 + V17 3 - V17 1, 2 (5) Eigenvalues for G4 were calculated numerically. Exact values can be calculated by using Cardano's formula. 2 2 One can easily verify that vectors wn(5) = [3/7, -1/7, 2/7, 2/7]T and w. ,(6) = [1/2, 0, 1/2, 0]T satisfy the equation G^'w = e, i = 5, 6. Since wT(i) e > 0, i = 5, 6, by Theorem 1.1 graphs G(5) and G(6) are EDM-graphs. Thus there are no NEDM-graphs on 4 nodes. In the case n = 5, there are 21 simple connected graphs (see Fig. 2). V J 13 20 21 Figure 2: Simple connected graphs on 5 nodes. Graphs Gf', i < 5, are the path graph P5, the cycle graph C5, the complete graph K5, the star graph S5 and the tree T5, respectively. Since they are EDM-graphs (see [1]), we (i) only need to analyse graphs G5 , i = 6,7,..., 21. A straightforward calculation shows that the graph distance matrix G5i) of the graph G (i) 6,7,..., 19, has exactly one positive eigenvalue and that there exists w e '. such that G^5iwc(i) = e and wT e > 0. By Theorem 1.1, graphs gf6, G^7, ■ g519) are EDM-graphs. We are left with graphs g520) and g521) (see Fig. 3). The characteristic polynomials of the corresponding graph distance matrices G(20) = 0 2 2 1 1 0 2 2 1 1 to 0 2 1 1 and g521) = 2 0 1 1 1 to 2 0 1 1 2 1 0 1 1 1 1 1 0 2 1 1 1 0 2 1 1 1 2 0 1 1 1 2 0 are pG(20) (A) = -(A + 2)3(A2 - 6A + 2), g5 PG(21) (A) = -(A + 1)(A + 2)(A3 - 3A2 - 12A + 2). 4 20 Figure 3: The graphs g520) and g521). Thus matrices g520) and g5,21) have spectra an(20) = {3 + V7, 3 -V7, -2, -2, -2} and an(21) = {5.2, 0.2,-1, -2, -2.4} . (21) Exact eigenvalues for G5 ) can be calculated by using Cardano's formula. Here they were calculated numerically. Since matrices G^20) and g5,21) have two positive eigenvalues, graphs g520) and g521) are NEDM-graphs. These are the smallest NEDM-graphs. An induced subgraph H of a graph G is a subset of the vertices V(G) together with all edges whose endpoints are both in this subset. Proposition 2.1. Let G be a simple connected graph and let H be its induced subgraph. If H is a NEDM-graph, the graph G is a NEDM-graph as well. Proof. Let n and m < n, denote the number of nodes in graphs G and H, respectively. Let us order vertices of the graph G in such a way that the first m vertices are the vertices of the graph H. Thus the distance matrix G of the graph G is of the form G H * where H is the distance matrix of the graph H. Every principal submatrix of an EDM has to be an EDM as well. Thus since H is not an EDM, neither is G. Therefore G is a NEDM-graph. □ All NEDM-graphs form a set of forbidden subgraphs of the class of EDM-graphs. Graphs g520) and g521) are the minimal forbidden subgraphs. All minimal forbidden subgraphs on 6 and 7 nodes can be seen in Fig. 4 and Fig. 5. Figure 4: NEDM-graphs for n = 6. Let m(n) be the number of NEDM-graphs on n nodes and let mnew(n) be the number of NEDM-graphs on n nodes for which none of the induced subgraphs is NEDM-graph. 10 Figure 5: NEDM-graphs for n = 7. We denote the number of non-isomorphic simple connected graphs on n nodes by g(n). Table 1 shows how numbers m(n) and mnew (n) grow with n. The calculations were done in the following way. By using program geng in Nauty ([12]) we generated all simple connected non-isomorphic graphs on n < 8 nodes. Then we applied Theorem 1.1 to determine whether a graph is an EDM-graph. Computations were done in Mathematica. n g(n) m(n) mnew (n) 5 21 2 2 6 112 27 3 7 853 341 13 8 11117 7946 48 Table 1: Number of NEDM-graphs compared to the number of all graphs on n nodes. 3 Bipartite graphs A quick observation shows that the graph g520) is bipartite (see Fig. 3). Let GUk,zn-k be a simple connected bipartite graph on n > 5 nodes, whose vertices are divided into two disjoint sets Uk = {u1, u2,..., uk}, Zn-k = {uk+1,uk+2,..., un}, к = 2,3,..., n - 2, such that every edge connects a vertex in Uk to a vertex in Zn-k (see Fig. 6). The sets Uk and Zn-k are called the partition sets. A graph join Gi + G2 of graphs Gi and G2 with disjoint vertex sets V(Gi), V(G2) and edge sets E (G1), E (G2) is the graph with the vertex set V (G1) U V (G2) and the edge set E (Gi) UE (G2 ) U {(u, v); u G V (Gi), v e V (G2)}. It is the graph union Gi U G2 with all the edges that connect the vertices of the first graph with the vertices of the second graph. The graph G Uk ,zn-k can also be written as the graph join of two empty graphs on к and n - к vertices, i.e., GUk,zn-k = Ok + On-k. The corresponding graph distance matrix is Gk.n—h 2(Ek,k — Ik ) Ek ,n-k En-k,k 2(En-k,n-k In-k) G Rn where Ep,q g Rpxq and Ip G Rpxp are the matrix of ones and the identity matrix, respectively. Uk U3 U2 Ul Un Uk 2 Uk l Figure 6: The graph Guk,zn Theorem 3.1. A simple connected bipartite graph GUk,Zn-k on n > 5 nodes and with partition sets Uk and Zn_k is a NEDM-graph. Proof. Since graphs GUk,Zn-k and GUn-k,zk are isomorphic, it is enough to see that the theorem holds true for k = 2, 3,..., |_n/2j. Let us analyse the eigenvalues of the graph distance matrix of G Uk,Zn-k. A simple computation shows that ui,i = [ej - ef, 0T]T solves the equation Gk,n_kui,i = -2ui,i for all i = 2, 3,..., k, and that u2,j = [0T, ej - ej]T, solves the equation Gk,n_ku2,j = -2u2,j for all j = 2,3,... ,n — k. Therefore Gk,n_k has an eigenvalue -2 with multiplicity n - 2. Now let us take u = [a eT, eT]T. The relation Gk,n_ku = Au yields the system of equations 2(k - 1)a + n - k = Aa, ka + 2(n - k - 1) = A, which has solutions 2k - n ± v7(n - 2k)2 + k(n - k) ai,2 =-k-, Ai,2 = n - 2 ± y/(n - 2k)2 + k(n - k). Relations n > 5 and 2 < k < |n/2j imply that ai 2 and Ai 2 are well-defined. Since Ai > 0 and Ai • A2 = 3(k - 2)(n - 2 - k) + 2(n - 4) > 0, we conclude that A2 > 0. Thus, by Theorem 1.1, the graph Guk,Zn-k is a NEDM-graph. ' n □ Remark 3.2. For k = 1, the graph GUk,Zn-k is the star graph Sn, which is an EDM-graph. 4 Graph subdivision Let G be a graph. A subdivision of an edge in G is a substitution of the edge by a path. For example, an edge of the cycle Cn can be subdivided into three edges, resulting in the cycle graph Cn+2. Recall the NEDM-graph g520). It contains a 4-cycle c connecting nodes 2, 3, 4 and 5 (see Fig. 7). We can construct larger NEDM-graphs by performing a subdivision of the cycle c. Let G^0 be a graph on n nodes, obtained by subdividing the cycle c in the graph g520) as seen in Fig. 7. Such graphs are g520) Fig. 5). ' G^ and g5270) g74) (see Fig. 4 and 20 5,n Figure 7: Construction of graphs G5 (20) Let ei denote the standard basis and let Cn be the graph distance matrix of the cycle graph Cn (see [9]). The matrix Cn is a circulant matrix (see [3]), generated by its first row: n — 1 n — 1 n — 3 0, 1, ...,-, -, -,..., 1, n odd, '' 2 2 2 n — 2 n n — 2 0, 1, ...,-, —, -,..., 1, n even. ' ' ' 2 ' 2' 2 ' ' ' We will use the notation сПм) for the (i, j)-th element of the matrix Cn. The structure of the matrix Cn implies сП1'2) = сП2'3) = 1, сП1'3) = 2, n > 4, (4.1) and ( L(n — 1)/2j, i =1, cna(n+4)/2J) = J [n/2j, i = 2, n > 3. \ L(n — 2)/2j, i = 3, Theorem 4.1. Graphs g5V , n > 5, are NEDM-graphs. Proof. The graph distance matrix of the graph G^, n > 5, is (4.2) G(20) = G5,n = 0 (Cn-1 + 21>2 eT (Cn-1 +21 )' C^ 1 By Theorem 1.1 it is enough to show that there exists x g Rn, such that xTe = 0 and xT G52n)x > 0. Let us take x = [-yTe, yT]T, with nr(-ei + e2 - e^ + e(n+s)/2, n odd, nr(-ei + (e2 - es)) + e(n+2)/2, n even. We will show that xTG52n)x = yTCn-iy - 2(yTe) (eTCn-iy + 2(yTe2)) > 0. (4.3) From Г ^, n odd, Г ■, n odd, yTe = I , n even, and ^e2 = \ , n even, it follows that T (20) T i (n - 3)(eTCn-iy + n - ^, n odd, xTofnx = yTCn-iy +< 2, „\ „ , L ^ (4.4) T5'" y n-iy '' n!-M(eTC„-iy + , neven. Firstly, let n be odd. Terms in the relation (4.4) simplify to yTC y = (n - 1)2 (C(i,2) C(i,S) + C(2,S)j y Cn-iy =--2-^Cn-i - Cn-i + Cn-i J - _ (n - 1) (C(i.("+S)/2) - с(2,(n+S)/2) + с(S,(n+S)/2) j eTC y = C(2,(n+S)/2) n - 1 (C(i,2) + C( e2 Cn-1y = C„-i --^ (^n-i + Cn By (4.1) and (4.2), yTCn-iy = - ————, e'T Cn-iy = and .Tr< (n - 1)(n - 5) ^^ n - 1 '* Cn-iy =--^-, e2 Cn-iy =--, T (20) x G5,n)x = n -1 which satisfies the requirement (4.3) for all n > 5. When n is even, the terms in the relation (4.4) simplify to yTCn-iy = - (n - ffi - 1) ((n - 2)(nCni!i) - nCn-3 + (n - 1)^) + + 2nCn-(in+2)/2) - 2(n - 1)(Cn-(in+2)/2) - Cn-(in+2)/2)^ , eTC y = n - 1 C(2,(n+2)/2) n - 2 (C(i,2) + n - 1 C(2,3) j eT Cn-iy = ~Cn-i - — (^n-i + ~Cn-i ) . By (4.1) and (4.2), yTC y (n - 1)2(n - 2)(n - 4) eTC y n - 2 y Cn-iy =--2n-, eT Cn-iy = — and xT G(20)x _ n - 2 which satisfies the requirement (4.3) for all n > 5. □ (21) Similarly, we can subdivide cycles of the graph £5 and produce NEDM-graphs (see Fig. 8). The graph g521) contains a 3-cycle c connecting nodes 3, 4 and 5. Let £52П) be a graph on n nodes, obtained by subdividing the cycle c in the graph g521) as seen in Fig. 8. Such graphs are ^ _ g521), ^ _ G(63) and g5271) _ Gr6) (see Fig. 4 and Fig. 5). 21 5,n Figure 8: Graphs £5^. Theorem 4.2. Graphs gf!, n > 5, are NEDM-graphs. Proof. The graph distance matrix of the graph g52^), n > 5, is "0 1 T uT G(21) _ G5,n _ 1 0 T u v Cn—2 where 1 + (-1)n u _ (Cn-2 + I)e2 + e - y----e(„+2)/2, v _ (Cn-2 + I)e2 + y, and y _ ^i(=2+1)/2 ek. Analogous to the proof of Theorem 4.1, we can show that for x _ a [a, -a, zT] , where 2 n —3 2 ' n odd, n even, and z n—1 e1 - 3e2 - e(n+1)/2, n odd, n_2 n_4 ■e1 - 4e2 - e(n+2)/2, n even, the expression xTG^x _ zTCn—2z + 2a (uTz - vTz - a) is positive. Relations Ci—22) _ 1, C(1,(n+1)/2) _ C(2,(n+1)/2) n — 3 C (1,(n+2)/2) n—2 n - 4 2 , C n—2 n—2 (2,(n+2)/2) _ n - 2 n— 2 a imply and T u z which yields 2, n odd, 1, n even, Z Cn-2Z = T v z 3 — n, n odd, 4 — n, n even, (n-3)(n+1) лл --2 + ), n odd, (n-2)(n-4) 2 , n even, xTG5,n x 2, n odd, □ Thus by Theorem 1.1, the matrix Gj2^ is a NEDM. 5 Systems with no solution When verifying whether a graph G with the corresponding graph distance matrix G is an EDM-graph, by Theorem 1.1 one can check if there exists a solution of the equation Gw = e, such that wT e > 0. For n > 7 there exist graphs, for which the equation Gw = e has no solution. Let Gk,n-k be the graph join of a complete graph Kk and an empty graph On-k, n > 7, k = 2,3,.'..,n — 3, i.e., Gk,n-k = Kk + On-k • The graph Kk contains vertices 1,2,..., k and the graph On-k contains vertices k +1, k + , n. Thus the corresponding graph distance matrix is 2 Gk2n k Ek,k — Ik E k2n - k En— k,k 2(En-k,n-k Jn-k) For n = 7 and k = 3 the equation G3 4 w = e has no solution since the ranks of the matrix G3,4 and its augmented matrix [G3,4|e] are different, rank(G3,4) = 6 and rank([G3,4|e]) = 7. The same holds true if n = 7 and k = 4. Thus by Theorem 1.1 matrices G3 4 and G4 3 are not EDMs. On the other hand, for n = 8 the equation Gk,8-k w = e has a solution for all k g {3,4,5}. In general, the matrix Gk,n-k is a NEDM. Theorem 5.1. The graph Kk + On-k, n > 7, k = 2,3,..., n — 3, is a NEDM-graph. Proof. Let Gk,n-k be the graph distance matrix of the graph Kk + On-k, n > 7, k = 2, 3, . . . , n - 3. For k = 2 we take 1 2 1 T w = -[4 — n, 4 — n, 1,1,..., 1] . We can verify that G2,n-2w = e and wTe = (6 — n)/2 < 0. Thus by Theorem 1.1 the matrix G2,n-2 is a NEDM. Now let k = 3, 4, . . . , n - 3. For n = 7 the proof has already been done above. For n > 8 let u = [a eT, eT 1T, where vectors e are of sizes k and n — k, respectively. The relation Gk,n-ku = Au yields the system of equations a(k — 1) + n — k = Aa, ak + 2(n — k — 1) = A, 2, n even. with solutions ai,2 = 2k - 2n +1 ± V4(n - k)(n - k - 1) + (k + 1)2) Ai,2 = 1 (2n - k - 3 ± V4(n - k)(n - k - 1) + (k + 1)2) . Relations n > 8 and 3 < k < n - 3 imply that a1,2 and A1,2 are well-defined. Since A1 > 0 and A1 • A2 = (n - 3 - k)(k - 3) + n - 7 > 0, we conclude that A2 > 0. Thus, by Theorem 1.1, graph Kk + On-k is aNEDM-graph. □ Remark 5.2. For k = 1 and k = n - 1, the graphs Kk + On-k are the star graph Sn and the complete graph Kn, respectively, which are EDM-graphs. Remark 5.3. For k = n - 2, the graph Kn-2 + O2 is an EDM-graph. The graph distance matrix Gn-2,2 has eigenpairs (-2, [0T, eT - eT ]T ) , (-1, [eT - eT, 0T ]T ) , i = 2, 3,...,n - 2, and with (a1,2, [«1,2 eT, eT]Tj «1,2 n - 5 ± Vn2 - 2n + 9 2(n - 2) and A12 - 1 ±yjn2 - 2n + 9 The eigenvalue A1 is obviously positive. From A1 • A2 = -2 it follows that A2 < 0. One can easily verify that w = (1/2) [0T, eT] solves the equation Gn-2,2w = e. Since wTe = 1, Theorem 1.1 implies that Gn-2 2 is EDM. 6 Conclusion In Section 4 we studied subdivisions of graphs. Not all graph subdivisions result in NEDM- graphs. Consider subdividing graph £5 sponding graph distance matrix (20) as in Fig. 9 and denoting it by H. The corre- 0122321 1012212 2101222 H = 2 2 10 12 1 3221012 2122103 1221230 has eigenvalues aH = {10.4,0, -0.2, -0.6, -2.2, -3.4, -4}, which were calculated numerically. Exact eigenvalues could be obtained using Cardano's formula. One can easily verify that vector wh = [1/2, -1/2, 1/2, -1/2, 1/2, 0, 0]T solves the equation HwH = e. Since wHe 1/2, by Theorem 1.1 the graph H is an EDM-graph. 2 5 6 20 5 Figure 9: A subdivision of the graph g520). References [1] R. Balaji and R. B. Bapat, Block distance matrices, Electron. J. Linear Algebra 16 (2007), 435-443. [2] J. Dattorro, Convex Optimization and Euclidean Distance Geometry, Meboo, 2008. [3] G. H. Golub and C.F. van Loan, Matrix computations, Johns Hopkins Studies in the Mathematical Sciences, third ed., Johns Hopkins University Press, 1996. [4] J. C. Gower, Euclidean distance geometry, Math. Sci. 7 (1982), 1-14. [5] T. L. Hayden, R. Reams and J. Wells, Methods for constructing distance matrices and the inverse eigenvalue problem, Linear Algebra Appl. 295 (1999), 97-112. [6] G. Jaklic, J. Modic, On properties of cell matrices, Appl. Math. Comput. 216 (2010), 20162023. [7] G. Jaklic, J. Modic, A note on "Methods for constructing distance matrices and the inverse eigenvalue problem", Linear Algebra Appl. 437 (2012), 2781-2792. [8] G. Jaklic and J. Modic, Inverse eigenvalue problem for Euclidean distance matrices of size 3, Bull. Aust. Math. Soc. 87 (2013), 82-93. [9] G. Jaklic and J. Modic, On Euclidean distance matrices of graphs, Electron. J. Linear Algebra 26 (2013), 574-589. [10] G. Jaklic, J. Modic, Euclidean graph distance matrices of generalizations of the star graph, Appl. Math. Comput. 230 (2014), 650-663. [11] G. Jaklic, J. Modic, Cartesian products of EDM-graphs, submitted. [12] B. D. McKay, Practical Graph Isomorphism, Congr. Numer. 30 (1981), 45-87. [13] I. J. Schoenberg, Metric spaces and positive definite functions, Trans. Amer. Math. Soc. 44 (1938), 522-536. [14] G. Young, A. Householder, Discussion of a set of points in terms of their mutual distances, Psychometrika 3 (1938), 19-22. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 165-186 Sparse line deletion constructions for symmetric 4-configurations Leah Wrenn Berman University of Alaska Fairbanks, 513 Ambler Lane, Fairbanks, AK 99775, USA William H. Mitchell University ofWisconsin - Madison, 480 Lincoln Dr., Madison, WI 53706, USA Received 15 April 2013, accepted 24 May 2014, published online 21 November 2014 A 4-configuration is a collection of points and lines in the Euclidean plane such that each point lies on four lines and each line passes through four points. In this paper we introduce a new family of these objects. Our construction generalizes a 2010 result of Berman and Grunbaum in which suitable 4-configurations from the well-understood celestial family are altered to yield new configurations with reduced geometric symmetry groups. The construction introduced in 2010 removes every other line of a symmetry class from the celestial configuration; here we we give conditions under which every p-th line can be removed, for p e {2, 3,4, • • • }. The geometric symmetry groups of the new configurations we obtain are of correspondingly smaller index as subgroups of the symmetry group of the underlying celestial configuration. These sparse constructions can also be repeated and combined to yield a rich variety of previously unknown 4-configurations. In particular, we can begin with a configuration with very high geometric symmetry—the dihedral symmetry of an m-gon for m quite large—and produce a configuration whose only geometric symmetry is 180° rotation. Keywords: Configurations. Math. Subj. Class.: 05B30, 51E30, 05C60 E-mail addresses: lwberman@alaska.edu (Leah Wrenn Berman), whmitchell@math.wisc.edu (William H. Mitchell) Abstract 1 Introduction An n-configuration is a set of n points and n lines with the property that each point lies on n lines and each line passes through n points. Configurations can be investigated as geometric objects or more generally as combinatorial objects where the lines are abstract sets of points. In this work we take the geometric perspective and consider points and lines in the Euclidean plane. Although such geometric objects were studied in the 19th century and several theorems on 3-configurations were proved, no illustration of a geometric 4-configuration appeared in print until much more recently, in [4]. Since then many more examples have been introduced. In this paper we give a technique that produces a large new class of 4-configurations, including 4-configurations with very few symmetries. We emphasize that by a symmetry of a configuration we mean an isometry of the plane which maps the configuration to itself, as opposed to the more general notion of combinatorial symmetry. The collection of symmetries of a configuration, or its symmetry group, partitions the points and lines into orbits, called the symmetry classes of points, and the symmetry classes of lines. One frequently studied class of 4-configurations is the celestial family. Its members have the property that every point lies on exactly two lines from each of two symmetry classes of lines, and every line is incident with two points from each of two symmetry classes of points. Figure 2 gives an example of a celestial 4-configuration. The first published 4-configuration, in [4], was of this class, and more examples appeared in [6]. The first discussion of celestial configurations as a family appeared in a paper called Polycyclic Configurations by Marko Boben and Tomaž Pisanski [2], where they were investigated as a particular class of polycyclic 4-configurations. Branko Griinbaum's 2009 monograph Configurations of Points and Lines [3] gives a detailed analysis of the construction method and theory for celestial 4-configurations. In that reference Grünbaum refers to them as k-astral 4-configurations. However, he also uses the term "k-astral" to describe configurations which have k symmetry classes of points and k symmetry classes of lines; while celestial 4-configurations have this property, there are many other 4-configurations with this property that are not celestial. We reserve the term "k-astral" for the more general class of configurations with k symmetry classes of points and lines, and use the term "celestial" to refer to 4-configurations with the particular symmetry restrictions described above. In [1], one author (LWB) developed two procedures which modify suitable celestial configurations to yield new 4-configurations. In the first of these, every other line from a particular symmetry class is deleted and then an equal number of new lines that pass through the center of the configuration—diameters—are added in such a way that the resulting structure is a (noncelestial) 4-configuration. The number of points and lines remains unchanged at the end of the construction since one diameter is added for every line removed. In the second procedure, particular elements of certain symmetry classes of points and of lines are both deleted and then diameters are added in such a way that every point is incident with four lines and every line is incident with four points, with a net loss of both points and lines. In this paper we generaliže the first of those procedures. We refer to this generaližed procedure as sparse line deletion or p-sparse line deletion because in general it is possible to delete a smaller number of lines than in the old construction. The new configurations obtained in this way differ qualitatively from those introduced in [1] in that they exhibit a wider variety of symmetry groups compared to the symmetries of the underlying celestial configurations. In particular, despite beginning with a configuration with a high degree of geometric symmetry, we can obtain configurations of quite low symmetry by repeating the sparse line deletion construction, in contrast to the previous construction. Figure 1 depicts three examples of these new objects; beginning with celestial configurations with di8, di2 and die symmetry, we develop configurations with d6, d4 and d4 symmetry, respectively. The paper is organized as follows. In Section 2 we review the theory and notation for celestial configurations. We correct a minor notational ambiguity from [1] and give new results describing the incidences of the diameters in a series of lemmas. In Section 3 we describe the p-sparse line deletion construction. In Section 4 we show how the construction may be carried out several times simultaneously to yield a rich variety of new configurations. In Section 5 we give examples of configurations obtained by a related, but poorly understood technique applicable in the case where each symmetry class contains an odd number of objects. We close by mentioning several questions that deserve further study. All figures in this paper were generated using the free software Matplotlib [5]. 2 Celestial configurations A celestial configuration is a 4-configuration with a high degree of geometric symmetry; specifically, such a configuration has the property that every point is incident with exactly two lines from each of two symmetry classes, and every line is incident with exactly two points from each of two symmetry classes. If a celestial configuration has k symmetry classes of points and of lines, we refer to it as a k-celestial configuration. Each k-celestial configuration consists of a composite number mk of points and mk lines for some m. The points are the vertices of k concentric regular m-gons, and the configuration exhibits m-fold dihedral symmetry (that is, dm symmetry). An example of a 3-celestial configuration is shown in Figure 2. In that figure, the three symmetry classes of points are distinguished by color (red, green and blue), and the three symmetry classes of lines are distinguished in the same way (also red, green, and blue). Each green line contains two red points and two green points (and similarly for the other two classes of lines), and each blue point lies on two red and two blue lines (and similarly for the other two classes of points). Celestial configurations will serve as the building blocks of all of the new 4-configurations described in this paper. One useful feature of celestial configurations is the fact that every celestial configuration may be described by a configuration symbol m#(si,ti; s2,t2; ■■ ■ ; sk,tk ) which encodes a geometric construction algorithm. The integers Si,U,m in the configuration symbol must satisfy several constraints for the construction to yield a 4-configuration; in this case we say the symbol is valid. The constraints are: m > 7, k > 2,1 < si,ti < m for all i, and 1. (order condition) adjacent entries in the sequence (si,ti, s2, ■ ■ ■ ,tk) (taken cyclically) are distinct; (a) 18#(13*, 7; 8, 6); D* (b) 12#(23*, 4; 1, 2; 4,1); D*, (c) 16#(54*, 3; 4, 5; 3,4); D*, Figure 1: Three new 4-configurations. (a), the 3-sparse line deletion 18#(13*, 7; 8,6); D*, with d6 symmetry. (b) the 3-sparse line deletion 12#(23*, 4; 1,2; 4,1); D*, with d4 symmetry. (c) the 4-sparse line deletion 16#(54*, 3; 4, 5; 3,4); D*, with d4 symmetry. ("0)2 Figure 2: The 3-celestial configuration 8#(2,1; 3, 2; 1,3). The labeling updates Figure 2a in [1]. Throughout the paper we use red for v0 and L0, blue for v1 and L1, and green for v2 and L2. 2. (even condition) J2k=1 si +t% is even; and 3. (cosine condition) nk=icos (smn) = nk=icos (im); 4. (substring condition) the symbol m#(L) is invalid whenever L is a proper contiguous substring of (s1,t1; • • • ; sk,tk). As an example illustrating contiguity, (3, 2; 1,4) and (4, 7; 5, 3) are contiguous substrings of (5, 3; 2,1; 4, 7) but (5, 2; 4, 7) is not. The cosine condition is satisfied automatically if the sets S = js1,...,sfc} and T = {t1,... ,tk} are equal, in which case the configuration is called trivial. All the configurations in this paper, with the exception of those in Figures 1a and 3, are formed from trivial celestial configurations. More information on these conditions can be found in [3, Chapter 3]. We now turn to the construction algorithm encoded by the symbol. 2.1 Geometric construction algorithm (celestial configurations) We write P V Q for the line passing through points P and Q and L Л M for the intersection of lines L and M. In the symbols (vjj and (Ljj, the second index j is to be interpreted modulo m. The construction algorithm to produce a celestial configuration given a valid configuration symbol is as follows. 1. Begin with the vertices of aregular m-gon; e.g. take (v0)j = (cos (, sin (), for 0 < i < m. Let v0, written without a second subscript, denote the collection of these points. 2. Given points vj, define (Lj )i = (vj )i V (vj)i+sj+1, for 0 < i < m. We denote by L j the collection of these lines. 3. Given lines Lj, define (vj+i)i = (Lj)i Л (Lj)i-tj, for 0 < i < m, and let vj+i denote the collection of these points. 4. Repeat the previous two steps until the line class Lk-1 is obtained using the parameter sk. Stop before constructing the points vk ; if the symbol is valid, the set of points vk that would be constructed in the next step would coincide setwise with the points vo. For future reference we list all of the incidences explicitly in Table 1. Table 1: Incidences between members of point and line classes in the celestial configuration m#(s1, t1;... ; sk, tk). The quantity S is defined by S = i=1 s - ti. Object Incidences (Lj)i, 0 < j < k — 1 (vj )i (vj )i+Sj+1 (vj+1)i, (vj+1 )i+tj + i (Lk-1 )i (vk-1)i (vk-1)i+Sj + i (v0)i+Ä (v0)i+d+tfc (v0 )i (L0 )i (L0 )i-s1 (Lk-1)i-s (Lk-1)i-đ-tk (vj)i, 0 < j < k — 1 (Lj )i (Lj)i sj+1 (Lj-1)i (Lj-1)i-tj . 2.2 Lines through the origin The vertices in a given point class vj of a celestial configuration form a regular m-gon. For each integer I it follows that the angle Z(vj)0O(vj)e is an integer multiple of 2n/m (that is, an even multiple of n/m). A slightly weaker statement holds for points in different symmetry classes: for i = j, it is still true that the angle Z(vi)0O(vj-)i is an integer multiple of n/m. In the constructions we consider we will add lines through the center of the configuration (although the center is not one of the points of the configuration). We denote by Dj the line through the origin that makes an angle of j m radians with the line O V (v0)0 (conventionally a horizontal line) for j = 0,1, • • • , m — 1. For j > m or j < 0 we reduce modulo m so that Dm = D0 = O V (v0)0. This notation is more flexible than the concept of diametral type introduced in [1] and does not require m to be even. We refer to all of the Dj as diameters. With this notation we restate some useful facts on celestial configurations. 1. Suppose that m is even and (vj)i lies on Da. Then (vj)i+ m also lies on Da so that Da passes through two points of vj. However, if q is odd then Da+q passes through no points of vj. Hence if m is even, each diameter passes through either zero or two points from each symmetry class. 2. Suppose that m is odd. Then each diameter is incident with exactly one point of each symmetry class. 3. Let 0 < j < k — 1. If (vj)0 lies on Da then (vj+1)0 lies on Da+Sj+i-tj+i. By combining (1) and (3) we see that if m and (s1 +11) are even, then the even-numbered diameters pass through two points from each of v0 and v1 while the odd-numbered diameters miss all of the points in v0 and v1. We now give three lemmas providing specific information on the incidences of the diameters. This information is conveniently expressed in terms of the constants [ßj } defined by ß0 = 0 and ßj =^2 Sq - tq, j = 1 ••• - 1. q= 1 Lemma 2.1. For all i and j, the point (vj )i lies on the diameter Dßj +2i. Proof. By definition, (v0)0 lies on D0. Applying (3) repeatedly we see that (vj)0 lies on diameter Dßj. It follows that (vj)i lies on Dßj+2i. □ Lemma 2.2. For 0 < £ < m, and 0 < j < k, the diameter Di passes through the following points of vj : none (vj ) fzßi , (vj ) m+i-ßj 2 2 (vj ) m+l-ßj 2 (vj ) i—j m even, ßj — £ odd; m even, ßj — £ even; m odd, ßj — £ odd; m odd, ßj — £ even. Proof. Lemma 2.1 states that for each i the point (vj )j lies on Dßj+2i, so it suffices to solve the congruence £ = ßj + 2i (m) for i. Equivalently we solve 2i = £ — ßj (m). If m is odd, this equation has one solution because 2 is a generator of the cyclic group Z/mZ. This solution depends on the parity of £ — ßj as indicated. If m is even then 2i and 2i — m are always even, so there is no solution if £ — ßj is odd. If £ — ßj is even then both ßj □ and m+2 ßj are solutions, as indicated. Lemma 2.3. If 0 < j, £ < k and 0 < i < m, the points of vf> sharing a diameter with (vj )i are none Ы i+ ßj - , (v£)i+ m+ßj -i+ 2 (v£)i+ m + ßj -i+ 2 (v£)i+ ßj -ßi m even, ßi — ßj odd; m even, ßi — ßj even; m odd, ßi — ßj odd; m odd, ßi — ßj even. Proof. Lemma 2.1 implies that (vj )i lies on Dßj +2i. Lemma 2.2 then states which points of vi lie on this diameter. Writing £ = ßj + 2i and j = £ to match the notation of Lemma 2.2, we find that the following points of v~- lie on Df none (vj ) l-ß- , (vf ) m+1-ß- (vf) m+i -ß (vf)l-ß- m even, ßf — £ odd; m even, ßf — j even; m odd, ßf — j odd; m odd, ßf — j even. 2 2 2 In other words, the following points of vi lie on Dßj+2i: : , (vi) m + ßj +2i- none (vi ) ßj +2i-2 (vi ) m+ßj + 2i-ߣ 2 (vi ) ßj +2i-ߣ m even, ßi — ßj odd; m even, в — ßj even; m odd, ßi — ßj odd; m odd, ßi — ßj even. The result follows. □ 2 2 3 Sparse line deletion Consider the celestial configuration 18#(5,1; 4,6), illustrated in Figure 3a. Suppose we delete the lines (L0)k, k = 0, 3, 6,9,12,15; the resulting structure is not a configuration because some of the points, shown larger in Figure 3b, have lost an incidence. We say that these points have been affected by the line deletion. Note that the affected points of v0 lie on the same diameters as the affected points of v1, and each diameter that has any affected point incident with it in fact is incident with two points from each of the two symmetry classes. In addition, each affected point is missing precisely one line. Therefore, if we add the six diameters {D0, D4, D6, Di0, Di2, D16}, we obtain the 4-configuration depicted in Figure 3c. This is an example of the 3-sparse line deletion construction. We call this construction sparse in comparison with the construction given in [1], because we remove only one-third of the lines L0 instead of one-half. Figure 4 shows the result of the construction technique described in [1], which was called odd deletion in that work and which corresponds to 2-sparse deletion in the terminology of the present work, beginning from the same celestial configuration 18#(5,1; 4,6). The example of Figure 4 also serves to correct an error from [1], where it was claimed incorrectly that the construction would work only for k-celestial configurations with k > 3. The following theorem gives necessary conditions for the procedure described above to succeed, given parameters m, si,ti of the celestial configuration and a sparsity p. The proof shows that the affected points all lie on a particular set of diameters, and that all points on these diameters are affected. The case p = 2 was proven in [1]. Theorem 3.1 (p-Sparse Line Deletion). Let p > 2, and let C be a celestial 4-configuration with symbol m#(s1,t1; s2,t2; ■ ■ ■ ; sk ,tk ) satisfying the following conditions: (i) p does not divide s^ (ii) m is even, and either Щ = 0 (mod p) or Щ- = s1 (mod p). (iii) The points lying on even-numbered diameters are precisely those of v0 and v1, i.e.: If k = 2, then s1 +11 and s2 +12 are both even. If k > 3, then si + ti is odd for i = 2, i = k, and even otherwise. (iv) The following sets coincide when reduced modulo p: s1 + t1 s1 — t1 {0,s1} = (b) 18#(53*, 1; 4, 6) (c) 18#(53*, 1;4, 6); D* Figure 3: The 3-sparse line deletion construction. (a) The celestial configuration 18#(5,1; 4, 6) with (L0)n drawn thicker for n = 0 mod 3. (b) Lines (L0)n for n = 0 mod 3 have been deleted and the points affected by the deletion are drawn larger. This structure is denoted 18#(53*, 1; 4,6) and is not a 4-configuration; the notation 53* is explained in Theorem 3.1. (c): The 4-configuration 18#(53*, 1; 4,6); D* obtained from (b) by adding diameters. Figure 4: The 2-sparse line deletion configuration 18#(52*, 1; 4,6); D*. All of the diameters have been added, so the other constructions considered in this paper are sparse in comparison. In the notation of [1] this would have been denoted 18#(5*, 1; 4,6); D. Remove from C the lines (L0)np, 0 < n < m. Add the diameters passing through the affected points of v0, i.e. D2np, D2(np+si), for 0 < n < [ 2p ]. Then the resulting structure C is again a 4-configuration, which we denote as m#(sp* ,ti;... ; sk, tk); D*. Proof. We verify that each object in the new structure has exactly four incidences. Each line (Lj )j of C that is not deleted still has exactly four incidences in C ' since no points are added or deleted in this construction. The added diameters also pass through exactly four points. To see this, note that by condition (iii) the classes v0 and vi and no others lie on even-numbered diameters. Condition (ii) implies that m is even, so each even-numbered diameter passes through two points from each of v0 and vi and no others. Consider now the points (vj)j with j > 1. By condition (iii) these lie on odd-numbered diameters. They therefore do not gain any incidence from the added diameters, and they do not lose any incidence either since the deleted lines are chosen from L0 and these lines are incident only with points of v0 and vi (again by condition (iii)). It remains only to show that each point of v0 and vi lies on exactly four lines after diameters are added. We begin with the points v0. A point (v0)j lies on two lines of L0, namely (L0)j and (L0)i-si. Because si ф 0 (mod p) by condition (i), at most one of these lines is deleted. Because we add a diameter if and only if it passes through an affected point of v0, the affected points regain their lost incidence and have exactly four incidences. Hence all points (v0)j have at least four incidences in C'. We must still check that none of them have five, i.e., that no unaffected point of v0 lies directly across the origin from an affected point of v0 on the same diameter. We therefore suppose that (v0)i is affected by the deletion, i.e. i = 0 (mod p) or i = s1 (mod p), and we show that its reflection (v0)i+ m across the origin is also affected. To do so, we must show that either i + Щ = о (mod p) or i + Щ = s1 (mod p). We consider the two cases of condition (ii). In the first case, where Щ = 0 (mod p) the desired congruence is immediate. In the second case we have Щ = s1 (mod p), so 2s1 = m = 0 (mod p). However, i + Щ is congruent to i + s1, since we are in the case where Щ = s1 (mod p), and this is now congruent to either s1 or 2s1 = 0 (mod p), according to whether i = 0 or i = s1. Hence all of the points v0 have exactly four incidences in the new structure C'. Now consider the points of v1. We begin by showing that a point (v1)i can lose at most one incidence when the lines (L0)np, 0 < n < [2rp], are deleted. Indeed, (v1)i lies on only two lines from the first line class, namely (L0)i and (L0)i_tl. Thus, we need to show that t1 = 0 (mod p). This follows from condition (iv). If t1 were congruent to 0 (mod p), we would have Sl_^ = Sl_^ +11 = Sl+^ (mod p). These numbers cannot be congruent, however, since one is congruent to 0 and the other to s1. This shows that each point (v1)i will lose either zero or one incidence when the lines (L0)np are removed. It follows that each line deletion affects two points of v1 as well as two points of v0, so the same number of points are affected in each of these point classes. Finally, we argue that the affected points of class v1 are precisely those that lie on the added diameters. Because v1 contains the same number of affected points as v0, it suffices to show that each affected point of class v1 lies on one of the diameters added previously. A counting argument then guarantees that no unaffected point lies on an added diameter. Since в1 = s1 - t1 is even by condition (iii) and m is even by condition (ii), Lemma 2.3 implies that each point (v1)i shares a diameter with (v0)i+i(sl_tl). The affected points of v1 are those lying on (L0)q where q = 0 (mod p), namely (v1)q and (v1)q+tl. It therefore suffices to show that if i = 0 or t1 (mod p), then i + 1(s1 — t1) = 0 or s1 (mod p), since 0 and s1 are the remainders modulo p of the indices of affected points in v0. But this is equivalent to condition (iv). Hence the affected points of v1 lie on added diameters in C'. This completes the proof. □ 3.1 Notation The notation of [1] may be extended to these generalized p-sparse constructions. If each p-th line of the class L0 has been deleted from the celestial configuration m#(s1, t1 ;... ; sk, tk), we denote the resulting incidence structure by m#(sf*, t1;... ; sk, tk); it is not a configuration. The notation m#(s1*,t1;... ; sk,tk) that was used in [1] should now be written as m#(s2*, t1;... ; sk, tk) since all of those constructions were 2-sparse. We append the symbol D* to the end of the sequence to indicate that for 0 < i < m we add the diameter Di if any of the points on Di have been affected by the line deletion. For brevity we do not explicitly state the indices of the added diameters. These can be recovered if necessary: under the conditions of Theorem 3.1, the added diameters are Di with 2 = 0 or 2 = s1 (modp). Hence if m#(s1,t1;... ; sk ,tk ) is a celestial configuration, (a) 24#(24*, 10; 7, 2; 10, 7); D* (b) 24#(26*1, 10; 7, 2; 10, 7); D* Figure 5: The celestial symbol 24#(2,10; 7,2; 10, 7) satisfies the hypotheses of Theorem 3.1 for both p = 4 and p = 6, yielding two new configurations. then m#( si ,t\;... ; s k ,tk ) is an incidence structure formed by removing each p-th line in L0, and Theorem 3.1 asserts that m#(sj*,t1;... ; sk,tk); D* is again a configuration under certain conditions on m, sj, tj, and p. We will need more powerful notation in the next section. In our examples so far, we deleted the lines (L0)q for all q = 0 (mod p). The construction works equally well if we delete instead the lines (L0)q for all q = b (mod p), where 0 < b < p (to see this, rotate the configuration through an angle of -2nb/m radians, perform the same operation, then rotate back). If b = 0 we write b following the asterisk in the superscript of s1; for clarity we may also do this even if b = 0. The construction outlined in Theorem 3.1 also works if instead of deleting every p-th line in class L0, we instead delete every p-th line in class Lj-1, provided the symbol satisfies the (suitably shifted) conditions of Theorem 3.1. We therefore use the notation m#(--- ;sP*b,tj;•••) to indicate deletion of each line (Lj-1)q, with q = b (mod p). See Figure 5b for an example with j = 0 and b =1. The next section details a generalization of this deletion technique, in which several deletions on the same set of lines are performed simultaneously; to denote this, we write m# #(••• ; sj ,tj ; ••• ) to indicate deletion of each line (Lj-1)q, with q = b1 or q = b2 or q = b3 (mod p). Figure 6: The constructions illustrated in Figure 5 have been carried out simultaneously to obtain 24#(24*0'6*1,10; 7,2; 10,7); D*. This procedure degrades the symmetry group from d24 to d2. 4 Repetition of sparse line deletion 4.1 Multiple deletions within the first line class Consider the celestial configuration 24#(2,10; 7,2; 10,7). This symbol satisfies the conditions for Theorem 3.1 for both p = 4 and p = 6. We can delete the lines (L0)q for q = 0,4, 8, • • • and add diameters to obtain 24#(24*0,10; 7,2; 10,7); D*, depicted in Figure 5a. The affected points of v0 are those (v0)j with i = 0 or i = s1 = 2 (mod 4). This leaves all of the (v0)j with odd i untouched. On the other hand, if we delete all lines (L0)q with q = 1 (mod 6), only points (v0)j of odd index will be affected: see Figure 5b. We may therefore perform both constructions together to obtain the configuration 24#(24*0'6*1,10; 7, 2; 10, 7); D*, depicted in Figure 6. We have now added all but two of the even-numbered diameters, and the deletion is "sparse" only in comparison with the construction given in [1]. The resulting configuration has only the four symmetries of a rectangle, compared to the 48 symmetries of the underlying celestial configuration. That is, the new symmetry group has index 12 in the original group. For 2-sparse line deletion the index is at most 4. This indicates that the more general procedure can give qualitatively novel configurations. Many celestial configurations admit p-sparse line deletions for several values of p. A naive exhaustion search by machine using the conditions of the theorem uncovered several extreme examples. The celestial configuration 48#(13,11; 20,13; 11, 20) admits p-sparse line deletion with p = 2,3,4, 6 or 12. With 80#(12,28; 23,12; 28, 23) we can take p = 5, 8,10, or 20. By repeating and combining the p-sparse line deletions for some- what larger values of p, we can obtain a large number of new 4-configurations. Even in the relatively small case of 24#(2,10; 7, 2; 10,7) we can obtain the configurations illustrated in Figure 7 in addition to those from Figures 5 and 6. 4.2 Combining deletions in the first and third classes in 4-celestial configurations Another possibility for repetition of the p-sparse line deletion construction arises in the special case of 4-celestial configurations. Suppose that the hypotheses of Theorem 3.1 hold for the celestial symbol m#(si,ti; s2,t2; s3,t3; s4,t4) with p = p1. Suppose further that they hold for the symbol m#(s3,t3; S4,14; si, ti; S2,t2) with p = p2. Beginning from the first symbol m#(s1, t1; s2, t2; s3, t3; s4, t4), we may then perform p1-sparse deletion on the lines L0 and add even-numbered diameters to recover a configuration (as in Theorem 3.1). We may additionally perform p2-sparse deletion on the lines L2 and add odd-numbered diameters to recover yet another configuration of a family not available previously. For example, the 4-celestial configuration 20#(6,4; 3,6; 7, 3; 4,7) admits a 5-sparse deletion on both L0 and L2. We can delete the lines (L0)q with q = 0 (mod 5) and add even-numbered diameters, or we can delete the lines (L2)q with q = 0 (mod 5) and add odd-numbered diameters to obtain a 4-configuration. We can also do both; in this case we arrive at the configuration 20#(65*, 4; 3, 6; 75*, 3; 4,7); D*, depicted in Figure 8b. By rotating the first construction we obtain 20#(65ф1, 4; 3,6; 75*0, 3; 4,7); D*, depicted in Figure 8c. These three objects are different, at least in the geometric sense that they differ by more than an isometry, illustrating the very large number of new configurations available through this method. Finally we note the possibility of repeating deletions within L0 and also repeating deletions within L2. An example is 20#(65*0'2,4; 3,6; 75*0'4,3; 4,7); D*; see Figure 8a and note again the very small symmetry group. 5 Constructions with an odd number of points per symmetry class Let C be a k-celestial configuration with symbol m#(s1, t1;... ; sk, tk). Suppose that m is odd. The hypotheses of Theorem 3.1 cannot hold; in this section we ask if there is another way to remove some lines of C and then add an equal number of diameters to recover a 4-configuration. We will give some examples where this succeeds and suggest a classification of the resulting configurations. We leave open the task of giving explicit construction algorithms with sufficient conditions on m, s, t and p. We claim that such a construction is possible only if k = 4. Indeed, since m is odd every diameter passes through exactly one point in each symmetry class; if the added diameters are lines in a 4-configuration then there must be exactly four classes. (a) 24#(24*0'1,10; 7, 2; 10, 7); D* (b) 24#(26*0,1, 10; 7, 2; 10, 7); D* (c) 24#(26*°>3, 10; 7, 2; 10, 7); D* Figure 7: Three more configurations arising from multiple modifications to the same celestial configuration as in Figures 5 and 6. Note that in the configuration shown in (a) we have deleted every red line (L0)q where q is congruent to 0 or 1 (mod 4). As a result all even-numbered diameters have been added, although this configuration cannot be constructed via 2-sparse line deletion. (a) 20#(66*°>2, 4; 3, 6; 76*°>4, 3; 4, 7); D* (b) 20#(66*, 4; 3, 6; 75*, 3;4, 7); D * . (c) 20#(66*\4; 3, 6;76*°, 3;4, 7); D * . Figure 8: Three configurations obtained from 20#(6,4; 3, 6; 7,3; 4,7) by performing 5-sparse deletion on both L0 and L2. Both odd- and even-numbered diameters have been added. The examples in previous sections proceeded in steps where some of the lines in one symmetry class were removed and diameters were added to yield a new configuration; in the more complicated examples several intermediate configurations were formed and destroyed along the way. With m odd such a scheme cannot work. Because each added diameter passes through points in four symmetry classes, while the lines of any line class Lj pass only through the two point classes v j, vj+i, we must simultaneously delete lines from more than one symmetry class. The necessity of coordinating these different classes of removed lines is the main challenge in this section. We propose the following classification for line deletion constructions. The ray from the origin through (v0)o passes through either zero points or one point from each of the classes vb v2, v3. If this ray passes through a point of class v j we say that v j is a cis class; otherwise we say that vj is a trans class (that is, trans classes are on the opposite side of the origin from points v0, while cis classes are on the same side of the origin as points v0). Hence v0 is always a cis class, and our Theorem 3.1 addresses the case where m is even and the set of cis classes is {v0, v1}. There are 23 = 8 possible sets of cis classes in a 4-celestial configuration. In Figures 9, 10, and 11 we give examples where m is odd and the cis classes are {v0, v1}, {v0, v1; v2, v3}, and {v0, v2, v3} respectively. It may be that for each of the eight possibilities one can find sufficient conditions for some line deletion procedure in the spirit of Theorem 3.1. This problem is beyond our scope here. 6 Questions for further study In Configurations of Points and Lines, Griinbaum wrote that "constructing new 4-configura-tions is still more of an art than a science" [3]. We now offer several possible directions for future work towards the ultimate goal of finding and classifying all 4-configurations. The technique we have explored here, the replacement of some lines of a celestial configuration with an equal number of diameters, can be extended further. The examples given in Section 5 should be systematized with explicit construction algorithms and sufficient conditions. There are also possibilites with m even that are not covered by Theorem 3.1. Figure 12 gives an example with m =12 where v0 and v2 are of cis type, in contrast to the situation of Theorem 3.1, where v0 and v1 are of cis type. This could be the first example of a new infinite family obtained by a more general construction. We also have yet to consider the "even deletion" procedure introduced in [1], in which points as well as lines are removed. This no doubt has a p-sparse generalization and could be worth exploring since the "even deletion" construction in [1] yielded previously unknown (254) configurations. We close by mentioning a related question. We say that two configurations are (combina-torially) isomorphic if there exists an incidence-preserving bijection between the two configurations. It is not clear how many of the configurations introduced here belong to new isomorphism classes in this combinatorial sense. For example, it is not known whether or not the configurations depicted in Figures 8b and 8c are combinatorially isomorphic. Even for the celestial configurations this question has not been solved. Figure 9: The configuration 15#(43*, 2; 1,4; 53*, 1; 2, 5); D*. The ray from the origin through (v0)0 (red) passes through (vi)_i (blue) but no points of v2 (green) or v3 (magenta), so the cis classes are v0 and v1. Figure 10: The configuration 27#(43*, 2; 8,4; 103*, 8; 2,10); D*. All four points on each added diameter lie on the same side of the origin, so all point classes are of cis type. The diameters could be extended through the origin without hitting other points because m is odd. Figure 11: The configuration 3Б#(125*, 13; 3,12; 75*, 3; 13, 7); D*. The cis classes are v0, v2, and v3. Figure 12: The configuration 12#(23*0, 5; 4, 3; 53*1,2; 3,4); D*. Here the cis classes are v0 and v2 . References [1] L. W. Berman and B. Grimbaum, Deletion constructions of symmetric 4-configurations. Part I, Contrib. Discrete Math. 5 (2010), 18-33. [2] M. Boben and T. Pisanski, Polycyclic configurations, European J. Combin. 24 (2003), 431-457, doi:10.1016/S0195-6698(03)00031-3. [3] B. Grünbaum, Configurations of Points and Lines, volume 103 of Graduate Studies in Mathematics, American Mathematical Society, Providence, RI, 2009. [4] B. Grunbaum and J. F. Rigby, The real configuration (2I4), J. London Math. Soc. (2) 41 (1990), 336-346, doi:10.1112/jlms/s2-41.2.336. [5] J. D. Hunter, Matplotlib: A 2d graphics environment, Computing In Science & Engineering 9 (2007), 90-95. [6] D. Marusic and T. Pisanski, Weakly flag-transitive configurations and half-arc-transitive graphs, European J. Combin. 20 (1999), 559-570, doi:10.1006/eujc.1999.0302. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 187-195 Bounds on the domination number of Kneser graphs Patrie R. J. Ostergard * Department of Communications and Networking, Aalto University School of Electrical Engineering, P.O. Box 13000, 00076 Aalto, Finland Zehui Shao f Key Laboratory of Pattern Recognition and Intelligent Information Processing, Institutions ofHigher Education ofSichuan Province, School of Information Science and Technology, Chengdu University, Chengdu, 610106, China Received 14 May 2013, accepted 17 August 2014, published online 28 November 2014 The Kneser graph KG„jfc has one vertex for each k-subset of an n-set and edges between vertices whenever the corresponding subsets are disjoint. A dominating set in a graph G = (V, E) is a subset S С V such that each vertex in V \ S is adjacent to at least one vertex in S. The domination number of KGn k, denoted by 7(n, k), is the minimum size of a dominating set in that graph. Combinatorial and computer-aided techniques for obtaining bounds on j(n, k) are here considered, and several new bounds are obtained. An updated table of bounds on 7(n, k) is presented for n < 21 and k < 5. Keywords: Dominating set, domination number, Kneser graph. Math. Subj. Class.: 05C69, 05C35 * Corresponding author. Supported in part by the Academy of Finland, Grant No. 132122. t Supported by the National Natural Science Foundation of China, Grant No. 61309015. E-mail addresses: patric.ostergard@aalto.fi (Patric R. J. Ostergard), zshao@cdu.edu.cn (Zehui Shao), xxdmaths@sina.com (Xiaodong Xu) Xiaodong Xu Guangxi Academy ofScience, Nanning, Guangxi 530007, China Abstract 1 Introduction Let G = (V, E) be a simple graph, that is, a graph having neither loops nor multiple edges. A dominating set in G is a subset S С V such that each vertex in V\S is adjacent to at least one vertex in S. The domination number 7(G) of G is the minimum size of a dominating set in G. The domination number has been extensively studied in the general case [11]. Due to a variety of applications, the case of n-cubes, G = Qn, is of particular interest [3]; most of the early work considered such graphs. In the current work, another specific type of graphs is considered, namely Kneser graphs. The Kneser graph KG„jfc has one vertex for each k-subset of an n-set and edges between vertices whenever the corresponding subsets are disjoint. If n < 2k, then KG„jfc has no edges, so we assume that n > 2k. We further denote the domination number of KGn k by 7(n, k). See [6, Chap. 7] for an in-depth discussion of Kneser graphs. General and specific bounds on 7(n, k) have been considered in a sequence of studies, including [2, 8, 10, 12, 18]. However, several of the best known bounds for small parameters were rather weak prior to this study. Indeed, the aim of the current work is to apply combinatorial and computer-aided techniques to the problem of improving upper and lower bounds on the domination number of Kneser graphs—and occasionally even find the exact value when the bounds meet. A total dominating set in a graph G = ( V, E) is a subset S С V such that each vertex in V is adjacent to at least one vertex in S. The minimum size of a total dominating set is called the total domination number, and the total domination number of KG„jfc is denoted by 7t(n, k). It is obvious that 7(n,k) < 7t(n, k). (1.1) Let C (v, k, t) denote the smallest number of k-subsets of a v-set, such that every t-subset of the v-set occurs in at least one of the k-subsets. Then 7t(n, k) = C (n, n - k, k), so by (1.1), 7(n,k) < C(n,n - k,k). (1.2) Exact values of and upper bounds on C (v, k, t) for v < 32, k < 16, and t < 5 can be found in [7]. The paper is organized as follows. Methods for obtaining upper and lower bounds on 7(n, k) are considered in Sections 2 and 3, respectively. The results are summarized in Section 4, where an updated table of bounds on 7(n, k) is presented for n < 21 and k < 5. 2 Upper Bounds and Exact Values Upper bounds for the domination number are commonly constructive, that is, explicit dominating sets prove the bounds. We here present various general results for upper bounds on 7(n, k); in some of the theorems, the exact value is in fact obtained. If n < 2k, then the Kneser graph consists of isolated vertices only. In the first nontrivial case, n = 2k, a dominating set must contain one vertex from each pair of disjoint k-sets. Theorem 2.1. For any k, 1 /2k 7(2k,k) = Hk It is easy to show that if n is large enough, then the smallest dominating set is obtained by taking k +1 disjoint k-sets. Theorem 2.2. If n > k2 + k, then y(n, k) = k + 1. The exact value of 7 (n, k) has also been determined for a range of values of n smaller than those covered by Theorem 2.2. Theorem 2.3 ([10]). If k > 3 and 4k2 + k < n < k2 + k, then Y(n,k) = k + 1 + k2 + k - n I_k/2J With increasing n, Y(n, k) turns out to be nonincreasing. Theorem 2.4 ([8, Proposition 4.2.4]). If n > 2k + 1, then Y(n + 1, k) < 7(n, k). Theorem 2.5 ([8, Theorem 4.5.1]). fk > 4 and7(n,k) < min{2k,n-k}, then Y(n, k) = Yt(n,k). We shall next see how dominating sets in certain Kneser graphs are related to a coloring problem for hypergraphs that has been extensively studied. We consider the case n = 2k + 1—such Kneser graphs are known as odd graphs—and view a dominating set S of the graph KG2fc+1jfc as the set of hyperedges in a hypergraph G' = (V', E') with | V= n, |E'| = |S|, and edges of size k (so the hypergraph is k-uniform). Now consider an arbitrary balaneed coloring of the vertices in V' with two colors, that is, k of the vertices are colored with one color and k + 1 with the other [21]. Since the vertex in the original Kneser graph that is labelled by the subset of the k vertices with the first color is dominated by some vertex s g S, the hyperedge in E' corresponding to s is unicolor. Hence, G' does not have a balanced coloring with two colors so that no hyperedge is unicolor, that is, G' is not 2-colorable in a balanced way. Hypergraphs that are 2-colorable (without requiring that the colorings be balanced) are said to have property B [13, Sect. 15.1]. Consequently, a hypergraph with appropriate parameters that does not have property B gives a dominating set in KG2fc+1jfc. Actually, this implication goes in the other direction as well. The upward shadow of a (k - 1)-subset of an n-set is the collection of all k-subsets of the n-set that contain the (k - 1)-subset. Lemma 2.6 ([8, Lemma 4.2.3]). If n > 2k + 1, then there exists a dominating set attaining Y (n, k) that does not contain the upward shadow of any (k — 1)-subset of the n-set. Theorem 2.7. There exists a dominating set S attaining Y(2k + 1, k) that can be trans-formedintoa k-uniform hypergraph G' = (V, E) with |V | = 2k + 1 and |E| = |S | without property B. Proof. Consider the hypergraph G' = (V', E') obtained from a dominating set attaining Y(2k +1, k) that does not contain the upward shadow of any (k - 1)-set. Such a dominating set exists by Lemma 2.6. If the vertices in V' are colored in a balanced way, then since the hypergraph was constructed from a dominating set in KG2fc+1jfc, there exists a unicolor hyperedge. Now consider a coloring of the vertices V' with k + 2 and k - 1 vertices of the two colors and denote the vertices in the former color class by U. If no hyperedge in E' is a subset of U, then the same holds for each (k + 1)-subset of U, so by the existence of a balanced coloring we get that (V' \ U) U v g E' for all v g U. But this is then an upward shadow and we have a contradiction (so a hyperedge that is a subset of U is indeed unicolor). The colorings with classes of size a and n — a where a > к + 3 are handled by considering an arbitrary subset of the larger class of size к + 2 and using the previous By Theorem 2.7 and results by Abbott and Liu [1] we now get that 24 < y(9, 4) < 26. Note, however, that in certain studies on uniform hypergraphs without property B a further assumption is made that the hyperedges must contain all pairs of vertices. Results for that variant of the problem, which is motivated by a more general question regarding property B, are not directly applicable here. This includes the results in [16, 21]. For certain parameters, we can say a lot about Y(2k + 1, к). Let 2 < t < к < v. An S (t, к, v) Steiner system is a collection of k-sets out of a v-set with the property that every t-subset of the v-set occurs in exactly one of the к-sets. The following result has been discussed both in the context of hypergraphs without property B [4] and dominating sets in KG2fc+i,fc [9]; see also [5, Lemma 11.8.3]. Theorem 2.8. There is a Steiner system S (к — 1, к, 2 к + 1) if and only if One may further use partial or exhaustive computational methods to determine bounds on y(n, к). Since upper bounds can be proven by finding a structure attaining the bound, nonexhaustive methods can be applied to such cases. For lower bounds, on the other hand, exhaustive methods are required; we shall consider such methods in the next section. In [19], the tabu search metaheuristic is applied to the problem of finding dominating sets in n-cubes. The algorithm takes the parameters of the instance and the desired size of the dominating set (which is, for example, one less than the best known upper bound), and searches for such a dominating set. The algorithm is also applicable to Kneser graphs—in fact, it is applicable to arbitrary graphs. The reader should consult [19] for details. Structures obtained in the current work and leading to new upper bounds on y(n, к) are listed in the Abstract. Some of the best known structures turn out to have nontrivial symmetries; these could further be used to narrow down the search space. We consider symmetries of dominating sets in terms of the labels (subsets) of the vertices. Two dominating sets are said to be equivalent if there is a permutation of the n-set that maps the vertices of one dominating set onto the vertices of the other. Such a mapping from a dominating set onto itself is an automorphism, and the set of all automorphisms of a dominating set forms the automorphism group of the dominating set. Such an automorphism group is isomorphic to a subgroup of the stabilizer subgroup of the dominating set in Aut(KGn,k). Note that it may happen that the automorphism group is a proper such subgroup; consider, for example, the mapping of the к-subsets of a 2^set to their complements. The nauty software [17] is a useful computational tool in this context. 3 Lower Bounds Several of the bounds in the previous section can be used also to get lower bounds. We shall here state two more general results that can be used to get best known lower bounds for small parameters. The first of these is the well-known volume bound obtained by dividing the total number of vertices with the number of vertices dominated by a single vertex. argument. □ Theorem 3.1. For any n and k, (n) Y(n, k) > -. ' ) > 1 + ("-) Theorem 3.2 ([8, Lemma 4.5.3]). Assume that n = ak, where a, k > 2. Then a Y(n, k) > --(y(n + k, k) — 1). a-1 To simplify the discussion of the techniques to obtain lower bounds on y(n, k), it is useful to think of a dominating set as a constant weight code [20]. The codewords of this code are of length n and have 1s in the coordinates given by the corresponding k-subset and 0s in the other coordinates. Theory and terminology from coding theory can then be directly applied. There are three general approaches to exhaustively search for codes with prescribed parameters [15, Chap. 7]: via subcodes, codeword by codeword, and coordinate by coordinate. There is no obvious way of constructing the current type of codes via subcodes, that is, codes obtained by considering the codewords with a 1 (alternatively, 0) in a given coordinate and deleting that coordinate. Some results can indeed be obtained in a backtrack search constructing the code word by word as in [24], which can be consulted for general details. The origins of the method of constructing codes coordinate by coordinate can be traced back to the 1960s [14], after which it has been developed further and become an efficient tool in the study of dominating sets, in particular in n-cubes and related graphs [22, 23]. See also [15, Sect. 7.2.2]. A version that has been applied to the hypergraph coloring problem discussed in the previous section can be found in [21]. The idea in the coordinate by coordinate approach can be described as a generalization of the following theorem. Theorem 3.3. Let D be a dominating set in KGn,k, and let D = D0 U D 1 such that Di consists of the vertices whose label has an i in the first coordinate. Then 1. |Di| + (n-k71)|Do|> (k-i), 2. (V)^+(i+m) Do|> (nk1). Proof. Let G = (V, E) be the KG„,fc Kneser graph, and let V = Vo U Vi so that the labels of the vertices in Vi have an i in the first coordinate. Then |V0| = ("k ^ and |V11 = (П-^. The result now follows as each vertex in D0 dominates 1 + )n-1-fc) vertices in V0 and (n-fc) vertices in V1, and each vertex in D 1 dominates (n-k71 ) vertices in V0 and 1 vertex in V. □ Theorem 3.3 can be generalized to an arbitrary number of specified coordinates. For example, with two specified coordinates we let D = D00 U D0 1 U D 10 U D 11, V = V00 U V0 1 U V1 0 U V11 and get four inequalities. For a small number of coordinates, the inequalities can be derived by hand, but when the number gets larger, it is convenient to form these computationally. When a code is constructed coordinate by coordinate, one first fixes the number of codewords and then start from the distributions of 0s and 1s in the first coordinate given by Theorem 3.3. In a backtrack exhaustive search, one may for the next couple of coordinates solve larger and larger systems of equations, but at some point one may start checking all possible candidates for the next coordinate and see whether the inequalities are fulfilled. At each level of the search tree, isomorph rejection should be carried out. For the sake of efficiency, one may also require that the number of 1s in the coordinates is either increasing or decreasing. Except for minor differences in details, the approach in [21] can be used. 4 Results Table 1 summarizes the best known bounds on and exact values for y(n, k), n < 21, k < 5. Indices are added to the entries, to give explanations of lower and upper bounds. If a bound can be obtained in several ways, we pick the explanation that in some sense is the nicest. We omit the index when the bound follows from Theorem 2.2. Table 1: Bounds on Y(n, k) for n < 21, k < 5 n\k 2 3 4 5 4 a3a 5 c3c 6 3 a10a 7 3 eye 8 3 cjc a35a 9 3 m 7c j to 1 10 3 b6b c15k a126a 11 3 b5b ®15c e66e 12 3 4 ®12h ®37-56fc 13 3 4 m 10^ j23-39k 14 3 4 dgh f16-31k 15 3 4 dgh g15-27h 16 3 4 b7b ®12-22h 17 3 4 b7b c11—17h 18 3 4 b6b c11—15h 19 3 4 b6b c11—14h 20 3 4 5 c11—12h 21 3 4 5 d11—12h Key to Table 1. Unmarked bounds are from Theorem 2.2. Bounds: a Theorem 2.1 b Theorem 2.3 c Theorem 2.4 d Theorem 2.5 and [7] e Theorem 2.8 f Theorem 3.1 g Theorem 3.2 h Eq. (1.2) and [7] 1 Exhaustive search, coordinate by coordinate j Exhaustive search, word by word k New constructive result, see Appendix 1 Abbott and Liu [1] (Theorem 2.7) m Gorodezky [8] Appendix We here list the structures that lead to new upper bounds on j(n, k). We first present structures that can be described as a set of orbits under the action of a permutation group, and finally list some explicit structures (we do not exclude the possibility that these, or better, bounds could also be obtained by a structure with some symmetry). Y(10,4) < 15: Generator of group: (1 2 3 4 5)(6 7 8 9 10) Orbit representatives: 1110010000, 1010000110, 1000001101 Y(13, 5) < 39: Generator of group: (1 2 3 4 5 6 7 8 9 10 11 12 13) Orbit representatives: 1101011000000, 1110001001000, 1101000100010 Y(12, 5) < 56: 111000110000, 100000111100, 010001011001, 010000111010, 001010000111, 011100010001, 110101000010, 100011010100, 000011111000, 110100100001, 110011001000, 011101001000, 101011001000, 000001100111, 011000100110, 110000010110, 001001100011, 100110001100, 111110000000, 000110010110, 100001101001,010011100001,010100001101,000010011011, 000111100100, 001110110000, 010101010100, 101010010001, 001110001010, 001000111010, 011011000010, 001100001101, 100101001010, 001001011001, 010110001010, 100001010011, 011010100001, 101000101100, 000100101011, 110000101100, 010010000111, 101101000010,011001000101, 100100110001, 100010000111, 011000011100, 110010010001, 010011100100, 101000010110, 000100011101, 000101110010, 101100100001, 001011100100, 111000001010, 001101010100, 010110110000 Y(14, 5) < 31: 10000011001100,00011110100000, 11100100100000, 00110101000010, 00110110000001, 00001011000110, 00010010100110, 01000011000011, 00000101101100,01011000010100, 10110000010001, 00100000100111, 10001010011000, 01101010000001, 01100010011000, 01010100011000, 11000000001101, 10010000101010, 01100001010100, 00001001110001, 10001111000000, 01001010110000, 00000001011011, 10000100010110, 11001100000010, 00101000101010, 00001100001101, 10111001000000, 01010001100001, 10000110110000, 00011010001100 References [1] H. L. Abbott and A. C. Liu, On property B of families of sets, Canad. Math. Bull. 23 (1980), 429-435. [2] E. Clark, Domination in Kneser graphs, unpublished manuscript, 1988. [3] G. Cohen, I. Honkala, S. Litsyn, and A. Lobstein, Covering Codes, North-Holland, Amsterdam, 1997. [4] H. L. de Vries, On property B and on Steiner systems, Math. Z. 153 (1977), 155-159. [5] C. D. Godsil, Algebraic Combinatorics, Chapman & Hall, New York, 1993. [6] C. Godsil and G. Royle, Algebraic Graph Theory, Springer-Verlag, New York, 2001. [7] D. M. Gordon and D. R. Stinson, Coverings, in: C. J. Colbourn and J. H. Dinitz (eds.), Handbook of Combinatorial Designs, 2nd ed., Chapman & Hall/CRC, Boca Raton, 2007, 365-373. [8] I. Gorodezky, Domination in Kneser graphs, Master's thesis, University of Waterloo, Canada, 2007. [9] P. Hammond and D. H. Smith, Perfect codes in the graphs Ok, J. Combin. Theory Ser. B 19 (1975), 239-255. [10] C. Hartman and D. B. West, Covering designs and domination in Kneser graphs, unpublished manuscript, 2003. [11] T. W. Haynes, S. T. Hedetniemi, and P. J. Slater, Fundamentals of Domination in Graphs, Marcel Dekker, New York, 1998. [12] J. Ivanco and B. Zelinka, Domination in Kneser graphs, Math. Bohem. 118 (1993), 147-152. [13] T. R. Jensen and B. Toft, Graph Coloring Problems, Wiley, New York, 1995. [14] H. J. L. Kamps and J. H. van Lint, The football pool problem for 5 matches, J. Combin. Theory 3 (1967), 315-325. [15] P. Kaski and P. R. J. Ostergard, Classification Algorithms for Codes and Designs, Springer, Berlin, 2006. [16] G. Manning, The M (4) problem of Erdos and Hajnal, Ph.D. dissertation, Northern Illinois University, 1997. [17] B. D. McKay, nauty user's guide (version 1.5), Technical Report TR-CS-90-02, Computer Science Department, Australian National University, Canberra, 1990. [18] J. C. Meyer, Quelques problemes concernant les cliques des hypergraphes h-complets et q-parti h-complets, in: C. Berge and D. Ray-Chaudhuri (eds.), Hypergraph Seminar, Lecture Notes in Mathematics, Vol. 411. Springer-Verlag, Berlin, 1974, 127-139. [19] P. R. J. Ostergard, Constructing covering codes by tabu search, J. Combin. Des. 5 (1997), 7180. [20] P. R. J. (Ostergard, Classification of binary constant weight codes, IEEE Trans. Inform. Theory 56 (2010), 3779-3785. [21] P. R. J. (Ostergard, On the minimum size of 4-uniform hypergraphs without property B, Discrete Appl. Math. 163 (2014), 199-204. [22] P. R. J. (Ostergard and U. Blass, On the size of optimal binary codes of length 9 and covering radius 1, IEEE Trans. Inform. Theory 47 (2001), 2556-2557. [23] P. R. J. (Ostergard and A. Wassermann, A new lower bound for the football pool problem for 6 matches, J. Combin. Theory Ser. A 99 (2002), 175-179. [24] P. R. J. (Ostergard and W. D. Weakley, Classification of binary covering codes, J. Combin. Des. 8 (2000), 391-401. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 197-207 The expected values of Kirchhoff indices in the random polyphenyl and spiro chains* Guihua Huang College of Mathematics and Computer Science, Hunan Normal University, Changsha, Hunan 410081, P. R. China Meijun Kuang College of Mathematics and Computer Science, Hunan Normal University, Changsha, Hunan 410081, P. R. China Hanyuan Deng t College of Mathematics and Computer Science, Hunan Normal University, Changsha, Hunan 410081, P. R. China Received 14 March 2013, accepted 22 April 2014, published online 28 November 2014 The Kirchhoff index Kf (G) of a graph G is the sum of resistance distances between all pairs of vertices in G. In this paper, we obtain exact formulas for the expected values of the Kirchhoff indices of the random polyphenyl and spiro chains, which are graphs of a class of unbranched multispiro molecules and polycyclic aromatic hydrocarbons. Moreover, we obtain a relation between the expected values of the Kirchhoff indices of a random polyphenyl and its random hexagonal squeeze, and the average values for the Kirchhoff indices of all polyphenyl chains and all spiro chains with n hexagons, respectively. Keywords: Expected value, average value, Kirchhoff index, resistance distance, polyphenyl chain, spiro chain. Math. Subj. Class.: 05C12, 05C80, 05C90, 05D40 * Project supported by Hunan Provincial Natural Science Foundation of China(13JJ3053). t Corresponding Author. E-mail addresses: 380026412@qq.com (Guihua Huang), 1075998525@qq.com (Meijun Kuang), hydeng@hunnu.edu.cn (Hanyuan Deng) Abstract 1 Introduction Based on the electrical network theory, Klein and Randić [13] introduced the concept of resistance distance. A connected graph G with vertex set {v1,v2, ■ ■ ■ , vn} is viewed as an electrical network N by replacing each edge of G with a unit resistor, the resistance distance between vi and vj, denoted by rG(vi,vj) or r(vi,vj), is the elective resistance between them as computed by the methods of the theory of resistive electrical networks based on Ohm's and Kirchhoff's laws in N. The Kirchhoff index of G, denoted by Kf (G), is the sum of resistance distances between all pairs of vertices in G, namely, Kf (G) = £ rG(vi,vj ) i 2) hexagons, where Hk is the k-th hexagon of PPCn attached to Hk-1 by a cut edge uk-1ck, k = 2, 3, ■ ■ ■ , n. A vertex v of Hk is said to be ortho-, meta- and para-vertex of Hk if the distance between v and ck is 1,2 and 3, denoted by ok, mk and pk, respectively. Examples of ortho-, meta-, and para-vertices are shown in Figure 1. Except the first hexagon, any hexagon in a polyphenyl chain has two ortho-vertices, two meta-vertices and one para-vertex. X2 X3 Figure 1: A polyphenyl chain PPCn with n hexagons, cn = x\ and ortho-vertices on = x2lx6, meta-vertices mn = x3,x5, and para-vertex pn = x4 in Hn. A polyphenyl chain PPCn is a polyphenyl ortho-chain if uk = ok for 2 < k < n — 1. A polyphenyl chain PPCn is a polyphenyl meta-chain if uk = mk for 2 < k < n — 1. A polyphenyl chain PPCn is a polyphenyl para-chain if uk = pk for 2 < k < n — 1. The polyphenyl ortho-, meta- and para-chain with n hexagons are denoted by On, Mn and Pn, respectively. For n > 3, the terminal hexagon can be attached to meta-, ortho-, or para-vertex in three ways, which results in the local arrangements we describe as PPCn+1, PPCn+1, PPCn+1, see Figure 2. Figure 2: The three types of local arrangements in polyphenyl chains. A random polyphenyl chain P PC (n,p1,p2) with n hexagons is a polyphenyl chain obtained by stepwise addition of terminal hexagons. At each step k(= 3,4, • • • , n), a random selection is made from one of the three possible constructions: (i)PPCk-1 ^ PPC1 with probability p1, (ii)PPCk-1 ^ PPC2 with probability p2, (iii)PPCk-1 ^ PPCf with probability 1 — p1 — p2 where the probabilities p1 and p2 are constants, irrespective to the step parameter k. Specially, the random polyphenyl chain PPC(n, 1,0) is the polyphenyl meta-chain Mn, PPC(n, 0,1) is the polyphenyl orth-chain On, and PPC(n, 0,0) is the polyphenyl para-chain Pn, respectively. Also, a spiro chain SPCn with n hexagons can be regarded as a spiro chain SPCn-1 with n — 1 hexagons to which a new terminal hexagon has been adjoined, see Figure 3. _ x6 x 5 SPCn-1 Un-^Xi X X2 Хз Figure 3: A spiro chain SPCn with n hexagons. For n > 3, the terminal hexagon can also be attached in three ways, which results in the local arrangements we describe as SPCn+1, SPCn+1, SPCn+1, see Figure 4. SPCn срл 1 срл2 срлЗ SPCn+1 SPCn+1 SPCn+1 Figure 4: The three types of local arrangements in spiro chains. A random spiro chain S PC (n,p1,p2) with n hexagons is a spiro chain obtained by stepwise addition of terminal hexagons. At each step k(= 3,4, • • • , n), a random selection is made from one of the three possible constructions: (i)SPCfc_1 ^ SPC1 with probability pb (ii)SPCfc-1 ^ SPC2 with probability p2, (iii)SPCfc_1 ^ SPC3 with probability 1 - p1 - p2 where the probabilities p1 and p2 are constants, irrelative to the step parameter k. Similarly, the random spiro chain SPC(n, 1,0), PPC(n, 0, l) and PPC(n, 0,0) are the spiro meta-chain Mn, the spiro orth-chain On and the spiro para-chain Pn, respectively. For a random polyphenyl chain PPC(n,p1,p2) and a random spiro chain SPC(n, p1,p2), their Kirchhoff indices are random variables. In this paper, we will obtain exact formulas for the expected values E(Kf (PPC(n,p1,p2))) and E(Kf (SPC(n,p1,p2))) of the Kirchhoff indices of random polyphenyl and spiro chains, respectively. 2 Main results 2.1 The Kirchhoff index of the random polyphenyl chain In this section, we will consider the Kirchhoff index of the random polyphenyl chain. Theorem 2.1. For n > 1, the expected value of the Kirchhoff index of the random polyphenyl chain PPC(n,p1,p2) is E(Kf(PPC(n,p1,p2))) = (15 -p1 -4p2)n3 + (3p1 + 12p2 + 8)n2 - (2p1 + 8p2 + y)n Proof. Note that the polyphenyl chain PPCn is obtained by attaching PPCn-1 a new terminal hexagon by an edge, we suppose that the terminal hexagon is spanned by vertices x1, ж2, ж3, ж4, ж5, же, and the new edge is un-1x1 (see Fig.1). Then (i) For any v e PPCn-1, r(xi,v) = r(un-i,v) + 1, r(x2,v) = r(un-i,v) + 1 + 5, r(x3, v) = r(un-i, v) + 1 + 3, r(x4, v) = r(un_i, v) + 1 + 3, r(x5, v) = r(u„_i, v) + 1 + 3, r(x6, v) = r(u„_i, v) + 1 + 5; (ii) PPCn_i has 6(n - 1) vertices; (iii) For k G {1, 2, 3,4, 5, 6}, £ r(xk,x,) = 35 i=i So, we have r(xi r(x2 r(x3 r(un_i|PPCn_i) + 1 X 6(n - 1) + 35 r(un_i|PPCn_i) + (1 + f) X 6(n - 1) + 35 PPCn) PPCn) PPCn) = r(un_i|PPCn_i) + (1 + 4) X 6(n - 1) + 35 r(un_i|PPCn_i) + (1 + 3) X 6(n - 1) + 35 r(x4|PPCn) r(x5|PPCn) = r(x3|PPCn_i) r(xf|PPCn) = r(x2|PPCn_i) where r(x|G) = J2 r(x, y), and yev(g) 66 Kf (PPCn) = Kf (PPCn_i) + 6r(un_i|PPCn_i) + 71n - 36 - 1 ^ r(vi, vj) i=i j=i Then 35 = Kf (PPCn_i) + 6r(un_i|PPCn_i) + 71n - 36 - 35 35 Kf (PPCn+i) = Kf (PPCn) + 6r(un|PPCn) + 71n +35 (2.1) For a random polyphenyl chain PPC(n,pi,p2), the resistance number r(un|PPC(n, pi, p2)) is a random variable, and its expected value is denoted by Un = E(r(un|PPC (n,pi,p2))). By the expectation operator and (1), we can obtain a recursive relation for the expected value of the Kirchhoff number of a random polyphenyl chain PPC(n, pi, p2 ) 35 E (Kf (PPC (n + 1,pi,p2)) = E (Kf (PPC (n,pi,p2))) + 6Un + 71n + 35 (2.2) Now, we consider computing Un. (i) If PPCn ^ PPCn+i with probability pi, then un coincides with the vertex x3 or x5. Consequently, r(un|PPCn) is given by r(x3|PPCn) with probability pi. (ii) If PPCn ^ PPCn+i with probability p2, then un coincides with the vertex x2 or x6. Consequently, r(un|PPCn) is given r(x2 |PPCn) with probability p2. (iii) If PPCn ^ PPCn+i with probability 1 - pi - p2, then un coincides with the vertex x4. Consequently, r(un|PPCn) is given by r(x4|PPCn) with probability 1 -pi -P2. From (i)-(iii) above, we immediately obtain Un =r(x3\PPCn)pi + r(x2\PPCn)p2 + r(x4\PPCn )(1 - Pi - P2 ) 35 =Pi[r(un-i|PPC(n - 1,Pi,P2)) + 14(n - 1) + —1 6 35 + P2[r(Un-i\PPC(n - 1,Pi,P2)) + 11(n - 1) + —1 6 35 + (1 - pi - p2)[r(un-i\PPC(n - 1,pi,p2)) + 15(n - 1) + —] 6 By applying the expectation operator to the above equation, we obtain 55 Un = Un-i + (15 - pi - 4p2)n + pi +4p2 - — 6 And Ui = E(r(ui\PPC(1,pi,p2))) = 365, using the above recurrence relation, we have тт (15 -pi - 4p2) 2 , tpi , о Un =-^-n + + 2p2 - 3)n From (2), E(Kf (PPC (n +1,pi,p2)) = E(Kf (PPC(n,pi,p2))) + 6[(i5-p2-4p2)n2 + ( f + 2p2 - f )n] + 71n + f = E(Kf (PPC(n,pi,p2))) + (45 - 3pi - 12p2)n2 + (3pi + 12p2 + 61)n + f and E(Kf (PPC (1,pi,p2))) = f. Using the above recurrence relation, we have E(Kf (PPC (n,pi,p2))) = (15 - pi - 4p2)n3 + (3pi + 12p2 +8)n2 - (2pi + 8p2 + y)n. □ Specially, by taking (pi,p2) = (1,0), (0,1) or (0,0), respectively, and Theorem 2.1, we have Corollary 2.2. ([8]) The Kirchhoff indices of the polyphenyl meta-chain Mn, the poly-phenyl ortho-chain On and the polyphenyl para-chain Pn are __15 Kf (Mn) = 14n3 + 11n2 - уn __27 Kf (On) = 11n3 + 20n2 - — n Kf (Pn) = 15n3 + 8n2 - у n 2.2 The Kirchhoff index of the random spiro chain In this section, we will consider the Kirchhoff index of the random spiro chain. Theorem 2.3. For n > 1, the expected value of the Kirchhoff index of the random spiro chain SPC (n,p1,p2) is E(Kf (SPC(n,p1,p2))) = (f - 36p1 - |p2)n3 + (12p1 + |p2 + 1Ц)п2 .25 50 5. — ( — p i ---p2--)n. v 18^ 9 2 6' Proof. Note that the spiro chain SPCn is obtained by attaching SPCn-1 a new terminal hexagon, we suppose that the terminal hexagon is spanned by vertices x1, x2, x3, x4, x5, x6, and the vertex x1 is un-1 (see Fig.3). Then (i) For any v e SPCn-1, r(x1, v) = r(un-1,v), r(x2, v) = r(un-1, v) + 5, r(x3, v) = r(un-1, v) + 3, r(x4, v) = r(un-1, v) + 3, r(x5, v) = r(un-1, v) + 3, r(x6, v) = r(un-1, v) + 6; (ii) SPCn-1 has 5(n - 1) + 1 vertices; 6 (iii) For k e {1, 2, 3,4, 5, 6}, £ r(xk,xH) = 35. i=1 So, we have r(x1|SPCn ) = r(un-1|SPCn-1) + 35 r(x2|SPCn ) = r(un-1|SPCn-1) + § x(5n-4)+§ +§ + f+4 = r(un-1 |SPCn-1) + 25 x (n - 1)-+ 35 r(x3|SPCn) = r(un-1|SPCn-1) + f x (n - 1) + 35 r(x4|SPCn) = r(un-1|SPCn-1) + f x (n - 1) + 35 r(x5|SPCn ) = r(x3|SPCn-1) r(x6|SPCn ) = r(x2|SPCn-1) where r(x|G) = J2 r(x, y), and yev (G) Kf (SPCn) = Kf (SPCn-1) + 5r(un-1|SPCn-1)+ (2.3) 175(n - 1) or 1 ЛЛ , —li-L +35 - 2EEr(vi,vj) i=1 j=1 = Kf (SPCn-1) + 5r(un-1| SPCn-1) + 175n 35 3 6 Then 175n 35 Kf (SPCn+1) = Kf (SPCn) + 5r(un|SPCn) + -— + — 6 ■ 2 (2.4) For a random spiro chain SPC(n, p1, p2 ), the resistance number r (un |SPC(n, p1, p2 )) is a random variable, and its expected value is denoted by Un = E (r(un|SPC (n,p1 ,p2))). By the expectation operator and (3), we can obtain a recursive relation for the expected value of the Kirchhoff number of a random spiro chain SPC(n, p1, p2 ) 175n 35 E (Kf (SPC (n + 1,p1,p2 )) = E(Kf (SPC (n,p1,p2))) + 5Un + — + — (2.5) 62 Now, we consider computing Un. (i) If SPCn ^ SPCn+1 with probability pi, then un is the vertex x3 or x5. Consequently, r(un\SPCn) is given by r(x3|SPCn) with probability pi. (ii) If SPCn ^ SPC'2+1 with probability p2, then un is the vertex x2 or x6. Consequently, r(un\SPCn) is given r(x2\SPCn) with probability p2. (iii) If SPCn ^ SPC3+1 with probability 1 - p1 - p2, then un is the vertex x4. Consequently, r(un\SPCn) is given by r(x4\SPCn) with probability 1 - p1 - p2. From (i)-(iii) above, we immediately obtain Un =r(x3\SPCn)pi + r(x2\SPCn)p2 + r(x4\SPCn)(1 - pi - p2) 20 35 =pi[r(un-i\SPC(n - 1,pi,p2)) + 2-(n - 1) + —] 3 6 25 35 + p2[r(un-i\SPC(n - 1,pi,p2)) + -r(n - 1) + "d 6 6 15 35 + (1 - pi - p2)[r(un-i\SPC (n - 1,pi,p2)) + ^r(n - 1) + —1 26 By applying the expectation operator to the above equation, we obtain 15 5 10 5 10 5 Un = Un-i + (y - ^pi - yp2)n + ^pi + yp2 - 3 And Ui = E(r(ui\SPC(1,pi,p2))) = 35, using the above recurrence relation, we have 15 5 5 2 25 5 5 Un = (T - 12pi - 3p2)n + (12 + 12pi + 3p2)n From (4), E (Kf (SPC (n +1,pi,p2)) = = E(Kf (SPC(n,pi,p2))) + 5[(f - i52pi - |p2)n2 + ( 1 + Ì2pi + 3 p2 )n] + Щ n + f and E(Kf (SPC (1,pi,p2))) = f. Using the above recurrence relation, we have E (Kf (SPC (n,pi,p2))) = (25 - 36 pi - 25 p2 )n3 + (12pi + ^ + Цу 25 50 5 — (— p +--p2--)n. v18^ 9 2 6' □ Specially, by taking (pi,p2) = (1,0), (0,1) or (0,0), respectively, and Theorem 2.3, we have Corollary 2.4. ([8]) The Kirchhoff indices of the spiro meta-chain Mn, the spiro orthochain On and the spiro para-chain Pn are 50 3 25 2 5 Kf (Mn) = 50 n3 + 25 n2 - 5 n Kf (On) = — n3 + —n2 - 85n jy n> 36 1 4 18 Kf (Pn)=25 n3+125 n2+5 n. 2.3 A relation between E(Kf (PPC)) and E(Kf (SPC)) Since a spiro chain can be obtained from a polyphenyl chain by squeezing off its cut edges, it is straightforward by Rayleigh short-cut principle in the classical theory of electricity that the Kirchhoff index of the spiro chain is less than the polyphenyl chain. In fact, a relation between the Kirchhoff indices of a polyphenyl chain and its corresponding spiro chain obtained by squeezing off its cut edges was given in [8]. Here, we can also obtain a relation between the expected values of their Kirchhoff indices of the random polyphenyl chain PPC(n,p1,p2) and the random spiro chain SPC(n,p1,p2) with the same probabilities p1 andp2 from Theorems 2.1 and 2.3. Theorem 2.5. For a random polyphenyl chain PPC(n,p1,p2) and a random spiro chain SPC (n, p1, p2 ) with n hexagons, the expected values of their Kirchhoff indices are related as 50E(Kf (PPC(n,p1,p2))) = 72E(Kf (SPC(n,p1,p2))) + 300n3 - 350n2 - 335n. Theorem 2.5 also shows that the expected value of Kirchhoff index of the random spiro chain is less than the random polyphenyl chain. In fact, for n > 2, E(Kf (SPC(n, p1,p2))) < 35E(Kf (PPC(n,p1,p2))). The reason is quite obvious. Dividing both sides of the equation in Theorem 2.5 yields 36 67 E(Kf (PPC(n,p1,p2))) = -E(Kf (SPC(n,p1,p2))) + 6n3 - 7n2 - — n 25 10 and it is easily seen that for n > 2, 6n3 - 7n2 - 67n > 0. 2.4 The average value of the Kirchhoff index Let Gn is the set of all polyphenyl chains with n hexagons. The average value of the Kirchhoff indices with respect to G n is Kfavr (Gn) = E Kf (G). |Gn| GeG„ In order to obtain the average value of the Kirchhoff indices with respect to Gn, we only need to take p1 = p2 = 1 in the random polyphenyl chain PPC(n,p1,p2), i.e., the average value of the Kirchhoff indices with respect to Gn is just the expected value of the Kirchhoff index of the random polyphenyl chain PPC(n,p1,p2) for p1 = p2 = 1. From Theorem 2.1, we have Theorem 2.6. The average value of the Kirchhoff indices with respect to Gn is Kfavr(Gn) = 40n3 + 13n2 - 53n. 36 Similarly, let Gn is the set of all spiro chains with n hexagons. The average value of the Kirchhoff indices with respect to Gn is Kfavr (Gn) = E Kf (G). |Gn| oeGn And the average value of the Kirchhoff indices with respect to Gn is just the the expected value of the Kirchhoff index of the random spiro chain S PC (n,pl,p2) for pi = p2 = 1. From Theorem 2.3, we have Theorem 2.7. The average value of the Kirchhoff indices with respect to Gn is . 275 3 125 2 40 Kfavr [Gn) = -54n + -9-n - 27n. Acknowledgments. The authors would like to thank the anonymous referees for their valuable comments and suggestions on the original manuscript. References [1] D. Babic, D. J. Klein, I. Lukovits, S. Nikolic, N. Trinajstic, Resistance- distance matrix: A computational algorithm and its application, Int. J. Quantum Chem. 90 (2002), 166-176. [2] Y. Bai, B. Zhao, P. Zhao, Extremal Merrifield-Simmons index and Hosoya index of polyphenyl chains, MATCH Commun. Math. Comput. Chem. 62 (2009), 649-656. [3] R. B. Bapat, I. Gutman, W. J. Xiao, A simple method for computing resistance distance, Z. Naturforsch 58a (2003), 494-498. [4] D. Bonchev, E. J. Markel, A. H. Dekmezian, Long chain branch polymer chain dimensions: application of topology to the Zimm-Stockmayer model, Polymer 43 (2002), 203-222. [5] X. Chen, B. Zhao, P. Zhao, Six-membered ring spiro chains with extremal Merrifild-Simmons index and Hosoya index, MATCH Commun. Math. Comput. Chem. 62 (2009), 657-665. [6] H. Deng, On the minimal Kirchhoff indices of graphs with a given number of cut edges, MATCH Commun. Math. Comput. Chem. 63 (2010), 171-180. [7] H. Deng, Wiener indices of spiro and polyphenyl hexagonal chains, Mathematical and Computer Modelling 55 (2012), 634-644. [8] H. Deng, Z. Tang, Kirchhoff indices of spiro and polyphenyl hexagonal chains, accepted by Util. Math. [9] T. Došlic, F. Mal0y, Chain hexagonal cacti: Matchings and independent sets, Discrete Math. 310 (2010), 1676-1690. [10] T. Doslic, M. S. Litz, Matchings and independent sets in polyphenylene chains, MATCH Commun. Math. Comput. Chem. 67 (2012), 313-330. [11] D. R. Flower, On the properties of bit string-based measures of chemical similarity, J. Chem. Inf Comput. Sci. 38 (1998), 379-386. [12] Q. Guo, H. Deng, The extremal Kirchhoff index of a class of unicyclic graphs, MATCH Com-mun. Math. Comput. Chem. 61 (2009), 713-722. [13] D. J. Klein, M. Randic, Resistance distance, J. Math. Chem. 12 (1993), 81-95. [14] D. J. Klein, Graph geometry, graph metrics and Wiener, MATCH Commun. Math. Comput. Chem. 35 (1997), 7-27. [15] D. J. Klein, H. Y. Zhu, Distances and volumina for graphs, J. Math. Chem. 23 (1998), 179-195. [16] D. J. Klein, Resistance-distance sum rules, Croat. Chem. Acta 75 (2002), 633-649. [17] Q. R. Li, Q. Yang, H. Yin, S. Yang, Analysis of by-products from improved Ullmann reaction using TOFMS and GCTOFMS, J. Univ. Sci. Technol. China 34 (2004), 335-341. [18] J. L. Palacios, Closed-form formulas for Kirchhoff index, Int. J. Quantum Chem. 81 (2001), 135-140. [19] J. L. Palacios, Resistance distance in graphs and random walks, Int. J. Quantum Chem. 81 (2001), 29-33. [20] S. TepavCeviC, A. T. Wroble, M. Bissen, D. J. Wallace, Y. Choi, L. Hanley, Photoemission studies of polythiophene and polyphenyl films produced via surface polymerization by ionassisted deposition, J. Phys. Chem. B 109 (2005), 7134-7140. [21] Y. J. Yang, X. Y. Jiang, Unicyclic graphs with extremal Kirchhoff index, MATCH Commun. Math Comput. Chem. 60 (2008), 107-120. [22] Y. Yang, H. Zhang, Kirchhoff index of linear hexagonal chains, Int. J. Quantum Chem. 108 (2008), 503-512. [23] W. Yang, F. Zhang, Wiener index in random polyphenyl chains, MATCH Commun. Math. Comput. Chem. 68 (2012), 371-376. [24] H. Zhang, Y. Yang, Resistance distance and Kirchhoff index in circulant graphs, Int. J. Quantum Chem. 107 (2007), 330-339. [25] W. Zhang, H. Deng, The second maximal and minimal Kirchhoff indices of unicyclic graphs, MATCH Commun. Math. Comput. Chem. 61 (2009), 683-695. [26] H. Zhang, X. Jiang, Y. Yang, Bicyclic graphs with extremal Kirchhoff index, MATCH Commun. Math. Comput. Chem. 61 (2009), 697-712. [27] P. Zhao, B. Zhao, X. Chen, Y. Bai, Two classes of chains with maximal and minimal total n-electron energy, MATCH Commun. Math. Comput. Chem. 62 (2009), 525-536. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 209-222 Extending patches to fullerenes Christina Graves The University of Texas at Tyler, Department of Mathematics Tyler, TX 75799, USA Jennifer McLoud-Mann University of Washington Bothell, School of STEM Bothell, WA 98011, USA Kristen Stagg Rovira The University of Texas at Tyler, Department of Mathematics Tyler, TX 75799, USA Received 10 May 2013, accepted 7 June 2014, published online 8 December 2014 In this paper we consider fullerene patches that can be extended to pseudconvex patches. We show that all fullerene disks with three or fewer pentagons can be extended to pseudo-convex patches, and that all pseudoconvex patches can be extended to fullerenes. Keywords: Fullerenes, fullerene patches, boundary codes, pseudoconvex patches. Math. Subj. Class.: 05C10, 05C75, 92E10 1 Preliminaries A fullerene is a trivalent planar graph whose faces consist solely of hexagons and pentagons. A fullerene patch, or patch, is similar; it is a planar graph where all faces are hexagons and pentagons except one outer face, with vertices not on the outer face having degree 3 and vertices on the outer face having degree 2 or 3. An easy way to create a fullerene patch from a fullerene is to trace a closed circuit on a fullerene and delete all faces on one side of the circuit. However, it is not so easy to create a fullerene from a fullerene patch. In fact, there are many fullerene patches that cannot be extended to a fullerene as we will see later. We do find a family of patches that can be extended to fullerenes. E-mail addresses: cgraves@uttyler.edu (Christina Graves), jmcloud@uw.edu (Jennifer McLoud-Mann), kstagg@uttyler.edu (Kristen Stagg Rovira) Abstract Figure 1: A patch and its complement. The patch on the left has boundary code 223(2223)2 and side parameters [1,0,1,0,1] and the patch on the right has boundary code 332(3332)2. An obvious characteristic of interest of a fullerene patch is the boundary. The boundary code of a patch is a sequence of 2's and 3's corresponding to the valences of the vertices on the outer face listed in cyclic order. It does not matter which vertex we start with nor which direction we travel around the patch; hence we make no distinction between a boundary code and its cyclic permutations or inverse. Given a boundary code s, the complement of s, denoted sc is a sequence of 2's and 3's with a 2 every place s has a 3, and a 3 every place s has a 2. To determine if a fullerene patch extends to a fullerene, we need to consider the following question: Given a fullerene patch П with boundary code s, does there exist a fullerene patch nc with boundary code sc? If Пс exists, we can identify the vertices and edges on the boundary of П with the corresponding vertices and edges on the boundary of Пс to get a fullerene (see Figure 1). The general question of interest is known in the literature as the PentHex Puzzle: Given a sequence of 2's and 3's, does there exist a fullerene patch with that sequence as its boundary code? Some variations on this question were explored in [3], [1], [6], and [4] among other places. For large patches, the boundary code can be unwieldy to work with. Generalizing the definitions from [8], [7], [5] and [2], we define the following. Definition 1.1. 1. A break edge is an edge on the boundary whose endpoints are both of degree two. 2. A bend edge is an edge on the boundary whose endpoints are both of degree three. 3. A side of a patch is a path on the boundary between a consecutive pair of break edges, including the break edges. The length of a side is the number of degree three vertices on the side. 4. A straight side is a side with no bend edges. 5. A bent side is a side containing at least one bend edge. A straight segment of a bent side is a subpath of a bent side between either a break edge and the closest bend edge, or between two consecutive bend edges. We can now think of the boundary of a patch as sections of straight sides and straight segments connected by break edges or bend edges. A patch with no bend edges, called a pseudoconvex patch, has only straight sides and thus has boundary code 2(23)£l 2(23)^2 • • • 2(23)£s. Rather than writing a sequence of 2's and 3's to describe our boundary, we describe the boundary by the lengths of the straight sides \l1, l2,..., 4], called side parameters of the patch. If the patch has bend edges, we would still like to consolidate the information found in the boundary code. A bent side with consecutive straight segments of lengths ai, a,2, .. . at can be described by (a1,... ,at). Thus, a patch with five straight sides of length 5 followed by a bent side with five straight segments of length 1 (see Figure 3) can be described by the side parameters [5,5, 5, 5, 5, (1,1,1,1,1)]. A patch with no break edges has zero sides and thus cannot be described by side parameters. If such a patch also has no bend edges, then we describe it by its boundary code (23)£o where 4 is the number of faces on the boundary. If a patch has no break edges but does have bend edges, then we describe it by its boundary code 3(32)ai-13(32)°2-1 • • • 3(32)at-1 where a1, a2,... ,at represent the lengths of the straight segments in cyclic order. The advantage of describing a patch by its side parameters rather than its boundary code is that this notation makes it effortless to find the number of sides (and hence break edges) of a patch. In fact, we can also tell exactly how many pentagons a patch must have by using Euler's formula. Lemma 1.2. In afullerene patch, the number of break edges s, the number of bend edges e, and the number of pentagons p are related by p = 6 — s + e. Proof. It is well-known (see for instance [1]) that the number of pentagons in a patch is equal to 6 — d2 + d3 where di is the number of degree i vertices on the boundary. By definition, the number of degree 2 vertices and the number of degree three vertices on the boundary are the same except on a break edge or a bend edge. Each break edge increases the number of degree 2 vertices by 1, and each bend edge increases the number of degree 3 vertices by 1. □ In this paper, we investigate two specific types of fullerene patches: pseudoconvex patches and fullerene disks or simply disks. A disk of radius r is a fullerene patch with a central face, and every face on the boundary is distance (measured as graph distance in the dual) exactly r from the central face. Our ultimate goal is to find a family of patches that can be extended to fullerenes. We begin by showing that a pseudoconvex patch can be extended to a fullerene. We then show that all disks with three or fewer pentagons can be extended to fullerenes. 2 Pseudoconvex Patches Because a pseudoconvex patch has no bend edges, its boundary can be described by its side parameters [4, • • • , 4] if it has s > 1 sides or by the boundary code (23)£° if it has zero sides. The relationships between the lengths of these sides were explored in detail and summarized in Theorem 3.6 of [8]. We include a weaker lemma here. Lemma 2.1. The side parameters [4, ■■■ ,4] of a pseudoconvex patch, excluding the patch with side parameters [0,0,0,0,0], satisfy the inequality 4 + • • • + £s > 6 — s for 1 < s < 6. Proof. If s = 1 or s = 2, the result follows directly from [8]. If s = 3, the side parameters are not of the form [0,1, k] or [0,0, k] for k > 0 so the result holds. In the case with s = 4 sides, at least two of the side lengths are non-zero and the result holds. Finally, if s = 5 or s = 6, all parameters are nonnegative so the sum is nonnegative. Thus the inequality is satisfied except in the patch consisting of a single pentagon with side parameters [0, 0, 0,0,0]. □ To extend a pseudoconvex patch to a fullerene, we need to find a complement for each pseudoconvex patch. The following lemma gives a constructive method for finding the complement patch. Lemma 2.2. There exists a patch with boundary code 3(32)£l3(32)^2 ••• 3(32)£s if 4 + 4 +-----+ 4 > 6 — s and 1 < s < 6. Proof. Start with the patch having boundary code 2(23)5 (i.e. one side of length 5) as shown in Figure 2. Add 4 + 4 + • • • + 4 + s — 6 rings of hexagons to this patch to create a patch with side parameters [4 +-----+ 4 + s — 1]. Next, add a pentagon to the break edge and add hexagons on the boundary everywhere else to yield a patch with boundary code (23)£i+-+£*+s. We now add hexagons and pentagons to the boundary in the following way. First put a pentagon somewhere on the boundary. Moving clockwise around the boundary, place 4 hexagons followed by a pentagon; then place 4 hexagons followed by a pentagon, and continue this process until the patch has a completely new outer ring of hexagons and pentagons (see Figure 2). This constructed patch has the desired boundary. □ Theorem 2.3. All pseudoconvex patches can be extended to fullerenes. Proof. For a pseudoconvex patch consisting of one pentagon with side parameters [0,0,0,0,0], we use a stereographic projection of the dodecahedron onto the plane as the complement patch. For a pseudoconvex patch having side parameters [4,..., 4] satisfying the condition 4 + ••• + 4 > 6 — s, we create the complement patch with boundary code 3(32)£l 3(32)^2 • • • 3(32)£s as described in Lemma 2.2 and identify the boundaries to create a fullerene. Given a pseu-doconvex patch with no sides and boundary code (23)£°, we create a second patch (32)£° identical to the first and then identify corresponding edges and vertices appropriately to create a fullerene. □ Figure 2: Creating a patch with boundary code 3(32)23(32)33(32)0. The 2(23)5 patch is shown on the left, the (23)8 patch in the middle, and the desired patch on the right. 3 Disks Because pseudoconvex patches extend to fullerenes, a patch that extends to a pseudoconvex patch also extends to a fullerene. One type of patch to explore is a fullerene disk. We will show that all disks with three or fewer pentagons extend to pseudoconvex patches; however, a disk with four or more pentagons may not extend to a fullerene. Consider a disk with four pentagons having side parameters [5, 5,5,5,5, (1,1,1,1,1)] as shown in Figure 3. If this patch could be extended to a fullerene, we would need to place a face adjacent to the four shaded pentagons. Such a face would have to have at least seven edges which is not allowed in a fullerene. Thus we restrict our attention to disks with three or less pentagons. Figure 3: A disk with side parameters [5,5,5,5, 5, (1,1,1,1,1)]. This patch cannot be extended to a fullerene. Disks are a nice family of patches to study because all disks can be constructed by repeatedly adding layers of faces to a smaller disk. A layer is a collection of faces placed on the boundary of a disk of a radius r so that each new face is distance r + 1 from the i —Ì-1...........................i-\— i (3.1.1) a2 (3.1.3) (3.1.4) Figure 4: Depiction of Lemma 3.1 and how adding hexagons changes side parameters. central face and and no faces from the original disk are on the new boundary. The word layer can be slightly misleading, though, because there can be faces added which are not on the boundary of the new patch. Figures 5, 6, and 7 show partial patches with one layer added. Because many of the remaining arguments involve adding layers of hexagons to existing patches, it is convenient to discuss how such an action affects the side parameters. Adding a row of hexagons to a side of a patch means that each edge of the original boundary is incident with a newly added hexagon. Figure 4 and Lemma 3.1 demonstrate this action. Lemma 3.1. 3.1.1. Adding a row of hexagons to the boundary of a straight side increases its length by 1. 3.1.2. Adding a partial layer of hexagons to a bent side with middle parameters larger than 1 keeps the outer parameters the same and decreases the middle parameters by 1. 3.1.3. Adding a row of hexagons to a straight segment a2 of the bent side (ai, a2) with a i > 1, not including the break edge, decreases ai by 1, keeps a2 the same, and increases the side following a2 by 1. 3.1.4. Adding a row of hexagons to the straight segment a2 of a bent side (ai, a2, a3) with ai,a3 > 1, decreases ai and a3 by 1 and increases a2 by 1. (3.1.2) The lemma tells us how the side parameters of a patch change when we add a layer of hexagons for almost all situations. We will consider special cases for adding a layer of hexagons for patches having a bent side containing a middle parameter equal to one in later proofs. The next three lemmas give the side parameters for all disks with three or fewer pentagons. Lemma 3.2. A disk with at most one pentagon has side parameters satisfying one of the following: A1. [ti,^2,4,4,4,4] B1. [ti ,t2,ts,t4,t5] B2. [ti,t2,t3,t4,t5, (01,02)] Proof. A disk consisting only of hexagons has the form [r, r, r, r, r, r]. For a disk containing one pentagon on the boundary, there are two cases. Either the disk consists of only one face, a pentagon, and has side parameters [0,0,0,0,0], or the disk can be constructed by adding faces to a disk containing no pentagons. Starting with a disk satisfying condition A1, adding a layer of faces with a pentagon on a break edge and hexagons everywhere else yields a B1 patch, and adding a layer with a pentagon on a straight side yields a B2 patch. If a disk with one pentagon does not have its pentagon on the boundary, then the patch can be viewed as a disk containing one pentagon on the boundary with layers of hexagons added to it. Using Lemma 3.1, adding layers of hexagons to a B1 or B2 patch yields a B1 or B2 patch respectively. □ Lemma 3.3. A disk with two pentagons has side parameters satisfying one of the following: C1. [ti,t2,t3,t4] C2. [ti,t2,t3,t4, (01,02)] C3. [ti,t2,t3,t4, (01,02), (61,62)] C4. [ti,t2,t3, (0i,02),t4, (61,62)] C5. [ti,t2, (0i,02),t3,t4, (61,62)] C6. [ti,t2,t3,t4,t5, (01,02,03)] Proof. We start by considering disks containing two pentagons with different distances from the central face, and the pentagon farthest from the central face on the boundary. Starting with a disk containing one pentagon satisfying condition B1, adding a layer of faces with a pentagon on a break edge yields a C1 patch, and adding a layer with a pentagon on a straight side would yield a C2 patch. Starting with a disk containing one pentagon satisfying condition B2, adding a layer of faces with a pentagon on a break edge results in a C2 patch, adding a layer with a pentagon on a straight side yields a C3, C4, or C5 patch, and adding a layer with a pentagon on a straight segment of the bent side would yield a C6 patch. / ~r A > J Figure 5: Adding a pentagon to a bend edge. The red edges are bend edges in the new patch. When adding a layer of faces with a pentagon on the bend edge of a B2 patch, there are three cases to consider. If a1 = a2 = 1, the resulting patch is a C1 patch. If 1 = a1 < a2, the new patch is a C2 patch. Lastly, if a1, a2 > 1, the resulting patch is a C6 patch, with the middle component of the bent edge having length 1 (see Figure 5). Now consider disks where both pentagons are the same distance from the central face and are both on the boundary. There are no disks of radius 0 containing two pentagons, so we can construct these patches by starting with a disk containing no pentagons and adding the pentagons to the same layer. Starting with a patch with no pentagons satisfying condition A1, adding a layer with two pentagons on two different break edges yields a C1 patch, adding a layer with one pentagon on a break edge and one on a straight side would result in a C2 patch, adding a layer with two pentagons on two different straight sides would yield a C3, C4, or C5 patch, and adding a layer with two pentagons on the same straight side would result in a C6 patch. Notice every disk we have constructed has a pentagon on the boundary. To construct disks with two pentagons neither of which is on the boundary, we simply construct a disk with two pentagons at least one of which is on the boundary, and add layers of hexagons. By Lemma 3.1, adding layers of hexagons to a Ci patch yields a Ci patch for 1 < i < 5. When considering a C6 patch, we assume without loss of generality that ai < a3. Adding less than a2 + a1 - 1 layers of hexagons to a C6 patch results in a C6 patch, and adding at least a2 + a1 - 1 layers of hexagons to a C6 patch would yield a C1 or C2 patch. The C1 patch results if a1 = a3 and the C2 patch results otherwise. □ Lemma 3.4. A disk with three pentagons has side parameters satisfying one ofthefollow- D1. [ti,t2M D2. [ti,Ì2,h, (ai,a2)] D3. [ti,Ì2,h, (ai,a2), (61,62)] D4. [h,l2, (ai,a2),4, (61,62)] D5. [^2,4, (ai,a2), (61,62), (01,05)] D6. [^2, (ai,a2),4, (61,62), (01,02)] D7. [h, (ai,a2),^2, (61,62)^3, (01,02)] D8. [4,4,4,4, (ai,a2,a3)] D9. [4,4,4,4, (ai,a2,a3), (bi, Л )] D10. [4, 4, 4, (ai, a2, a3), 4, (bi, ,b2 )] D11. [4, 4, (ai,a2,a3),4,4, (bi, ,b2 )] D12. [4, 4, 4, (ai, a2, a3, a4)] Proof. We begin by showing that that all disks with three pentagons having at least one pentagon on the boundary satisfy the criteria given. To show that all three-pentagon disks have the listed side parameters, we add layers of hexagons to three-pentagon disks with a pentagon on the boundary. First, consider a three-pentagon disk with exactly one pentagon on the boundary. Such disks are created by starting with a two-pentagon disk and adding a layer of faces containing exactly one pentagon. Starting with a C1 patch, adding a layer of faces with a pentagon on a break edge yields a D1 patch, and adding a layer with a pentagon on a straight side yields a D2 patch. Starting with a C2 patch, adding a layer of faces with a pentagon on a break edge yields a D2 patch, and adding a layer of faces with a pentagon on a straight side yields a D3 or D4 patch. If a layer of faces is added to a C2 patch with a pentagon on a bend edge, then (following the proof of Lemma 3.3) a D1, D2, or D8 patch is obtained. If a layer is added with a pentagon on a straight straight segment of a bent side of a C2 patch, the resulting disk is a D8 patch. Starting with a C3 - C5 patch, adding a layer of faces with a pentagon on a straight side yields a D 5 - D 7 patch, and adding a layer with a pentagon on a straight segment of a bent side yields a D9, D10, or D11 patch. Adding a layer of faces with a pentagon on a break edge of a C3 patch yields a D3, D4, or D8 patch, with a D8 patch occurring if the break edge is between the two bent sides. Adding a layer with a pentagon on a break edge of a C4 or C5 patch yields a D3 or D4 patch. Finally, adding a layer of faces with a pentagon on a bend edge on a C3 - C5 patch yields a D2, D3, D4, D9, D10, or D11 patch. The last case involves adding a layer of faces with exactly one pentagon to a C6 patch. Adding a layer of faces with a pentagon on a break edge results in a D8 patch, adding a layer with a pentagon on a straight side yields a D8 - D11 patch, and adding a layer with a pentagon on a straight segment of a bent side results in a D12 patch. When adding a layer of faces with a pentagon on a bend edge of a C6 patch there are a few cases to consider. Without loss of generality, assume that the pentagon is added to the bend edge between the straight segment of length ai and the straight segment of length a2. If ai > 1 and a2 > 2 the resulting patch is a D12 patch with the new bent side having lengths (ai - 1,1, a2 - 2, a3) (see Figure 6). In the cases were either ai = 1 or a2 g {1,2}, some care needs to be taken. In these cases, the length of a3 can affect the type of new patch; however, in every case, the resulting patch is a D2 or D8 patch. Figure 7 shows these cases in full detail. Now consider the three-pentagon disks with exactly two pentagons on the boundary. To create such disks, we start with either a B1 or B2 patch. Adding a layer of faces with two pentagons to a B1 patch is very similar to adding two pentagons to an A1 patch, which was explored in the proof of Lemma 3.3. Thus adding a layer with two pentagons on a B1 patch results in a D1 - D4 or D8 patch. Using similar arguments as before, adding a layer Figure 6: Adding a pentagon to a bend edge in a C6 patch. The red edges are bend edges in the new patch. Figure 7: Adding a pentagon to a bend edge in a C6 patch. The red edges are bend edges in the new patch. with two pentagons to a B2 patch results in a D2 - D12 patch if neither pentagon is on a bend edge. Adding a layer with one of the pentagons on a bend edge requires the cases seen in Figure 5. Adding a layer with one pentagon on the bend edge and one pentagon on a break edge yields a D1, D2, or D8 patch. Adding a layer with one pentagon on the bend edge and one pentagon on a straight side results in a D2 - D4 or D9 - D11 patch. Finally, adding a layer with one pentagon on the bend edge and one pentagon on a straight segment of the bent side (which can only be done if the straight segment is greater than 1) results in a D8 patch. Three-pentagon disks with all three pentagons on the boundary are formed by adding a layer to a disk of hexagons. Starting with an A1 patch, adding a layer with three pentagons results in a D1 - D11 patch using similar arguments to those above. Notice every patch we have constructed has a pentagon on the boundary. For those that do not, we simply construct a previous patch and add layers of hexagons. By Lemma 3.1, adding layer of hexagons to a Di patch yields a Di patch for 1 < i < 7. For disks that have a bent side with at least three straight segments, the resulting patch's side parameters differ based on the number of layers added. For instance, adding one layer of hexagons to a patch with the side parameters [f ь f2, f3, f4, (a1,a2,a3)] and a2 > 1 results in a patch with side parameters \l1 + 1, l2 + 1, f3 + 1, f4 + 1, (a1,a2 - 1, a3)]. Furthermore, adding a layer of hexagons to a patch with side parameters [f 1, l2, f 3, f 4, ( a 1, 1, a3 )] yields a patch with side parameters [4 + 1, f2 + 1, f3 + 1, f4 + 1, (a1 - 1,1, a3 - 1)]. Thus, adding less than a2 + min{a1, a3 } - 1 rings to a D j patch will yield a Dj patch for 8 < j < 11. Adding at least a2 + min{a1, a3} - 1 rings to a D8 patch yields a D1 patch if a1 = a3 or D2 patch otherwise. Adding at least a2 + min{a1, a3} - 1 rings to a D9 patch yields a D2 patch if a1 = a3, a D3 patch if a1 < a3, or a D8 patch if a1 > a3. Adding at least a2 + min{a1, a3} - 1 rings to a D10 or D11 patch yields a D2 patch if a1 = a3 or D4 patch otherwise. Now let us consider adding rings of hexagons to a D12 patch; without loss of generality assume a2 < a3. Adding less than a2 + min{a1 - 1, |_a2 j} rings yields another D12 patch. Adding exactly a2 + min{a1 - 1, |_°2 j} rings yields a D1, D2, or D8 patch. From arguments above, adding further rings of hexagons gives a D1, D2, or D8 patch. □ Now that we have shown the side parameters of disks with three or fewer pentagons fall into different classes, we show that any patch with these side parameters can be extended to a pseudoconvex patch. Lemma 3.5. Any patch with one of the following descriptions of its side parameters can be extended to a pseudoconvex patch: 1. [4 *2,.. .,4] 2. [4 Ì2,.. f (a1, a2)] 3. [4 Ì2,.. f . , f s , (a1,a2), (61,62)] 4. [4 Ì2,.. f . , f s , (a1,a2),fs+1 ,...,ft, (61, 62)] 5. [4 Ì2,.. f • : ^s ? (a1,a2), (61,62), (c1,C2)] 6. Ì2,.. . , f , (a1,a2), (61, 62), fs+1,. . ,ft, (C1,c2)] 7. [4 Ì2,.. . , f , s (a1,a2),fs+1, ...,ft, (61, 62), f t+1, .. u Figure 8: Filling in the bent side (a1,a2) and partially filling in the bent side (ab a2, a3). 8. [4,4,..., 4, (ai,a2,a3)j 9. [4,4,..., 4, (ai,a2), (61,62,63)] 10. [ii, l2,..., Is, (ai, a2), 4+i, ...,tu (61, 62, 63)] 11. [ii, 4,.. ., 4, (ai, a2, a3, a4)]. Proof. Let П be a patch with one of the following descriptions above. If П = [4,4,..., is], then it has no bend edges and is pseudoconvex. If П = [4,4,..., is, (ai, a2)], then adding ai rows of a2 hexagons to the straight segment a2, not including the break edge, yields the pseudoconvex patch [4 + ab i2,..., 4-ъ 4 + a2] by Lemma 3.1 (see Figure 8). If the side parameters of П are of the form [4,4 ,...,4, (ai,a2), (6i,62)] or [4,4,..., 4, (ai, a2), 4+ъ ..., 4, (6i, 62)], then adding ai rows of a2 hexagons to the straight segment a2, not including the break edge, yields either [4, 4, ..., 4 + a2, (6i + ai, 62)] or [4,4,..., 4 + a2,^s+i + ab...,4, (6i, 62 )] which are both extendable by a previous case. If П has one of the forms [4, 4, .. ., 4, (ai,a2), (6i, 62), (ci,c2)], [4, 4,.. ., 4, (ai, a2), (6i, 62), 4+i,. .., 4, (ci, C2)], or [4,4,..., 4, (aba2),4+i,...,4, (6b62),4+i,...,4, (ci,c2)], then adding ai rows of a2 hexagons to the straight segment a2, not including the break edge, yields a patch with one of the descriptions [ei, 4, .. ., is + a2, (6i + ai, 62), (ci, c2)], [ei, 4, .. ., is + a2, (6i + ai, 62), is+i,. .., it, (ci, c2)], or [ii,i2, .. ., is + a2,is+i + ai,. .. ,it, (6i, 62),it+i, .. ., 4, (ci, c2)] which are all extendable by previous cases. Suppose П = [£b £2,..., £s, (ai, a2, a3)]. Without loss of generality, assume a1 < a3. Adding a1 -1 rows of hexagons to the a2 straight segment results in a patch with parameters [£i, £2,. .. ,£s, (1, ai + a2 - 1, a3 - ai + 1)] by Lemma 3.1. Adding one more row yields a patch whose parameters are [£1, £2,. .. ,£s-1, (£s + ai + a,2,a,3 - ai)] if a1 < a3 or [£s + ai + a2 + £1, £2, £3, .. ., £s-i] if a1 = a3. Either of these is extendable by a previous case. If П has the form [£1, £2,. .., £s, (ai,a2), (61,62,63)] or [£1, £2, ..., £s, (ai ,a2),£s+i,...,£t, (61,62,63)], then adding a1 rows of a2 hexagons to the straight segment a2, not including the break edge, yields a patch satisfying the form [£1, £2,.. ., £s + a2, (61 + ai, 62, 63)] or [£1, £2,..., £s + a2, £s+i + ai,..., £t, (61, 62, 63)] which are both extendable by previous cases. Suppose П = [£1, £2,..., £s, (a1, a2, a3, a4)]. Without loss of generality, we may assume that a1 < a4. Note that a1 < a3 + a4. We have the following four cases to consider: (i) 0 < a1 < a3, (ii) a1 = a3, (iii) a3 < a1 < a3 + a4, and (iv) a1 = a3 + a4. In each of the four cases we will add a1 rows of hexagons to the straight segment a2 in order to create a patch which is extendable by a previous case. For (i) and (ii), we begin by adding a1 - 1 rows of hexagons to create a patch with parameters [£1, £2,..., £s, (1, a1 + a2 - 1, a3 - a1 + 1, a4)] by Lemma 3.1. Adding one more row will yield [£i,£2,... ,£s-i, (£s + ai + a2, a3 - ai, a4)] in case (i) and [£1, £2, .. ., £s-i, (£s + ai + a2 - 1,1,a4 - 1)] in case (ii). For (iii) and (iv), we begin by adding a3 - 1 rows of hexagons to straight side a2 to yield [£1, £2,..., £s, (a1 - a3 + 1, a2 + a3 - 1,1, a4)] by Lemma 3.1. Adding another row yields a [£1, £2,..., £s, (a1 - a3, a2 + a3 - 1,1, a4 - 1)] patch. For (iii), adding a1 - a3 -1 rows gives [£i,£2,..., £s, (1, a2 + a3 - 1,1, a4 + a3 - ai)] and finally adding one more row gives a [£i,£2,. .. ,£s-i, (£s + a2 + a3 - 1,1, a4 + a3 - ai - 1)] Figure 9: Partially filling in the bent side (a1, a2, a3, a4). patch (see Figure 9). For (iv), adding a1 — a3 — 2 rows of hexagons gives \l1, l2,..., 4, (2, a2 + a3 — 1,1,1)], an additional row yields [4, 4,. .., 4, (1,a2 + аз, 4)], and finally adding another row gives [4,4,..., 4-ъ (4 + a2 + a3, 4)]. □ Theorem 3.6. All disks containing at most three pentagons extend to a fullerene. Proof. All disks containing at most three pentagons will have side parameters satisfying the conditions of Lemma 3.2, Lemma 3.3, or Lemma 3.4. By Lemma 3.5, these disks can be extended to pseudoconvex patches. By Theorem 2.3, those pseudoconvex patches can be extended to fullerenes. □ References [1] P. Bonsma and F. Breuer, Finding fullerene patches in polynomial time. In ISAAC 2009, volume 5878 of LNCS (2009), 750-759. [2] J. Bornhoft, G. Brinkmann and J. Greinus, Pentagon-hexagon-patches with short boundaries, European Journal of Combinatorics 24 (2003), 517-529. [3] G. Brinkmann and A. W. M. Dress, PentHex puzzles: A reliable and efficient top-down approach to fullerene-structure enumeration, Advance in Applied Math 21 (1998), 473-480. [4] G. Brinkmann, U.V. Nathusius and A.H.R. Palser, A constructive enumeration of nanotube caps, Discrete Applied Mathematcis 116 (2002), 55-71. [5] G. Brinkmann and N. Van Cleemput, Classification and Generation of Nanocones, Discrete Applied Mathematics 159 (2011), 1528-1539. [6] M. Deza, P. W. Fowler, and V. Grishukhin. Allowed boundary sequences for fused polycyclic patches and related algorithmic problems. J. Chem. Inf. Comput. Sci. 41 (2001), 300-308. [7] J.E. Graver and C. Graves, Fullerene Patches I, Ars Mathematica Contemporanea 3 (2010), 104-120. [8] C. Graves and J. McLoud-Mann, Side lengths of pseudoconvex fullerene patches, Ars Mathematica Contemporanea 5 (2012), 291-302. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 223-242 Fast recognition of partial star products and quasi cartesian products* Marc Hellmuth f Center for Bioinformatics, Saarland University, D - 66041 Saarbrücken, Germany Wilfried Imrich Chair of Applied Mathematics, Montanuniversität, A-8700 Leoben, Austria Tomas Kupka Department of Applied Mathematics, VSB-Technical University of Ostrava, Ostrava, 70833, Czech Republic Received 12 August 2013, accepted 16 December 2013, published online 8 December 2014 This paper is concerned with the fast computation of a relation d on the edge set of connected graphs that plays a decisive role in the recognition of approximate Cartesian products, the weak reconstruction of Cartesian products, and the recognition of Cartesian graph bundles with a triangle free basis. A special case of d is the relation S*, whose convex closure yields the product relation a that induces the prime factor decomposition of connected graphs with respect to the Cartesian product. For the construction of d so-called Partial Star Products are of particular interest. Several special data structures are used that allow to compute Partial Star Products in constant time. These computations are tuned to the recognition of approximate graph products, but also lead to a linear time algorithm for the computation of S* for graphs with maximum bounded degree. Furthermore, we define quasi Cartesian products as graphs with non-trivial S*. We provide several examples, and show that quasi Cartesian products can be recognized in linear time for graphs with bounded maximum degree. Finally, we note that quasi products can be recognized in sublinear time with a parallelized algorithm. *We thank Lydia Ostermeier for her insightful comments on graph bundles, as well as for the suggestion of the term "quasi product". This work was supported in part by ARRS Slovenia and the Deutsche Forschungsgemeinschaft (DFG) Project STA850/11-1 within the EUROCORES Program EuroGIGA (project GReGAS) of the European Science Foundation. This paper is based on part of the dissertation of the third author. t Corresponding Author Abstract Keywords: Cartesian product, quasi product, graph bundle, approximate product, partial star product, product relation. Math. Subj. Class.: 05C15, 05C10 1 Introduction Cartesian products of graphs derive their popularity from their simplicity, and their importance from the fact that many classes of graphs, such as hypercubes, Hamming graphs, median graphs, benzenoid graphs, or Cartesian graph bundles, are either Cartesian products or closely related to them [5]. As even slight disturbances of a product, such as the addition or deletion of an edge, can destroy the product structure completely [2], the question arises whether it is possible to restore the original product structure after such a disturbance. In other words, given a graph, the question is, how close it is to a Cartesian product, and whether one can find this product algorithmically. Unfortunately, in general this problem can only be solved by heuristic algorithms, as discussed in detail in [8]. That paper also presents several heuristic algorithms for the solution of this problem. One of the main steps towards such algorithms is the computation of an equivalence relation d|Sv (W)* on the edge-set of a graph. The complexity of the computation of d|Sv (W)* in [8] is O(nA4), where n is the number of vertices, and Д the maximum degree of G. Here we improve the recognition complexity of d|Sv (W)* to O(mA), where m is the number of edges of G, and thereby improve the complexity of the just mentioned heuristic algorithms. A special case is the computation of the relation S* = d|Sv (V(G))*. This relation defines the so-called quasi Cartesian product, see Section 3. Hence, quasi products can be recognized in O(mA) time. As the algorithm can easily be parallelized, it leads to sublinear recognition of quasi Cartesian products. When the given graph G is a Cartesian product from which just one vertex was deleted, things are easier. In that case, the product is uniquely defined and can be reconstructed in polynomial time from G, see [1] and [3]. In other words, if G is given, and if one knows that there is a Cartesian product graph H such that G = H \ x, then H is uniquely defined. Hagauer and Žerovnik showed that the complexity of finding H is O(mn(A2 + m)). The methods of the present paper will lead to a new algorithm of complexity O(mA2 + A4) for the solution of this problem. This is part of the dissertation [13] of the third author, and will be the topic of a subsequent publication. Another class of graphs that is closely related to Cartesian products are Cartesian graph bundles, see Section 3. In [11] it was proved that Cartesian graph bundles over a triangle-free base can be effectively recognized, and in [14] it was shown that this can be done in O(mn2) time. With the methods of this paper, we suppose that one can improve it to O(mA) time. This too will be published separately. 2 Preliminaries We consider finite, connected undirected graphs G = (V, E) without loops and multiple edges. The Cartesian product G1\2G2 of graphs Gi = (V1, Ei) and G2 = (V2,E2) E-mail addresses: marc.hellmuth@bioinf.uni-sb.de (Marc Hellmuth), imrich@unileoben.ac.at (Wilfried Imrich), tomas.kupka@teradata.com (Tomas Kupka) is a graph with vertex set Vi x V2, where the vertices (u1,v1) and (u2, v2) are adjacent if u\u2 G Ei and vi = v2, or if v\v2 G E2 and ui = u2. The Cartesian product is associative, commutative, and has the one vertex graph Ki as a unit [5]. By associativity we can write GiDG2D • • • □ Gk for a product G of graphs Gi, G2,..., Gk and can label the vertices of G by the set of all k-tuples (vi,v2,... ,vk ), where vj G Gì for 1 < i < k. If v is labeled (vi; v2,..., vk), then we call vj its ith coordinate. One says two edges have the same Cartesian color if their endpoints differ in the same coordinate. A graph G is prime if it is non-trivial, and if the identity G = Gi^G2 implies that Gi or G2 is the one-vertex graph Ki. A representation of a graph G as a product Gi^G2^ • • • □Gk of prime graphs is called a prime factorization of G. It is well known that every connected graph G has a prime factor decomposition with respect to the Cartesian product, and that this factorization is unique up to isomorphisms and the order of the factors, see Sabidussi [15]. Furthermore, the prime factor decomposition can be computed in linear time, see [10]. Following the notation in [8], an induced cycle on four vertices is called chordless square. Let the edges e = vu and f = vw span a chordless square vuxw. Then f is the opposite edge of the edge xu. The vertex x is called top vertex (w.r.t. the square spanned by e and f). A top vertex x is unique if |N(x) П N(v)| = 2, where N(u) denotes the (open) 1-neighborhood of vertex u. In other words, a top vertex x is not unique if there are further squares with top vertex x spanned by the edges e or f together with a third distinct edge g. Note that the existence of a unique top vertex x does not imply that e and f span a unique square, as there might be another square vuyw with a possible unique top vertex y. Thus, e and f span a unique square vuxw only if |N(u) П N(w)| = 2. The degree deg(u) := |N(u)| of a vertex u is the number of edges that contain u. The maximum degree of a graph is denoted by Д and a path on n vertices by Pn. We now recall the Breadth-First Search (BFS) ordering of the vertices v0, vi;..., vn-i of a graph: select an arbitrary, but fixed vertex v0 g V (G), called the root, and create a sorted list of vertices. Begin with v0; append all neighbors vi;..., vdeg(V0) of v0 to the list; then append all neighbors of vi that are not already in the list; and continue recursively with v2, v3,... until all vertices of G are processed. 2.1 The Relations ö, a and the Square Property. There are two basic relations S and a, among other relations that are defined on the edge set of a given graph, that play an important role in the field of Cartesian product recognition. In the sequel we shall also use the notation R* for the transitive closure of a relation R, that is, R* is the smallest transitive relation containing R. Definition 2.1. Two edges e, f G E (G) are in the relation SG, if one of the following conditions in G is satisfied: (i) e and f are adjacent and it is not the case that there is a unique square spanned by e and f, and that this square is chordless. (ii) e and f are opposite edges of a chordless square. (iii) e = f. Clearly, this relation is reflexive and symmetric but not necessarily transitive. The transitive closure S*G is an equivalence relation. If adjacent edges e and f are not in relation 6, that is, if Condition (i) of Definition 2.1 is not fulfilled, then they span a unique square, and this unique square spanned by e and f is chordless. We call such a square the unique chordless square (spanned by e and f ). Two edges e and f are in the product relation aG if they have the same Cartesian colors with respect to the prime factorization of G. The product relation aG is a uniquely defined equivalence relation on E(G) that contains all information about the prime factorization1. Furthermore, 6G and 6*G are contained in aG. If there is no risk of confusion we write 6 or a for 6G or aG, respectively. We say an equivalence relation p defined on the edge set of a graph G has the square property if the following three conditions hold: (a) For any two edges e = uv and f = uw that belong to different equivalence classes of p there exists a unique vertex x = u of G that is adjacent to v and w. (b) The square uvxw is chordless. (c) The opposite edges of any chordless square belong to the same equivalence class of p. From the definition of 6 it easily follows that 6 is a refinement of any such p. It also implies that 6*, and thus also a, have the square property. This property is of fundamental importance, both for the Cartesian and the quasi Cartesian product. We note in passing that a is the convex hull of 6*, see [12]. 2.2 The Partial Star Product This section is concerned with the partial star product, which plays a decisive role in the local approach. As it was introduced in [8], we will only define it here, list some of its most basic properties, and refer to [8] for details. Let G = (V, E) be a given graph and Ev the set of all edges incident to some vertex v g V. We define the local relation dv as follows: dv = ((Ev X E) U (E X Ev)) n 6g C 6{Na[v]), where (N2G[v]) denotes the induced closed 2-neigborhood of v in G. In other words, dv is the subset of 6G that contains all pairs (e, f ) g 6g, where at least one of the edges e and f is incident to v. Clearly d*, which is not necessarily a subset of 6, is contained in 6*, see [8]. v Let Sv be a subgraph of G that contains all edges incident to v and all squares spanned by edges e, e' g Ev where e and e' are not in relation d*. Then Sv is called partial star product (PSP for short). To be more precise: Definition 2.2 (Partial Star Product (PSP)). Let Fv C E \ Ev be the set of edges which are opposite edges of (chordless) squares spanned by e, e' g Ev that are in different d* classes, that is, (e, e') G d*. Then the partial star product is the subgraph Sv C G with edge set E' = Ev U Fv and vertex set Ue£E> e, which consists of the end vertices of the edges in E'. We call v the center of Sv, Ev the set of primal edges, Fv the set of non-primal edges, and the vertices adjacent to v primal vertices of Sv. 1For the properties of a that we will cite or use, we refer the reader to [5] or [9]. As shown in [8], a partial star product Sv is always an isometric subgraph or even isomorphic to a Cartesian product graph H, where the factors of H are so-called stars K1n. These stars can directly be determined by the respective d* classes, see [8]. Now we define a local coloring of Sv as the restriction of the relation d* to Sv : d|Sv := d*|s„ = {(e,f ) € d* | e,f G E (Sv )}. In other words, d|Sv is the subset of dv that contains all pairs of edges (e, f ) G d* where both e and f are in Sv and edges obtain the same local color whenever they are in the same equivalence class of d|Sv. As an example consider the PSP Sv in Figure 1(d). The relation d|Sv has three equivalence classes (highlighted by thick, dashed and double-lined edges). Note, S* just contains one equivalence class. Hence, d|Sv = S*Sv. For a given subset W С V we set d|Sv (W) = Uv£wd|Sv . The transitive closure of d|Sv (W) is then called the global coloring with respect to W. As shown in [8], we have the following theorem. Theorem 2.3. Let G = (V, E) be a given graph and d|Sv (V) = UveVd|Sv. Then d| Sv (V )* = SG. For later reference and for the design of the recognition algorithm we list the following three lemmas about relevant properties of the PSP. Lemma 2.4 ([8]). Let G=(V,E) be a given graph and Sv be a PSP of an arbitrary vertex v G V. If e, f G Ev are primal edges that are not in relation d v, then e and f span a unique chordless square with a unique top vertex in G. Conversely, suppose that x is a non-primal vertex of Sv. Then there is a unique chord-less square in Sv that contains x, and that is spanned by edges e, f G Ev with (e, f ) G d*. Lemma 2.5 ([8]). Let G=(V,E) be a given graph and f G Fv be a non-primal edge of a PSP Sv of an arbitrary vertex v G V. Then f is opposite to exactly one primal edge e G Ev in Sv, and (e,f ) G d|S„. Lemma 2.6 ([8]). Let G=(V,E) be a given graph and W С V such that (W} is connected. Then each vertex x G W meets every equivalence class of d|Sv (W)* in UveWSv. 3 Quasi Cartesian Products Given a Cartesian product G = ЛОБ of two connected, prime graphs A and B, one can recover the factors A and Б as follows: the product relation a has two equivalence classes, say E1 and E2, and the connected components of the graph (V(G), E1) are all isomorphic copies of the factor A, or of the factor Б, see Figure 1(a). This property naturally extends to products of more than two prime factors. We already observed that S is finer than any equivalence relation p that satisfies the square property. Hence the equivalence classes of p are unions of S*-classes. This also -*-1- -*-1" (a) The Cartesian product G = P3OC4. (c) A quasi Cartesian product, which is not a graph bundle. (b) A quasi Cartesian product, which is also a graph bundle. (d) The approximate product and PSP Sv, which is neither a quasi product nor a graph bundle. Figure 1: Shown are several quasi Cartesian products, graph bundles and approximate products. holds for a. It is important to keep in mind that a can be trivial, that is, it consists of a single equivalence class even when S* has more than one equivalence class. We call all graphs G with a non-trivial equivalence relation p that is defined on E(G) and satisfies the square property quasi (Cartesian) products. Since S* C p for every such relation p, it follows that S* must have at least two equivalence classes for any quasi product. By Theorem 2.3 we have d|Sv (V(G))* = S*. In other words, quasi products can be defined as graphs where the PSP's of all vertices are non-trivial, that is, none of the PSP's is a star K1n, and in addition, where the union over all d|Sv yields a non-trivial S*. Consider the equivalence classes of the relation S* of the graph G of Figure 1(b). It has two equivalence classes, and locally looks like a Cartesian product, but is actually reminiscent of a Möbius band. Notice that the graph G in Figure 1(b) is prime with respect to Cartesian multiplication, although S* has two equivalence classes: all components of the first class are paths of length 2, and there are two components of the other S*-class, which do not have the same size. Locally this graph looks either like Р3ПР3 or P2ПР3. In fact, the graph in Figure 1(b) is a so-called Cartesian graph bundle [11], where Cartesian graph bundles are defined as follows: Let B and F be graphs. A graph G is a (Carte- sian) graph bundle with fiber F over the base B if there exists a weak homomorphism2 p : G ^ B such that (i) for any u e V(B), the subgraph (induced by) p-1 (u) is isomorphic to F, and (ii) for any e e E(B), the subgraph p-1(e) is isomorphic to K2ÜF. The graph of Figure 1(c) shows that not all quasi Cartesian products are graph bundles. On the other hand, not every graph bundle has to be a quasi product. The standard example is the complete bipartite graph K3 3. It is a graph bundle with base K3 and fiber K2, but has only one S*-class. Note, in [8] we considered "approximate products" which were first introduced in [7,6]. As approximate products are the graphs that have a (small) edit distance to a non-trivial product graph, it is clear that every bundle and quasi product can be considered as an approximate product, while the converse is not true. For example, consider the graph in Figure 1(d). Here, S* has only one equivalence class. However, the relation d|Sv has, in this case, three equivalence classes (highlighted by thick, dashed and double-lined edges). Because of the local product-like structure of quasi Cartesian products we are led to the following conjecture: Conjecture 3.1. Quasi Cartesian products can be reconstructed in essentially the same time from vertex-deleted subgraphs as Cartesian products. 4 Recognition Algorithms 4.1 Computing the Local and Global Coloring For a given graph G, let W С V (G) be an arbitrary subset of the vertex set of G such that the induced subgraph (W} is connected. Our approach for the computation is based on the recognition of all PSP's Sv with v e W, and subsequent merging of their local colorings. The subroutine computing local colorings calls the vertices in BFS-order with respect to an arbitrarily chosen root v0 e W. Let us now briefly introduce several additional notions used in the PSP recognition algorithm. At the start of every iteration we assign pairwise different temporary local colors to the primal edges of every PSP. These colors are then merged in subroutine processes to compute local colors associated with every PSP. Analogously, we use temporary global colors that are initially assigned to every edge incident with the root v0. For any vertex v of distance two from a PSP center c we store attributes called first and second primal neighbor, that is, references to adjacent primal vertices from which v was "visited" (in pseudo-code attributes are accessed by v .FirstPrimalNeighbor and v.SecondPrimalNeighbor). When v is found to have at least two primal neighbors we add v to Tc, which is a stack of candidates for non-primal vertices of Sc. Finally, we use incidence and absence lists to store recognized squares spanned by primal edges. Whenever we recognize that two primal edges span a square we put them into the incidence list. If we find out that a pair of primal edges cannot span a unique chordless square with unique top vertex, then we move it into the absence list. Note that the above structures are local and are always associated with a certain PSP recognition subroutine (Algorithm 4.1). Finally, we will "map" local colors to temporary global colors via temporary vectors which helps us to merge local with global colors. 2A weak homomorphism maps edges into edges or single vertices. Algorithm 4.1 computes a local coloring for a given PSP and merges it with the global coloring d|Sv (W )* where W С V (G) is the set of treated centers. Algorithm 4.2 summarizes the main control structure of the local approach. Algorithm 4.1 (PSP recognition) Input: Connected graph G = (V, E), PSP center c e V, global coloring ö | S(W)*, where W с V is the set of treated centers and where the subgraph induced by W u c is connected. Output: New temporary global coloring ö | S ( W u c)*. 1. Initialization. 2. FOR every neighbor u of c DO: (a) FOR every neighbor w of u (except c) DO: i. IF w is primal w.r.t. c THEN add pair of primal edges (cu,cw) to absence list. ii. ELSEIF w was not visited THEN set w.FirstPrimalNeighbor = u. iii. ELSE (w is not primal and was already visited) DO: A. IF only one primal neighbor v (v = u) of w was recognized so far, then DO: • Set w.S econdPrimalN eighbor = u. • IF (cu, cv) is not in incidence list, then add w to the stack Tc and add the pair (cu, cv) to incidence list. • ELSE (cu and cv span more squares) add pair (cu, cv) to absence list. B. ELSE: • Add all pairs formed by primal edges evi, cv2, cu to absence list, where vi, v2 are first and second primal neighbors of w. 3. Assign pairwise different temporary local colors to primal edges. 4. FOR any pair (cu, cv) of primal edges cu and cv DO: (a) IF (cu, cv) is contained in absence list THEN merge temporary local colors of cu and cv. (b) IF (cu, cv) is not contained in incidence list THEN merge temporary local colors of cu and cv. (Resulting merged temporary local colors determine local colors of primal edges in Sc. We will reference them in the following steps.). 5. FOR any primal edge cu DO: (a) IF cu was already assigned some temporary global color di THEN i. IF local color b of cu was already mapped to some temporary global color d2, where d2 = di, THEN merge di and d2. ii. ELSE map local color b to di. 6. FOR any vertex v from the stack Tc DO: (a) Check local colors of primal edges cwi and cw2 (where wi,w2 are fìrst and second primal neighbor of v, respectively). (b) IF they differ in local colors THEN i. IF there was defined temporary global color di for vwi THEN A. IF local color b of cw2 was already mapped to some temporary global color d2, where d2 = di THEN merge di and d2. B. ELSE map local color b to d1. ii. IF there was already defìned temporary global color d1 for vw2 THEN: A. IF local color b of cw1 was already mapped to some temporary global color d2, where d2 = d1 THEN merge d1 and d2. B. ELSE map local color b to d1. 7. Take every edge e of the PSP Sc that was not colored by any temporary global color up to now and assign it d, where d is the temporary global color to which the local color of e or the local color of its opposite primal edge e' was mapped. (If there is a local color b that was not mapped to any temporary global color, then we create anew temporary global color and assign it to all edges of color b.) Algorithm 4.2 (Computation of d|Sv (W)*) Input: A connected graph G, W с V (G) s.t. the induced subgraph (W} is connected, and an arbitrary vertex v0 g W. Output: Relation ö|Sv (W)*. 1. Initialization. 2. Set sequence Q of vertices v0,v1,... ,vn that form W in BFS-order with respect to v0. 3. Set W' := 0. 4. Assign pairwise different temporary global colors to edges incident to v0. 5. FOR any vertex vi from sequence Q DO: (a) Use Algorithm 4.1 to compute ö | sv (W' u vi)*. (b) Add vi to W '. In order to show that Algorithm 4.1 correctly recognizes the local coloring, we define the (temporary) relations ac and ßc for a chosen vertex c: Two primal edges of Sc are • in relation ac if they are contained in the incidence list and • in relation ßc if they are contained in the absence list after Algorithm 4.1 is executed for c. Note, we denote by ac the complement of ac, which contains all pairs of primal edges of PSP Sc that are not listed in the incidence list. Lemma 4.1. Let e and f be two primal edges of the PSP Sc. If e and f span a square with some non-primal vertex w as unique top-vertex, then (e, f ) G ac. Proof. Let e = cu\ and f = cu2 be primal edges in Sc that span a square cu\wu2 with unique top-vertex w, where w is non-primal. Note, since w is the unique top vertex, the vertices ui and u2 are its only primal neighbors. W.l.o.g. assume that for vertex w no first primal neighbor was assigned and let first u1 and then u2 be visited. In Step 2a vertex w is recognized and the first primal neighbor u1 is determined in Step 2(a)ii. Take the next vertex u2. Since w is not primal and was already visited, we are in Step 2(a)iii. Since only one primal neighbor of w was recognized so far, we go to Step 2(a)iiiA. If (cu1, cu2) is not already contained in the incidence list, it will be added now and thus, (cu1, cu2) g ac. □ Corollary 4.2. Let e and f be two adjacent distinct primal edges of the PSP Sc. If (e, f ) e ac, then e and f do not span a square or span a square with non-unique or primal top vertex. In particular, ac contains all pairs (e, f ) that do not span any square. Proof. The first statement is just the contrapositive of the statement in Lemma 4.1. For the second statement observe that if e = cx and f = cy are two distinct primal edges of Sc that do not span a square, then the vertices x and y do not have a common non-primal neighbor w. It is now easy to verify that in none of the substeps of Step 2 the pair (e, f ) is added to the incidence list, and thus, (e, f ) e ac. □ Lemma 4.3. Let e and f be two primal edges of the PSP Sc that are in relation ßc. Then e and f do not span a unique chordless square with unique top vertex. Proof. Let e = cui and f = cu2 be primal edges of Sc. Then pair (e, f ) is added to the absence list in: a) Step 2(a)i, when u1 and u2 are adjacent. Then no square spanned by e and f can be chordless. b) Step 2(a)iiiA (else-condition), when (e, f ) is already listed in the incidence list and another square spanned by e and f is recognized. Thus, e and f do not span a unique square. c) Step 2(a)iiiB, when e and f span a square with top vertex w that has more than two primal neighbors and at least one of the primal vertices u1 and u2 are recognized as first or second primal neighbor of w. Thus e and f span a square with non-unique top vertex. □ Lemma 4.4. Relation ß* contains all pairs of primal edges (e, f ) of Sc that satisfy at least one of the following conditions: a) e and f span a square with a chord. b) e and f span a square with non-unique top vertex. c) e and f span more than one square. Proof. Let e = cu1 and f = cu2 be primal edges of the PSP Sc. a) If e and f span a square with a chord, then u1 and u2 are adjacent or the top vertex w of the spanned square is primal and thus, there is a primal edge g = cw. In the first case, we can conclude analogously as in the proof of Lemma 4.3 that (e, f ) e ßc. In the second case, we analogously obtain (e, g), (f, g) e ßc and therefore, (e, f ) e ß*. b) Let e and f span a square with non-unique top vertex w. If at least one of the primal vertices u1, u2 is a first or second neighbor of w then e and f are listed in the absence list, as shown in the proof of Lemma 4.3. If u1 and u2 are neither first nor second primal neighbors of w, then both edges e and f will be added to the absence list in Step 2(a)iiiB, together with the primal edge g = cu3, where u3 is the first recognized primal neighbor of w. In other words, (e, g), (f, g) e ßc and hence, (e, f ) e ß*. c) Let e and f span two squares with top vertices w and w', respectively and assume w.l.o.g. that first vertex w is visited and then w'. If both vertices u1 and u2 are recognized as first and second primal neighbors of w and w', then (cu1, cu2 ) is added to the incidence list when visiting w in Step 2(a)iiiA. However, when we visit w', then we insert (cui, cu2) to the absence list in Step 2(a)iiiA, because this pair is already included in the incidence list. Thus, (e, f ) G ßc. If at least one of the vertices w, w' does not have ui and u2 as first or second primal neighbor, then e and f must span a square with non-unique top vertex. Item b) implies that (e, f ) G ß*. □ Lemma 4.5. Let f be a non-primal edge and e1, e2 be two distinct primal edges of Sc. Let (ei,f ), (e2,f ) G dc. Then (ei, e2) G ß*c. Proof. Since the edge f is non-primal, f is not incident with the center c. Recall, by the definition of dc, two distinct edges can be in relation dc only if they have a common vertex or are opposite edges in a square. To prove our lemma we need to investigate the three following cases, which are also illustrated in Figure 2: a) Suppose both edges ei and e2 are incident with f. Then ei and e2 span a triangle and consequently (e1; e2) will be added to the absence list in Step 2(a)i. b) Let ei and e2 be opposite to f in some squares. There are two possible cases (see Figure 2 b)). In the first case ei and e2 span a square with non-unique top vertex. By Lemma 4.4, (e1; e2) G ß**. In the second case e1 and e2 span triangles with other primal edges e3 and e4. As in Case a) of this proof, we have (e1, e3) G ßc, (e3, e4) G ßc, (e4, e2) G ßc and consequently, (e1; e2) G ß* c) Suppose only e1 has a common vertex with f and e2 is opposite to f in a square. Again we need to consider two cases (see Figure 2 c)). Since e1 and f are adjacent and (e1,f ) G dc, we can conclude that either no square is spanned by e1 and f, or that the square spanned by e1 and f is not chordless or not unique. It is easy to see that in the first case the edges e1 and e2 are contained in a common triangle and thus will be added to the absence list in Step 2(a)i. In the second case e1,e2 span a square which has a chord or has a non-unique top vertex. In both cases Lemma 4.4 implies that e1 and e2 are in relation ß*. □ Lemma 4.6. Let e and f be distinct primal edges of the PSP Sc. Then (e, f ) G (ac U ßc)* if and only if (e, f ) G d*. Proof. Assume first that (e, f ) G ac U ßc. By Corollary 4.2, if (e, f ) G ac, then e and f do not span a common square, or span a square with non-unique or primal top vertex. In the first case, e and f are in relation SG and consequently also in relation dc. On the other hand, if e and f span square with non-unique top vertex then, by Lemma 2.4, e and f are in relation dc as well. Finally, if e and f span a square with primal top vertex w, then this square has a chord cw and (e, f ) G dc. If (e, f ) G ßc, then Lemma 4.3 implies that e and f do not span a unique chordless square with unique top vertex. Again, by Lemma 2.4, we infer that (e, f ) G d*. Hence, ac U ßc С dc, and consequently, (ac U ßc)* C d*. a) b) c) ) 1 f Figure 2: The three possible cases a), b), and c) that are investigated in the proof of Lemma 4.5. Now, let (e, f) G d*. Then there is a sequence U = (e = e1; e2,..., ek = f), k > 2, with (ei; ei+1) G dc for i G {1, 2,..., k - 1}. By definition of dc, two primal edges are in relation dc if and only if they do not span a unique and chordless square. Corollary 4.2 and Lemma 4.4 imply that all these pairs are contained in (ac U ßc)*. Hence, any two consecutive primal edges ež and ei+1 contained in the sequence U are in relation (ac U ßc ) *. Assume that there is an edge ež g U that is not incident to the center c and thus, non-primal. By the definition of dc, and since (ei-1,ei), (ei;ei+1) G dc, we can conclude that the edges ei-1 and ei+1 must be primal in Sc. Lemma 4.5 implies that ei-1 and ei+1 must be in relation ß*. Thus, if we remove the edge ež from U, we still can claim that all consecutive primal edges in U \ {ež} are in relation (ac U ßc)*. By removing all non-primal edges from U we therefore obtain a sequence U' = e = e1, e'2,..., ej = f of primal edges. By analogous arguments as before, all pairs (ei, ei+1) of U' must be contained in (ac U ßc)*. By transitivity, e and f are also in (ac U ßc)*. □ Corollary 4.7. Let e and f be primal edges of the PSP Sc. Then (e, f ) G (ac U ßc)* if and only if e and f have the same local color in Sc. Proof. This is an immediate consequence of Lemma 4.6, the local color assignment, and the merging procedure (Step 3 and 4) in Algorithm 4.1. □ Lemma 4.8. Let d|Sv (W)* be a global coloring associated with a set of treated centers W and assume that the induced subgraph (W} is connected. Let c be a vertex that is not contained in W but adjacent to a vertex in W. Then Algorithm 4.1 computes the global coloring d|Sv (W U c)* by taking W and c as input. Proof. Let W С V (G) be a set of PSP centers and let c G V (G) be a given center of PSP Sc where c G W and (W U c} is connected. In Step 2 of Algorithm 4.1 we compute the absence and incidence lists. In Step 3, we assign pairwise different temporary local colors to any primal edge adjacent to c. Two temporary local colors b1 and b2 are then merged in Step 4 if and only if there exists some pair of primal edges (eb e2) G (ac U ßc) where c c ei is colored with bi and e2 with b2. Therefore, merged temporary local colors reflect equivalence classes of (ac U ßc)* containing the primal edges incident to c. By Corollary 4.7, (ac U ßc)* classes indeed determine the local colors of primal edges in Sc. Note, if one knows the colors of primal edges incident to c, then it is very easy to determine the set of non-primal edges of Sc, as any two primal edges of different equivalence classes span a unique and chordless square. In Step 6, we investigate each vertex v from stack Tc and check the local colors of primal edges cwi and cw2, where wi and w2 are the first and second recognized primal neighbors of v, respectively. If cwi and cw2 differ in their local colors, then vwi and vw2 are non-primal edges of Sc, as follows from the PSP construction. Recall that the stack contains all vertices that are at distance two from center c and which are adjacent to at least two primal vertices. In other words, the stack contains all non-primal top vertices of all squares spanned by primal edges. Consequently, we claim that all non-primal edges of the PSP Sc are treated in Step 6. Note that non-primal edges have the same local color as their opposite primal edge, which is unique by Lemma 2.5. As we already argued, after Step 4 is performed we know, or can at least easily determine all edges of Sc and their local colors. Recall that local colors define the local coloring d|Sc. Suppose, temporary global colors that correspond to the global coloring d|Sv (W)* are assigned. Our goal is to modify and identify temporary global colors such that they will correspond to the global coloring d|Sv (W U c)*. Let Bi, B2,..., Bk be the classes of d|Sc (local classes) and Di,D2,... , D; be the classes of d|Sv (W)* (global classes). When a local class Bi and a global class D j have a nonempty intersection, then we can infer that all their edges must be contained in a common class of d|Sv (W U c)*. Note, by means of Lemma 2.6, we can conclude that for each local class Bj there is a global class D j such that Bi n D j = 0, see also [8]. In that case we need to guarantee that edges of Bj and D j will be colored by the same temporary global color. Note, in the beginning of the iteration two edges have the same temporary global color if and only if they lie in a common global class. In Step 5 and Step 6, we investigate all primal and non-primal edges of Sc. When we treat first edge e that is colored by some local color bi, that is e g Bi, and has already been assigned some temporary global color d j, and therefore e g D j, then we map bi to d j. Thus, we keep the information that e g Bi n D j. In Step 7, we then assign temporary global color d j to any edge of Sc that is colored by the local color bi. If the local color bi is already mapped to some temporary global color d j, and if we find another edge of Sc that is colored by bi and simultaneously has been assigned some different temporary global color dj,, then we merge dj and dj, in Step 5(a)i. Obviously this is correct, since Bi n Dj = 0 and Bi n Dj, = 0, and hence Dj, Dj, and Bi must be contained in a common equivalence class of d|Sv (W U c)*. Recall, for each local class Bi there is a global class Dj such that Bi n D j = 0. This means that every local color is mapped to some global color, and consequently there is no need to create a new temporary global color in Step 7. Therefore, whenever local and global classes share an edge, then all their edges will have the same temporary global color at the end of Step 7. On the other hand, when edges of two different global classes are colored by the same temporary global color, then both global classes must be contained in a common class of d|Sv (W U c)*. Hence, after the performance of Step 7, the merged temporary global colors determine the equivalence classes of d | Sv ( W U c) *. □ Lemma 4.9. Let G be a connected graph, W С V (G) s.t. (W} is connected, and v0 an arbitrary vertex of G. Then Algorithm 4.2 computes the global coloring d|Sv (W)* by taking G, W, and v0 as input. Proof. In Step 2 we define the BFS-order in which the vertices will be processed and store this sequence in Q. In Step 4 we assign pairwise different temporary global colors to all edges that are incident with v0. In Step 5 we iterate over all vertices of the given induced connected subgraph (W} of G. For every vertex we execute Algorithm 4.1. Lemma 4.8 implies that in the first iteration we correctly compute the local colors for Sv0, and consequently also d|Sv ({v0})*. Obviously, whenever we merge two temporary local colors of two primal edges in the first iteration, then we also merge their temporary global colors. Consequently, the resulting temporary global colors correspond to the global coloring d|Sv ({v0})* after the first iteration. Lemma 4.8 implies that after all iterations are performed, that is, all vertices in Q are processed, the resulting temporary global colors correspond to d|Sv (W )* for the given input set W С V (G). □ For the global coloring, Theorem 2.3 implies that d|Sv (V(G))* = S**. This leads immediately to the following theorem. Theorem 4.10. Let G be a connected graph and v0 an arbitrary vertex of G. Then Algorithm 4.2 computes the global coloring S* by taking G, V(G), and v0 as input. 4.2 Time Complexity We begin with the complexity of merging colors. We have global and local colors, and will define local and global color graphs. Both graphs are acyclic temporary structures. Their vertex sets are the sets of temporary colors in the initial state. In this state the color graphs have no edges. Every component is a single vertex and corresponds to an initial temporary color. Recall that we color edges of graphs, for example the edges of G or Sv. The color of an edge is indicated by a pointer to a vertex of the color graph. These pointers are not changed, but the colors will correspond to the components of the color graph. When two colors are merged, then this will be reflected by adding an edge between their respective components. The color graph is represented by an adjacency list as described in [5, Chapter 17.2] or [9, pp. 34 -37]. Thus, working with the color graph needs O(k) space when k colors are used. Furthermore, for every vertex of the color graph we keep an index of the connected component in which the vertex is contained. We also store the actual size of every component, that is, the number of vertices of this component. Suppose we wish to merge temporary colors of edges e and f that are identified with vertices a, respectively b, in the color graph. We first check whether a and b are contained in the same connected component by comparing component indices. If the component indices are the same, then e and f already have the same color, and no action is necessary. Otherwise we insert an edge between a and b in the color graph. As this merges the components of a and b we have to update component indices and the size. The size is updated in constant time. For the component index we use the index of the larger component. Thus, no index change is necessary for the larger component, but we have to assign the new index to all vertices of the smaller component. Notice that the color graph remains acyclic, as we only add edges between different components. Lemma 4.11. Let G0 = (V, E) be a graph with V = {v1,..., vk} and E = 0. The components of Go consist of single vertices. We assign component index j to every component {vj}. For i G {1,... ,k — 1} let Gi+1 denote the graph that results from G i by adding an edge between two distinct connected components, say C and C'. If |C| < |Cwe use the component index of C' for the new component and assign it to every vertex of C. Then every Gi is acyclic, and the total cost ofmerging colors is O(k log2 k). Proof. Acyclicity is true by construction. A vertex is assigned a new component index when its component is merged with a larger one. Thus, the size of the component at least doubles at every such step. Because the maximum size of a component is bounded by k, there can be at most log2 k reassignments of the component index for every vertex. As there are k vertices, this means that the total cost of merging colors is O(k log2 k). □ The color graph is used to identify temporary local, resp., global colors. Based on this, we now define the local and global color graph. Assigned labels of the vertices of the global color graph are stored in the edge list, where any edge is identified with at most one such label. Notice that the original graph is represented by an extended adjacency list, where for any vertex and its neighbor a reference to the edge (in the edge list) that connects them is stored. This reference allows to access a global temporary color from adjacency list in constant time. In every iteration of Algorithm 4.2, we recognize the PSP for one vertex by calling Algorithm 4.1. In the following paragraph we introduce several temporary attributes and matrices that are used in the algorithm. Suppose we execute an iteration that recognizes some PSP Sc. To indicate whether a vertex was treated in this iteration we introduce the attribute visited, that is, when vertex v is visited in this iteration we set v.visited = c. Any value different from c means that vertex v was not yet treated in this iteration. Analogously, we introduce the attribute primal to indicate that a vertex is adjacent to the current center c. The attribute tempLabel maps primal vertices to the indices of rows and columns of the matrices incidenceList and absenceList. For any vertex v that is at distance two from the center c we store its first and second primal neighbor w1 and w2 in the attributes FirstPrimalNeighbor and SecondPrimalNeighbor. Furthermore, we need to keep the position of vw1 and vw2 in the edge list to get their temporary global colors. For this purpose, we use attributes firstEdge and secondEdge. Attribute mapLocalColor helps us to map temporary local colors to the vertices of the global color graph. Any vertex that is at distance two from the center and has a least two primal neighbors is a candidate for a non-primal vertex. We insert them to the stack. The temporary structures help to access the required information in constant time: • v.visited = c vertex v has been already visited in the current iteration. • v.primal = c vertex v is adjacent to center c. • incidenceList[v.tempLabel,u.tempLabel] =0 pair of primal edges (cv, cu) is missing in the incidence list. • absenceList[v.tempLabel,u.tempLabel] = 1 pair of primal edges (cv, cu) was inserted to the absence list. • v.f irstPrimalNeighbor = u u is the first recognized primal neighbor of the non-primal vertex v. • v.firstEdge = e edge e joins the non-primal vertex v with its first recognized primal neighbor (it is used to get the temporary global color from the edge list). • b.mapLocalColor = d local color b is mapped to temporary global color d (i.e. there exists an edge that is colored by both colors). Note that the temporary matrices incidenceList and absenceList have dimension deg(c) X deg(c) and that all their entries are set to zero in the beginning of every iteration. Theorem 4.12. For a given connected graph G = (V, E) with maximum degree A and W С V, Algorithm 4.2 runs in O(|E|A) time and O(|E| + A2) space. Proof. Let G be a given graph with m edges and n vertices. In Step 1 of Algorithm 4.2 we initialize all temporary attributes and matrices. This consumes O(m+n) = O(m) time and space, since G is connected, and hence, m > n — 1. Moreover, we set all temporary colors of edges in the edge list to zero, which does not increase the time and space complexity of the initial step. Recall that we use an extended adjacency list, where every vertex and its neighbors keep the reference to the edge in the edge list that connects them. To create an extended adjacency list we iterate over all edges in the edge list, and for every edge uv = e G E(G) we set a new entry for the neighbor v for u and, simultaneously, we add a reference v.edge = e. The same is done for vertex v. It can be done in O(m) time and space. In Step 2 of Algorithm 4.2, we build a sequence of vertices in BFS-order starting with v0, which is done in O(m + n) time in general. Since G is connected, the BFS-ordering can be computed in O(m) time. Step 3 takes constant time. In Step 4 we initialize the global color graph that has deg(v0) vertices (bounded by A in general). As we already showed, all operations on the global color graph take O(A log2 A) time and O(A) space. We proceed to traverse all neighbors u1,u2,..., udeg (v0) of the root v0 g V (G) (via the adjacency list) and assign them unique labels 1, 2,..., deg(v0) in edge list, that is, every edge v0ui gets the label i. In this way, we initialize pairwise different temporary global colors of edges incident with v0 , that is, to vertices of the global color graph. Using the extended adjacency list, we set the label to an edge in the edge list in constant time. In Step 5 we run Algorithm 4.1 for any vertex from the defined BFS-sequence. In the remainder of this proof, we will focus on the complexity of Algorithm 4.1. Suppose we perform Algorithm 4.1 for vertex c to recognize the PSP Sc. The recognition process is based on temporary structures. We do not need to reset any of these structures, for any execution of Algorithm 4.1 for a new center c, except absenceList and incidenceList. This is done in Step 1. Further, we set here the attribute tempLabel for every primal vertex v, such that every vertex has assigned a unique number from {1,2, ... , deg(c)}. Finally, we traverse all neighbors of the center c and for each of them we set primal to c. Hence, the initial step of Algorithm 4.1 is done in O(deg(c)2) time. Step 2a is performed for every neighbor of every primal vertex. The number of all such neighbors is at most deg(c)A. For every treated vertex, we set attribute visited to c. This allows us to verify in constant time that a vertex was already visited in the recognition subroutine Algorithm 4.1. If the condition in Step 2(a)i is satisfied, then we add primal edges cu and cw to the absence list. By the previous arguments, this can be done in constant time by usage of tempLabel and absenceList. If the condition in Step 2(a)ii is satisfied, we set vertex u as first primal neighbor of vertex w. For this purpose, we use the attribute firstPrimalNeighbor. We also set w.firstEdge = e, where e is a reference to the edge in the edge list that connects u and w. This reference is obtained from the extended adjacency list in constant time. Recall, the edge list is used to store the labels of vertices of the global color graph for the edges of a given graph, that is, the assignment of temporary global colors to the edges. Using w.firstEdge, we are able to directly access the temporary global color of edge uw in constant time. Step 2(a)iii is performed when we try to visit a vertex w from some vertex u where w has been already visited before from some vertex v. If v is the only recognized primal neighbor of w, then we perform analogous operations as in the previous step. Moreover. if (cu, cv) is not contained in the incidence list, then we set u as second primal neighbor of w, add (cu, cv) to the incidence list and add w to the stack. Otherwise we add (cu, cv) to the absence list. The number of operations in this step is constant. If w has more recognized primal neighbors we process case B. Here we just add all pairs formed by cv1, cv2, cu to absence list. Again, the number of operations is constant by usage of tempLabel and matrices incidenceList and absenceList. In Step 3 we assign pairwise different temporary local colors to the primal edges. Assume the neighbors of the center c are labeled by 1,2,..., deg (c), then we set value u.tempLabel to cu. In Step 4a we iterate over all entries of the absenceList. For all pairs of edges that are in the absence list we check whether they still have different temporary local colors and if so, we merge their temporary local colors by adding a respective edge in the local color graph. Analogously we treat all pairs of edges contained in the incidenceList in Step 4b. Here we merge temporary local colors of primal edges cu and cv when the pair (cu, cv) is missing. To treat all entries of the absenceList and incidenceList we need to perform deg(c)2 iterations. Recall, the temporary local color of the primal edge cu is equal to the index of the connected component in the local color graph, in which vertex u.tempLabel is contained. Thus, the temporary local color of this primal edge can be accessed in constant time. As we already showed, the number of all operations on the local color graph is bounded by O(deg(c) log2 deg(c)). Hence, the overall time complexity of both Steps 3 and4 is O(deg(c)2). In Step 5 we map temporary local colors of primal edges to temporary global colors. For this purpose, we use the attribute mapLocalColor. The temporary global color of every edge can be accessed by the extended adjacency list, the edge list and the global color graph in constant time. Since we need to iterate over all primal vertices, we can conclude that Step 5 takes O(deg(c)) time. In Step 6 we perform analogous operations for any vertex from Stack Tc as in Step 5. In the worst case, we add all vertices that are at distance two from the center to the stack. Hence, the size of the stack is bounded by O(deg(c)A). Recall that the first and second primal neighbor w1 and w2 of every vertex v from the stack can be directly accessed by the attributes firstPrimalNeighbor and secondPrimalNeighbor. On the other hand, the temporary global colors of non-primal edges vw1 and vw2 can be accessed directly by the attributes firstEdge and secondEdge. Thus, all needful information can be accessed in constant time. Consequently, the time complexity of this step is bounded by O(deg(c)A). In the last step, Step 7, we iterate over all edges of the recognized PSP. Note, the list of all primal edges can be obtained from the extended adjacency list. To get all non-primal edges we iterate over all vertices from the stack and use the attributes firstEdge and secondEdge, which takes O(deg(c)A) time. The remaining operations can be done in constant time. To summarize, Algorithm 4.1 runs in O(deg(c)A) time. Consequently, Step 5 of Algorithm 4.2 runs in OEceW deg(c)A) = O(mA) time, which defines also the total time complexity of Algorithm 4.2. The most space consuming structures are the edge list and the extended adjacency list (O(m) space) and the temporary matrices absenceList and incidenceList (O(A2) space). Hence, the overall space complexity is O(m + A2). □ Since quasi Cartesian products are defined as graphs with non-trivial S*, Theorem 4.10 and 4.12 imply the following corollary. Corollary 4.13. For a given connected graph G = (V, E) with bounded maximum degree Algorithm 4.2 (with slight modifications) determines whether G is a quasi Cartesian product in O(IE|) time and O(|E|) space. 4.3 Parallel Processing The local approach allows the parallel computation of S* (G) on multiple processors. Consider a graph G with vertex set V(G). Suppose we are given a decomposition of V(G) = Wi U W2 U — UWfc into k parts such, that | Wi | « | W21 « • • • « | Wk |, where the subgraphs induced by W1,W2,..., Wk are connected, and the number of edges whose endpoints lie in different partitions is small (we call such a decomposition good). Algorithm 4.3 (Parallel recognition of 8*) Input: A graph G, and a good decomposition V(G) = W1 u W2 u • • • u Wk. Output: Relation SG. 1. For every partition Wi concurrently compute global coloring d|sv (Wi) (i e {1, 2,..., k}): V (a) Take all vertices of Wi and order them in BFS to get sequence Qi. (b) SetW' := 0. (c) Assign pairwise different temporary global colors to edges incident to first vertex in Qi. (d) For any vertex v from sequence Qi do: i. Use Algorithm 4.1 to compute ö|Sv (W' u v)*. ii. Move all edges that were treated in previous step and have at least one endpoint not in partition Wi to stack Ti. iii. Add v to W'. 2. Run concurrently for every partition Wi to merge all global colorings (i e {1, 2,..., k}): (a) For each edge from stack Ti, take all its assigned global colors and merge them. d Sv (Wi) Wi W4 W2 W3 Figure 3: Example - Parallel recognition of 6*. Then algorithm 4.2 can be used to compute the colorings d|Sv(Wi)*,d|Sv(W2)*,..., d|Sv(Wk)*, where every instance of the algorithm can run in parallel. The resulting global colorings are used to compute d|Sv(V (G))* = (d|Sv(Wi)* U d|Sv(W2)* U- • •Ud|Sv(Wfc )*)*. The sketch of the parallelization is summarized in Algorithm 4.3. Figure 3 shows an example of decomposed vertex set of a given graph G. The computation of global colorings associated with the individual sets of the partition can be done then in parallel. The edges that are colored by global color when the partition is treated are highlighted by bold black color. Thus we can observe that many edges will be colored by more then one color. Notice that we do not treat the task of finding a good partition. With the methods of [4] this is possible with high probability in O(log n) time, where n is the number of vertices. G References [1] W. Dörfler, Some results on the reconstruction of graphs, in: Infinite and finite sets (Colloq., Keszthely, 1973; dedicated to P. Erdos on his 60th birthday), Vol. I, North-Holland, Amsterdam, pp. 361-363. Colloq. Math. Soc. Janos Bolyai, Vol. 10, 1975. [2] J. Feigenbaum, Product graphs: some algorithmic and combinatorial results, Technical Report STAN-CS-86-1121, Stanford University, Computer Science, 1986, phD Thesis. [3] J. Hagauer and J. Žerovnik, An algorithm for the weak reconstruction of Cartesian-product graphs, J. Combin. Inform. System Sci. 24 (1999), 87-103. [4] S. Halperin and U. Žwick, Optimal randomized EREW PRAM algorithms for finding spanning forests, J. Algorithms 39 (2001), 1-46, doi:10.1006/jagm.2000.1146. [5] R. Hammack, W. Imrich and S. KlavZar, Handbook of Product Graphs, Discrete Mathematics and its Applications, CRC Press, 2nd edition, 2011. [6] M. Hellmuth, A local prime factor decomposition algorithm, Discrete Math. 311 (2011), 944965. [7] M. Hellmuth, W. Imrich, W. Klockl and P. F. Stadler, Approximate graph products, European J. Combin. 30 (2009), 1119 - 1133. [8] M. Hellmuth, W. Imrich and T. Kupka, Partial star products: A local covering approach for the recognition of approximate Cartesian product graphs, Math. Comput. Sci 7 (2013), 255-273. [9] W. Imrich and S. Klavžar, Product graphs, Wiley-Interscience Series in Discrete Mathematics and Optimization, Wiley-Interscience, New York, 2000. [10] W. Imrich and I. Peterin, Recognizing Cartesian products in linear time, Discrete Math. 307 (2007), 472-483. [11] W. Imrich, T. Pisanski and J. Žerovnik, Recognizing Cartesian graph bundles, Discrete Math. 167-168 (1997), 393-403. [12] W. Imrich and J. Žerovnik, Factoring Cartesian-product graphs, J. Graph Theory 18 (1994), 557-567, doi:10.1002/jgt.3190180604. [13] T. Kupka, A local approach for embedding graphs into Cartesian products, Ph.D. thesis, VSB-Technical University of Ostrava, 2013. [14] T. Pisanski, B. Žmazek and J. Žerovnik, An algorithm for k-convex closure and an application, Int. J. Comput. Math 78 (2001), 1-11, doi:10.1080/00207160108805092. [15] G. Sabidussi, Graph multiplication, Math. Z. 72 (1960), 446-457. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 243-259 Rational sums of hermitian squares of free noncommutative polynomials Kristijan Cafuta * University of Ljubljana, Faculty of Electrical Engineering, Laboratory of Applied Mathematics, Tržaška cesta 25, 1000 Ljubljana, Slovenia Igor Klep f The University ofAuckland, Department ofMathematics, Private Bag 92019, Auckland 1142, New Zealand Janez Povh * Faculty of Information studies in Novo mesto, Novi trg 5, 8000 Novo mesto, Slovenia Received 25 July 2013, accepted 13 March 2014, published online 11 January 2015 In this paper we consider polynomials in noncommuting variables that admit sum of hermitian squares and commutators decompositions. We recall algorithms for finding decompositions of this type that are based on semidefinite programming. The main part of the article investigates how to find such decomposition with rational coefficients if the original polynomial has rational coefficients. We show that the numerical evidence, obtained by the Gram matrix method and semidefinite programming, which is usually an almost feasible point, can be frequently tweaked to obtain an exact certificate using rational numbers. In the presence of Slater points, the Peyrl-Parrilo rounding and projecting method applies. On the other hand, in the absence of strict feasibility, a variant of the facial reduction is proposed to reduce the size of the semidefinite program and to enforce the existence of Slater points. All these methods are implemented in our open source computer algebra package NCSOStools. Throughout the paper many worked out examples are presented to illustrate our results. * Partially supported by the Slovenian Research Agency grants P1-0222 and L1-6722. t Supported by the Marsden Fund Council of the Royal Society of New Zealand. Partially supported by the Slovenian Research Agency grants P1-0222, L1-4292 and L1-6722. Part of this research was done while the author was on leave from the University of Maribor. * The author wishes to thank to Slovenian research agency for support via program P1-0383 and project L74119 and to Creative Core FISNM-3330-13-500033 Simulations project funded by the European Union. Abstract Keywords: Sum of squares, semidefinite programming, noncommutative polynomial, Matlab toolbox, commutator, cyclic equivalence, free positivity, real algebraic geometry, Motzkin polynomial, Bessis-Moussa-Villani (BMV) conjecture, NCSOStools. Math. Subj. Class.: Primary 13J30, 90C22; Secondary 08B20, 11E25, 90C90 1 Introduction In this paper we consider free noncommutative (nc) polynomials that are sums of hermitian squares (and commutators). We focus on the following important question: how to obtain a rational certificate (i.e., a symbolic proof) for such a decomposition when the given nc polynomial has rational coefficients and we have numerical (approximate) evidence of a sum of hermitian squares (and commutators) decomposition obtained by mathematical optimization methods (e.g. by using open-source software package NCSOStools)? 1.1 Notation Nc polynomials with real coefficients, denoted by R(X}, are (real) linear combinations of words in letters X1,..., Xn, including the empty word 1. We shortly denote by X the n-tuple of letters (Xi,... ,Xn ). These nc polynomials form a free algebra, which we equip with the involution * that fixes R and letters point-wise and thus reverses words, e.g. (X1X2X3 - X32Xi)* = X3X2X1 - 2X1X3. Hence R(X} is the *-algebra freely generated by n symmetric letters. The subset of R(X} consisting of all symmetric nc polynomials is denoted by SymR(X} := {f e R(X} | f = f*}. If V = (vi) is a (column) vector of nc polynomials vi e R(X}, then V* is the row vector with components v* and V1 denotes the row vector with components vi. The length of the longest word in an nc polynomial f e R(X} is the degree of f and is denoted by deg f. The degree of f in Xi, degi f, is the largest number of occurrences of the letter Xi in a monomial appearing in f. Similarly, the length of the shortest word appearing in f e R(X} is called the min-degree of f and denoted by mindeg f. Likewise, mindegi f is introduced. If the variable Xi does not occur in any monomial of f, then mindegi f = 0. The set of all nc polynomials of degree < d will be denoted by R(X} 0 for every n-tuple of real symmetric matrices A of the same order. Note that positivity implies trace positivity while the converse is not true. Helton [16] and McCullough [27] proved that a symmetric nc polynomial f is positive if and only if it can be decomposed as a sum ofhermitian squares (SOHS), that is, there exist nc polynomials g1,... ,gm such that f = J2 m=19i9i. We denote all nc polynomials that admit SOHS decompositions as m S2 := {f G SymR(X> | f = ^g^, gi G R(X>, m > l}. i= 1 For trace positivity there is no necessary and sufficient condition of this type but there exists an important sufficient condition, obtained using cyclic equivalence to SOHS [18]; for a more example specific approach to certificates for trace positivity we refer to [36]. Nc polynomials f,g G R(X> are cyclically equivalent (f ~c g) if and only if there exist nc polynomials pi,qi G R(X> such that k f - g = Y.(pq - qiPi). i=1 We call an element of the form [p, q] := pq - qp, where p, q G R(X>, a commutator. Cyclically equivalent nc polynomials have equal trace if they are evaluated at the same n-tuple of real symmetric matrices, since the trace of every commutator of matrices is zero. Therefore if f is cyclically equivalent to SOHS, it is trace positive. We denote the set of nc polynomials of this type by e2 := {f G R(X> |3g G S2 : f c£c g}. By definition, the elements in e2 are exactly the nc polynomials which can be written as sums of hermitian squares with commutators. Although any bivariate nc polynomial of degree at most 4 is trace positive if and only if it is a sum of (four) squares with commutators [5, 8], there are trace positive nc polynomials which are not members of e2. Probably the easiest example is the noncommutative Motzkin polynomial, XY4X + YX4 Y - 3XY2X +1 [18, Example 4.4]; see also Subsection 3.3.2. We also refer the reader to [19, Example 3.5] for more sophisticated examples obtained by considering the BMV conjecture. Cyclic equivalence is obviously an equivalence relation. It can be easily detected by the following remark. Remark 1.2 ([18]). (a) For words v, w G (X>, we have v ~c w if and only if there are words v1,v2 G (X> such that v = v1v2 and w = v2v1. That is, v ~c w if and only if w is a cyclic permutation of v. (b) Nc polynomials f = Ewe aw w and g = Ewe bw w (aw, bw G R) are cyclically equivalent if and only if for each word v g (X >, Y bw. (1.1) ■E(X) weyev E E(X) a w Example 1.3. Let f = 1 + X2 + 2X2Y - 2XY + 2XY2X G R(X, Y). Since f = (X + XY)*(X + XY) + (1 - YX)* (1 - YX) + [X2 - X,Y] + [XY,YX] it follows that f c~ (X + XY)*(X + XY) + (1 - YX)*(1 - YX), and therefore f G в2. 1.2 Motivation and related work There is s surge of interest in free real algebraic geometry in the last decade, partially due to many facets of applications. A nice survey on connections to control theory, systems engineering and optimization is given by de Oliveira, Helton, McCullough, Putinar [13]. Applications to quantum physics are explained e.g. by Pironio, Navascués, Acfn [32] who also consider computational aspects related to sums of hermitian squares. On the theoretical level, trace positive nc polynomials arise e.g. in the Lieb-Seiringer reformulation of the famous Bessis-Moussa-Villani (BMV) conjecture [2] from statistical quantum mechanics, which was recently proved by Stahl [39]. This connection will be explained in detail later to demonstrate the usage of our proposed algorithm. In addition, trace positive nc polynomials occur naturally in von Neumann algebras and functional analysis. For instance, Connes' embedding problem [12] on finite II i-factors is a question about the existence of a certain type of sum of hermitian squares certificates for trace positive nc polynomials [18]. Motivated by this intensive research in free real algebraic geometry we have developed NCSOStools [10] - an open source Matlab toolbox for solving such problems using semidefinite programming. As a side product our toolbox implements symbolic computation with free noncommuting variables in Matlab. 1.3 Contribution The main contribution of this paper is the following. Once we know that a given rational nc polynomial f can be decomposed as a sum of hermitian squares (with commutators), i.e., we have numerical evidence for the existence of such a decomposition, we aim to obtain an exact (rational) certificate. Following ideas from [31] (see also [17]) we propose an algorithm which under a strict feasibility assumption theoretically and practically always yields a rational certificate. On the other hand, in the absence of strict feasibility, a variant of the facial reduction [3] (in our case projecting onto the orthogonal complement of the nullspace of the analytic center) is used to reduce the size of the semidefinite program and enforce the existence of Slater points. We employ the noncommutative version of Motzkin's polynomial to demonstrate how the proposed algorithm as implemented in NCSOStools is used and provide new rational certificates for some instances of nc polynomials related to the Bessis-Moussa-Villani conjecture. 2 Nc polynomials and semidefinite programming 2.1 Semidefinite programming Semidefinite programming (SDP) is a generalization of linear programming (LP) where one looks for the optimum of a linear function over the intersection of an affine subspace with the cone of positive semidefinite matrices. Although this is a far-reaching extension of LP, there exists several methods that can solve semidefinite programs efficiently in theory and practice. Given s x s self-adjoint matrices C, Ai,..., Am of the same size over R and a vector b e Rm, we formulate a semidefinite program in standard primal form as follows: inf (C, G) s.t. (A,, G) = b,, i = l,...,m (PSDP) G ^ 0. Here (•, •) stands for the standard inner product of matrices: (A,B) = tr(B*A),and G ^ 0 means that G is positive semidefinite. If C = 0 or if C is not important, we call such a problem a semidefinite programming feasibility problem: t /л G ~ 0, • i (FSDP) s.t. (Aj,G) = b,, i = 1,..., m. The complexity of solving semidefinite programs is mainly determined by the order s of matrix variable G and the number of linear constraints m. Given e > 0, the interior point methods can find an e-optimal solution with polynomially many iterations, where each iteration takes polynomially many real number operations, provided that (PSDP) and its dual both have non-empty interiors of feasible sets and we have good initial points. The variables appearing in these polynomial bounds are s, m and log e (cf. [40, Chapter 10.4.4]). Many problems in control theory, system identification and signal processing can be formulated using SDPs [4, 30, 1]. Combinatorial optimization problems can often be modeled or approximated by SDPs [14,23, 34, 35, 33]. SDP has important role in real algebraic geometry, where it is used e.g. for finding sums of squares decompositions of polynomials or approximating the moment problem [22, 21, 26, 24], and in free real algebraic geometry [18, 20, 6], as is recalled in the following subsection. 2.2 Sums of hermitian squares (with commutators) and semidefinite programming Testing whether a given nc polynomial f e R(X) is an element of S2 can be done efficiently by using semidefinite programming [20, 10]. This is the Gram matrix method, which is based on the following proposition [16, 28], the noncommutative version of the classical result for commuting variables. Proposition 2.1. Suppose the nc polynomial f e Sym R(X ) is of degree < 2d and let Wd be the vector of all words w e (X ) of degree < d. Then f e S2 if and only if there exists a positive semidefinite matrix Gf (called a Gram matrix for f ) satisfying f = WJGf Wd. Example 2.2. Take f = 1 + X2 + XY + YX + 4YX2 Y + Y2 and let V = [1 X Y XY f be a subvector of W2. Then the Gram matrix for nc polynomial f corresponding to the vector V is "10 0 u' ru s 0 1 1 — u 0 G(u) := 0 1 — u 1 0 0 0 4 The question is: does there exist (at least one) u such that G(u) is a positive semidefinite matrix? Since G(2) = C * C for C 10 0 2 0 1-10 it follows that f = (1 + 2XY)*(1 + 2XY) + (X - Y)*(X - Y) e S2. As we saw in the last example we can sometimes replace Wd with a smaller subvector in the Gram matrix method. An algorithm (the Newton chip method) for reducing the size of needed word vector is presented in [20] and is implemented in NCSOStools. See also [29] for a strengthening. Similarly we can use semidefinite programming to test whether a given nc polynomial f e R(X} is an element of в2 as first observed in [19], see also [10, 7, 6]. The method behind it is a variant of the Gram matrix method: Proposition 2.3. Suppose that an nc polynomial f e R(X} is of degree < 2d and let Wd be as above. Then f e в2 if and only if there exists a positive semidefinite matrix Gf (called a tracial Gram matrix for f ) such that f ~c W*Gf Wd. Again we can sometimes replace the full word vector Wd with a smaller subvector. An algorithm (the Newton cyclic chip method) for reducing the size of needed word vector is presented in [6] and is implemented in NCSOStools. Following Proposition 2.1, we can decide whether an nc polynomial f is a sum of hermitian squares by solving a semidefinite programming feasibility problem in the matrix variable G, where the constraints (Aj, G} = 6j are implied by the fact that for each product of monomials w e {p*q | p,q e W} the following must be true: ] Gp,q = aw ? (2.1) p,qeW p* q=w where aw is the coefficient of w in f (aw = 0 if the monomial w does not appear in f). Since any input nc polynomial f is symmetric (so aw = aw* for all w), the corresponding SDP feasibility problem is as follows: G ^ 0 s.t. (Aw,G} = aw + aw* Vw e {p*q | p, q e W}, where Aw = Aw* is the symmetric matrix defined by 2; if u*v e {w,w*}, w* = w, (Aw )u,v \ 1; (SOHSsdp) 0 if u* v e { w, w* } , w* = w, otherwise. Similarly, following Proposition 2.3, an nc polynomial f is cyclically equivalent to a sum of hermitian squares if and only if there exists a positive semidefinite matrix G such that f c~ W *GW. Again, this is an SDP feasibility problem (FSDP) in the matrix variable G, where the constraints (Aj, G} = 6j are essentially equations (1.1), i.e., for each product of monomials v e {p*q | p, q e W} the following must be true: Gp,q =53 aw. (2.2) p,q£W w%cv * СУС p q ~ v The SDP feasibility problem is as follows [6, Corollary 4.5]: G У 0, s.t E 0. Together with the fact that f is cyclically sorted, [18, Proposition 4.2] implies that f is trace positive. Since f is of degree 4 in two variables it is a sum of hermitian squares with commutators [5, 8] (with real coefficients). On the other hand, if (FSDP) admits a feasible positive definite solution, then it admits a (positive definite) rational solution. More exactly, we have the following: Theorem 3.2 (Peyrl & Parrilo [31]). If an approximate feasible point G0 for (FSDP) satisfies 5 := min(eig(Go)) > ||«Ai,Go> - 6;);|| =: e, (3.1) then a (positive definite) rational feasible point G exists. It can be obtained from G0 in the following two steps (cf. Figure 1): (1) compute a rational approximation G of G0 with т := ||G-G0|| satisfying т 2+e2 < 52; (2) project G onto the affine subspace L given by the equations {Ai, G> = bi to obtain G. PsD G G T Go Figure 1: Rounding and projecting to obtain a rational solution Note that the results in [31] are stated for SDPs arising from sum of squares problems, but their results carry over verbatim to the setting of (the seemingly more) general SDPs. The rationalization scheme based on this Peyrl-Parrilo technique has been implemented in NCSOStools; see Example 3.5 for a demonstration. 3.2 Facial reduction Not all is lost, however, if the SDP solver gives a singular feasible point G0 for (FSDP). Suppose that z is a rational nullvector for G0. Let P be a change of basis matrix containing z as a first column and a (rational) orthogonal basis for the orthogonal complement {z}± as its remaining columns. Then P tG0P i.e., Go = P- for some symmetric G0. Hence bi = (Ai, G0> = tr(AiG0) = tr (AiP-t So if 00 0 G0 0 0 ' 0 G0 P- 00 0 G0 P -1 tr (P -1A,P 00 0 G0 ai ci Ci Ai P-1AiP-t then Ai is a symmetric matrix with rational entries and bi = tr ai cit 0 0 ci Ai 0 G 0_ tr(AiG0) = (Ai, G?0>. We have established a variant of the facial reduction [3] which applies whenever the original SDP is given by rational data and has a singular feasible point with a rational nullvector: Theorem 3.3. Let (FSDP), Ai and G?0 be as above. Consider the feasibility SDP G У 0 s. t. (Ai,G> = bi, i = l,...,m (1) (FSDP') is feasible if and only if (FSDP) is feasible. (2) (FSDP') admits a rational solution if and only if (FSDP) does. (FSDP') 1 t 3.3 Examples 3.3.1 BMV conjecture In their 2004 paper [25], Lieb and Seiringer gave the following purely algebraic reformulation of the Bessis-Moussa-Villani (BMV) conjecture [2] from quantum statistical physics, which was recently proved in the original formulation by Stahl [39]: Conjecture 3.4. For all positive semidefinite matrices A and B and all m e N, the polynomial p(t) := tr((A + tB)m) e R[t] has only nonnegative coefficients. The coefficient of tk in p(t) for a given m is the trace of Sm,k (A, B), where Sm,k (A, B) is the sum of all words of length m in the letters A and B in which B appears exactly k times. For example, S4,2 (A, B) = A2B2 + AB AB + AB2 A + BABA + B2 A2 + BA2B. Thus Sm k (X, Y) is an nc polynomial; it is the sum of all words in two variables X, Y of degree m in which Y appears exactly k times. Even though the motivating conjecture was proved, the related questions concerning nc polynomials remain interesting. In the last few years there has been much activity around the following question: which pairs (m, k) does Sm,k(X2, Y2) e в2 or Sm,k(X, Y) e в2 hold for? An affirmative answer (for all m, k) to the former would imply the BMV conjecture. This question has been resolved completely (see e.g. [19, 11, 9]), however only finitely many nontrivial Sm,k(X2, Y2) admit a e2-certificate. Adding to the current state of knowledge (nicely summarized in [11]), we shall use our computer algebra system NCSOStools to establish Sio,2(X, Y) e в2 and sm^x, Y) e в2. We also show that S2m,2(X, Y) e в2 holds for all m e N. Example 3.5. Consider the nc polynomial f = S10,2(X, Y), i.e., the sum of all words of degree 10 in the nc variables X and Y in which Y appears exactly twice. To prove that f e в2 with the aid of NCSOStools, proceed as follows: (1) Define two noncommuting variables: >> NCvars x y (2) Our nc polynomial f is constructed using BMV(10,2). For a numerical test whether f e в2, run >> p.obj = 0; >> [IsCycEq,G0,W,sohs,g,SDP_data] = NCcycSos(BMV(10,2),p); Using the SDP solver SDPT3, this yields a floating point Gram matrix G o Go = 5.0000 2.5000 -1.8851 1.6770 10.6424 0.8230 -2.7313 1.6770 -0.0899 0.8230 -1.8851 2.5000 2.5000 -1.8851 0.8230 -0.0899 8.7702 1.6770 -2.7313 0.8230 1.6770 -1.8851 8.7702 2.5000 5.0000 for the word vector W = [X 4Y X 3YX X 2 YX2 XYX3 YXY . The rest of the output: IsCycEq = 1 since f is (numerically) an element of в2; sohs is a vector of nc polynomials gi with f ~c J2 i 9ì9ì = g; SDP_data is the SDP data for (2.2) constructed from f. (3) To round and project the obtained floating point solution G0 following Theorem 3.2, feed G0 and SDP_data into RprojRldlt: >> [G,L,D,P,err]=RprojRldlt(G0,SDP_data,true) This produces a rational Gram matrix G for f with respect to W and its LDU decomposition PLDL^-Pt, where P is a permutation matrix, L lower unitriangular, and D a diagonal matrix with positive entries. We caution the reader that l,d, and G are cells, each containing numerators and denominators separately as a matrix. Finally, the obtained rational sum of hermitian squares certificate for f = S10 2 (X, Y) is cyc f ~ Xi9i9i for gi = X2YX2 + — X3YX + — XYX3 - X4Y - ^YX4 g2 = X3YX - 577 1535 XYX3 + 408 1535 X 4y + 188 1535 YX4 дз = XYX 3 + 1Ì909 X 4Y +.7613 YX 4 45984 g4 = X4Y - 296301 YX4 15328 647065 g5 = YX4 and A1 = 11, A. 1535 2= A3 = 11496 A4 = 647065 A5 = 1242629 176 3 1535 4 183936 5 647065 This example is not surprising, as it is a particular instance of a larger pattern: Proposition 3.6. For all m G N we have: S2m,2(X, Y) G в2. Proof. We first point out that for all m G N we have S2m,2(X,Y )= 2 X aYX в YX2m-2-a-e a+e<2m-2 2m-2 tv v2m-2-t 2 (2m - 2 - t + 1)YXtYX t=0 2m-2 - 2 (2m - 2 - t + 1)(YXtYX2m-2-t + YX2m-2-tYXt) t=0 2m 2 - 2 ((2m - 2 - t +1)YXtYX2m-2-t + (t + 1)YXtYX 2 t=0 2m-2 m 2 YXtYX t 2m-2-t 2m-2-t t=0 Note that for t = 2s we have YXtYX2m—2—t c£c Xm-s-1YX2sYXm—s—1 e S2, hence we next turn our attention to words YXtYX2m—2—t for odd t. In such cases we write t = 2s + 1 and observe that 2s + 1vv2m-3-2s YX 2s+1YX CyC _ ^(Xs+1YXm-s-2 + XSYXm—s —1)*(Xs+1YXm-s-2 + XSYXm-s-1jj _ _Xm-s-2yX2s+2yXm-s-2 _ _xm-s-1YX2«yxm-s-1 2 2 ' Therefore each word with odd t is cyclically equivalent to a hermitian square minus two hermitian squares. These two negative hermitian squares cancel out with the "even" words for t = 2s and t = 2s + 2. In fact, each word with odd t cancels one half of these two even terms, hence all even terms finally cancel out and only one half of the first and the last even term remains (these two terms are cyclically equivalent). Finally we get S2m,2 (X? Y) 2m-2 СУС m ^ (X S + 1YX m-s-2 + X SYX m-s-1)(X s + ^YX m-s-2 + x SYX m-s-1)* 2 t=0 + Xm-1 Y 2X m-1' □ Example 3.7. We conclude this subsubsection by showing S14,6(X, Y) e O2. We define two noncommuting variables and run NCcycSos as in the previous examples: >> NCvars x y >> [IsCycEq,G0,V,sohs,g,SDP_data] = NCcycSos(BMV(14,6)); However, this seems to be an infeasible problem. In fact, we shall use the generated data SDP_data to prove it is strongly infeasible by computing a rational hyperplane separating O2 and S14,6 (X, Y). Let P be the set of all nc polynomials p with degX p = mindegX p = 8 and degY p = mindegY p = 6. Obviously, S14,6(X, Y) e P. Each p e P can be represented by a 35 X 35 Gram matrix using the basis V from given as output of NCcycSos. An important observation is that p e O2 if and only if there is a positive semidefinite matrix G satisfying p c~ V*GV, cf. Proposition 2.3. Let L : P ^ R be a linear *-map nonnegative on O2 n P. It can be represented as p ^ (M, Gp) for a symmetric 35 x 35 matrix M, where Gp is a Gram matrix for p. Since L(S2) С [0, то), the matrix M is positive semidefinite. The fact that L(f) = 0 for all f c~ 0, can be modeled with constraints (M, H) = 0 for all H e A^, cf. [9, Section 2.2]. Here, A^ is the orthogonal complement of the span of the Av from Section 2.2 in the set of symmetric matrices. Clearly, it suffices to consider H from a linearly independent generating subset C of A^. To express L(S14 6(X, Y)) < 0, we first compute a Gram matrix for S14 6(X, Y). The matrix A = SDP_data.A and vector b = SDP_data.b model the linear constraints (Av, G) = bv for v e (X, Y) with degX v = 8, degY v = 6. Hence a symmetrized solution of the linear system >> SDP_data.A\SDP_data.b will be a Gram matrix G for S14,6(X, Y). Now consider the feasibility SDP M ^ 0 s. t. (M, G) = -35, УН e C : {M, H) = 0. (Here, -35 is just a convenient scaling factor.) Every feasible point induces a hyperplane separating в2 and S14,6(X, Y). Solving this SDP with SeDuMi (using the trivial objective function C = 0) yields a floating point solution M0 in the relative interior of the optimal face, see Remark 2.4, with minimal eigenvalue S = 0.3426 and residual norm e = 6.8 • 10-9. Thus we can find a rational feasible solution M as explained in Theorem 3.2, using RprojRldlt. This proves S14,6(X, Y) e в2. 3.3.2 Noncommutative Motzkin polynomial The nc polynomial fMot(X, Y) = XY4X + YX4Y - 3XY2X + 1 e R(X, Y) is a noncommutative version of the (commutative) Motzkin polynomial. The Motzkin polynomial is a well-known example of a (commutative) polynomial which is nonnegative on R2 but is not a sum of squares of polynomials. Similarly, fMot is an example of trace positive nc polynomial which is not a member of в2 [18, Example 4.4]. Indeed, since the (commutative) Motzkin polynomial is not a sum of squares of polynomials, fMot is not a member of в2. An alternative proof for trace positivity of fMot (X, Y) follows from the fact that fMot (X3, Y3 ) e в2, as we can show with the aid of the facial reduction procedure from Subsection 3.2. Example 3.8. Consider f = fMot(X3, Y3) = X3Y12X3 + Y3X12Y3 - 3X3Y6X3 + 1. To prove that f e в2 with the aid of NCSOStools, proceed as follows: (1) Define two noncommuting variables and the nc polynomial f: >> NCvars x y >> f = xA3*yA12*xA3 + yA3*xA12*yA3 - 3*xA3*yA6*xA3 + 1; (2) Define a custom vector of monomials W >> W = {''; 'x*y*y'; 'x*x*y'; 'x*x*y*y*y*y'; 'x*x*x*x*y*y'; 'x*x*x*y*y*y*y*y*y'; 'x*x*x*x*y*y*y*y*y'; 'x*x*x*x*x*y*y*y*y'; 'x*x*x*x*x*x*y*y*y'}; (3) For a numerical test whether f e в2, run >> param.V = W; [IsCycEq,G0,W,sohs,g,SDP_data] = NCcycSos(f,param); This yields a floating point Gram matrix G0 that is singular. (4) Try to round and project the obtained floating point solution G0, feed GO and SDP_data into RprojRldlt: >> [G,L,D,P,err] = RprojRldlt(G0,SDP_data) This exits with an error, since unlike in Example 3.5, the rounding and projecting alone does not yield a rational feasible point. (5) Instead, let us reexamine G0. A detailed look at the matrix reveals three nullvectors. We thus run our interactive procedure which aids the computer in reducing the size of the SDP as in Theorem 3.3. >> [G,SDP_data] = fac_reduct(f,param) This leads the computer to return a floating point feasible point G0 e R9x9 and the data for this SDP, SDP_data. It also stays in interactive mode and the user can inspect the matrix and enter the nullvector z to be used in the dimension reduction. We feed in three nullvectors as a matrix of three columns: K>> z = [0-10; -10 0; 0 0 1; 0-10; 0-10; -10 0; 0 0 1; -1 0 0; 0 0 1]; return Inside the interactive routine this enables the computer to produce a positive definite feasible G0 e R6x6. Hence we exit the interactive routine. K>> stop = 1; return Now, NCSOStools uses G0 to produce a rational positive semidefinite Gram matrix G for f, which proves f e в2. Like in the Example 3.5, the solution G is a cell containing two matrices with numerators and denominators of the rational entries of G. The reader can verify that f ~c W*GW exactly by doing rational arithmetic or approximately by computing floating point approximation for G and using floating point arithmetic. (6) To compute the LDU decomposition PLDLfP1 for the rational Gram matrix G of f with respect to W (where G,l,d are cells, each containing numerators and denominators separately as a matrix) run >> [L,D,P] = Rldlt(G) The obtained rational sum of hermitian squares certificate for fMot (X3 , Y3 ) is then 6 fMot(X3,Y3) Xc E Aig^gi i= 1 for and gi — 1 — 1X 2 2 Y 4 _ 1 2 X4Y2 92 — XY2 — 1X 3Y 6 2 — 1X 5y 4 2 93 — X 2Y — 1X 4Y5 2 — 1X 6Y3 2 94 — X2Y4 - X4Y2 95 — X3Y6 - -X 5Y 4 96 — X4Y5 - X6Y3 Ai — A2 — A3 — 1, A4 — A5 — Аб Remark 3.9. We point out that this yields a rational sum of squares certificate for f(x3,y3) where f (x, y) — 1 + x4y2 + x2y4 — 3x2y2 is the commutative Motzkin polynomial. 4 Conclusions In this paper we considered nc polynomials p in freely noncommuting variables which can be decomposed as a sum of hermitian squares (and commutators) with a special focus on nc polynomials with rational coefficients that admit rational decompositions. We explained how to obtain rational decompositions in theory and practice: if the related semidefinite programming problems have strictly feasible solutions then the algorithm we proposed - a variant of Peyrl-Parrilo rounding and projecting method - always yields a rational (i.e., exact symbolic) decomposition. In the absence of strict feasibility we proposed a variant of the facial reduction to reduce the size of the semidefinite program and enforce the existence of Slater points. We implemented both methods in our open source software package NCSOStools [10] and demonstrated them on several illustrative examples. References [1] M. Anjos and J. Lasserre, Handbook of Semidefinite, Conic and Polynomial Optimization, volume 166 of International Series in Operational Research and Management Science, Springer, 2012. [2] D. Bessis, P. Moussa and M. Villani, Monotonic converging variational approximations to the functional integrals in quantum statistical mechanics, J. Mathematical Phys. 16 (1975), 23182325. [3] J. Borwein and H. Wolkowicz, Facial reduction for a cone-convex programming problem, J. Austral. Math. Soc. Ser. A 30 (1980/81), 369-380. [4] S. Boyd, L. E. Ghaoui, E. Feron and V. Balakrishnan, Linear Matrix Inequalities in System and Control Theory, Studies in Applied Mathematics, SIAM, 1994. [5] S. Burgdorf and I. Klep, Trace-positive polynomials and the quartic tracial moment problem, C. R. Math. Acad. Sci. Paris 348 (2010), 721-726. [6] S. Burgdorf, K. Cafuta, I. Klep and J. Povh, Algorithmic aspects of sums of Hermitian squares of noncommutative polynomials, Comput. Optim. Appl. 55 (2013), 137-153. [7] S. Burgdorf, K. Cafuta, I. Klep and J. Povh, The tracial moment problem and trace-optimization of polynomials, Math. Program. 137 (2013), 557-578. [8] K. Cafuta, On matrix algebras associated to sum-of-squares semidefinite programs, Linear Multilinear Algebra 61 (2013), 1496-1509. [9] K. Cafuta, I. Klep and J. Povh, A note on the nonexistence of sum of squares certificates for the Bessis-Moussa-Villani conjecture., J. Math. Phys. 51 (2010), 083521, 10. [10] K. Cafuta, I. Klep and J. Povh, NCSOStools: a computer algebra system for symbolic and numerical computation with noncommutative polynomials, Optim. Methods and Softw. 26 (2011), 363-380,http://ncsostools.fis.unm.si/ [11] B. Collins, K. Dykema and F. Torres-Ayala, Sum-of-squares results for polynomials related to the Bessis-Moussa-Villani conjecture, J. Stat. Phys. 139 (2010), 779-799. [12] A. Connes, Classification of injective factors. Cases IIi, IITO, IIIa, A = 1., Ann. of Math. (2) 104 (1976), 73-115. [13] M. de Oliveira, J. Helton, S. McCullough and M. Putinar, Engineering systems and free semi-algebraic geometry, in: M. Putinar and S. Sullivant (eds.), Emerging applications of algebraic geometry, Springer, New York, volume 149 of IMA Vol. Math. Appl., pp. 17-61, 2008. [14] M. Goemans and D. Williamson, Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming, Journal of the Association for Computing Machinery 42 (1995), 1115-1145. [15] M. Halickà, E. de Klerk and C. Roos, On the convergence of the central path in semidefinite optimization, SIAM J. Optim. 12 (2002), 1090-1099. [16] J. Helton, "Positive" noncommutative polynomials are sums of squares, Ann. of Math. (2) 156 (2002), 675-694. [17] E. Kaltofen, B. Li, Z. Yang and L. Zhi, Exact certification in global polynomial optimization via sums-of-squares of rational functions with rational coefficients, J. Symbolic Comput. 47 (2012), 1-15. [18] I. Klep and M. Schweighofer, Connes' embedding conjecture and sums of Hermitian squares, Adv. Math. 217 (2008), 1816-1837. [19] I. Klep and M. Schweighofer, Sums of Hermitian squares and the BMV conjecture, J. Stat. Phys 133 (2008), 739-760. [20] I. Klep and J. Povh, Semidefinite programming and sums of hermitian squares of noncommutative polynomials, J. Pure Appl. Algebra 214 (2010), 740-749. [21] J. B. Lassere, Moments, positive polynomials and their applications, volume 1 of Imperial College Press Optimization Series, Imperial College Press, London, 2009. [22] J. B. Lasserre, Global optimization with polynomials and the problem of moments, SIAM J. Optim. 11 (2000/01), 796-817. [23] M. Laurent and F. Rendl, Semidefinite programming and integer programming, in: G. N. K. Aardal and R. Weismantel (eds.), Discrete Optimization, Elsevier, volume 12 of Handbooks in Operations Research and Management Science, pp. 393 - 514, 2005. [24] M. Laurent, Sums of squares, moment matrices and optimization over polynomials, in: Emerging applications of algebraic geometry, Springer, New York, volume 149 of IMA Vol. Math. Appl., pp. 157-270, 2009. [25] E. Lieb and R. Seiringer, Equivalent forms of the Bessis-Moussa-Villani conjecture, J. Stat. Phys. 115 (2004), 185-190. [26] M. Marshall, Positive polynomials and sums of squares, volume 146 of Mathematical Surveys and Monographs, American Mathematical Society, Providence, RI, 2008. [27] S. McCullough, Factorization of operator-valued polynomials in several non-commuting variables, Linear Algebra Appl. 326 (2001), 193-203. [28] S. McCullough and M. Putinar, Noncommutative sums of squares, Pacific J. Math. 218 (2005), 167-171. [29] C. Nelson, A real nullstellensatz for matrices of non-commutative polynomials, http:// arxiv.org/abs/1305.07 9 9 [30] P. Parrilo, Structured semidefinite programs and semialgebraic geometry methods in robustness and optimization, Ph.D. thesis, California Institute of Technology, 2000. [31] H. Peyrl and P. Parrilo, Computing sum of squares decompositions with rational coefficients, Theoret. Comput. Sci. 409 (2008), 269-281. [32] S. Pironio, M. Navascués and A. Acfn, Convergent relaxations of polynomial optimization problems with noncommuting variables, SIAM J. Optim. 20 (2010), 2157-2180. [33] J. Povh, Contribution of copositive formulations to the graph partitioning problem, Optimization: A Journal of Mathematical Programming and Operations Research (2011), 1-13. [34] J. Povh and F. Rendl, A copositive programming approach to graph partitioning, SIAM Journal on Optimization 18 (2007), 223-241. [35] J. Povh and F. Rendl, Copositive and semidefinite relaxations of the quadratic assignment problem, Discrete Optimization 6 (2009), 231-241. [36] R. Quarez, Some examples of trace-positive quaternary quartics, to appear in Proc. Amer. Math. Soc., http://hal.archives-ouvertes.fr/hal-0 0 685397 [37] B. Recht, M. Fazel and P. A. Parrilo, Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization, SIAM Rev. 52 (2010), 471-501. [38] C. Scheiderer, Sums of squares of polynomials with rational coefficients, to appear in J. Eur. Math. Soc., http://arxiv.org/abs/1209.2 976 [39] H. R. Stahl, Proof of the BMV conjecture, Acta Math. 211 (2013), 255-290. [40] H. Wolkowicz, R. Saigal and L. Vandenberghe, Handbook of semidefinite programming: Theory, Algorithms, and Applications, volume 27 of International Series in Operations Research & Management Science, Kluwer Academic Publishers, Boston, MA, 2000. /^creative ^commor Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 261-266 On mixed discriminants of positively definite matrix* Chang-Jian Zhao f Department of Mathematics, China Jiliang University, Hangzhou 310018, P. R. China Xiao-Yan Li Department of Mathematics, Hunan Normal University, Changsha 410000, P. R. China Received 2 November 2012, accepted 21 February 2014, published online 11 January 2015 Abstract In the paper, some new inequalities for the mixed discriminants of positively definite matrix are established, which are the matrix analogues of inequalities of the well-known mixed volumes function. Keywords: Mixed discriminants, symmetric matrix, mixed volume, Aleksandrov's inequality. Math. Subj. Class.: 15A09, 52A40 1 Introduction Let x1,... ,xn be a set of nonnegative quantities and let Ei(x) be the i-th elementary symmetric function of an n-tuple x = x(xi,..., xn) of non-negative reals is defined by E0(x) = 1 and EE i ( jx ) — ^^^ jx j i jx • • • jx j., 1 ^^ i ^^ 1 det(K) + det(L) det(Ki + Li)~ det(Ki) det(Li) ' An interesting proof is due to Bellman [4] (also see [2], p. 67). A generalization of (1.2) was established by Ky Fan [5] (also see [6-7]). Moreover, we assume all positive definite matrix are supposed to be symmetric in the article. There is a remarkable similarity between inequalities for symmetric functions (or determinants of symmetric matrices) and inequalities for the mixed volumes of convex bodies. In 1991, V. Milman asked if there is version of (1.1) or (1.2) in the theory of mixed volumes and it was stated as the following open question (see [8]): Question 1.1. For which values of i is it true that for very pair of convex bodies K and L in Rn, Wi(K + L) Wi(K ) + Wi(L) ? Wi+i(K + L) > Wi+i(K) + Wi+i(L) ? ( . ) The convex body is the compact and convex subsets with non-empty interiors in Rn. Wi(K) denotes the quermassintegral of convex body K and Wi+i (K) denotes the mixed volumes V(K,... , K, B,..., B). The sum + is the usual Minkowski vector sum and B n-i-1 i+1 denotes the unit ball. A theorem by Minkowski provides a fundamental relation between volume and operations of addition and multiplication of convex bodies by nonnegative reals: If K1,..., Km are convex bodies, m g N, then the volume of t1K1 + • • • + tmKm is a homogeneous polynomial of degree n in ti > 0 (see [14]). That is V(t1K1 + ••• + tmKm)= ^ V(Kh ,...,Kin )ti i ••• tin, 1 0, can be expanded as a polynomial in t: n n V(K + tB) = £ (. JWi(K)ti i=0 where Wi(K) := V(K,..., Kf, B,..., B) is the quermassintegral of convex body K. n—i i A partial answer (L must be a ball) of (1.3) was established by Gianopoulos, Hart-zoulaki and Paouris [9]). If K is a convex body and D is a ball in Rn, then for i = 0,..., n — 1 Wi(K + D) Wi(K ) + Wi (D) Wi+1(K + D) > Wi+1(K) + Wi+1 (D) (1.4) The answer to the above question is negative; it can be proved that (1.3) is true in full generality only when i = n — 1 or i = n — 2 (the details see [10]). Moreover, a dual inequality of (1.4) for the dual quermassintegral of star bodies was proved by Li and Leng [11]. In the paper, we establish some inequalities for mixed discriminants of positively definite matrix which are matrix analogues of some mixed volumes inequalities. 2 Mixed discriminants and Aleksandrov's inequality Recall that for positive definite n x n matrices K1,..., KN and X1,... ,XN > 0, the determinant of the linear combination AiKi +-----+ ANKN is a homogeneous polynomial of degree n in the Аг (see e.g. [12]), det(AiKi + ••• + An Kn )= 2 DK,..., K-„ )Ah ••• A-„, (2.1) i D(Ki,Ki, K3,..., Kn)D(K2, K2, K3,..., Kn), (2.4) with equality if and only if Ki=AK2 with positive number A. 3 Inequalities for mixed discriminants of positively definite matrix Theorem 3.1. Let K be symmetric positively definite matrix and I stands for the identity matrix and t > 0. If 0 < i < n — 1 and i G N, then the function Di(K + tI ) g(t) = Di+i(K + tI ) (3Л) is an increasing and concave function on [0, +то). Proof. If fi (s) = Di(K + si ), then by the linearity of the mixed discriminant we see that fi(s + £) = £ j=0 £j Di+j (K + si ) Hence Similarly, we obtain = fi(s) + e(n - i)fi+i(s) + o(e2). dfi(s) = lim f (s + £) - f (s) ds £ = (n - i)fi+1(s). dfi+i(s) ds = (n - i - l)fi+2(s). From (2.4), we obtain for all 0 < i < n f2+i(s) - fi(s)fi+2(s) > 0, with equality if and only if K = pi. From (3.2), (3.3) and (3.4), we have dfi(s) f ^ л dfi+i(s) —;—fi+i(s) - f(s)-1- ds ds (3.2) (3.3) (3.4) Therefore = fi+i(s) + (n - i - 1)(fi+i(s) - fi(s)fi+2(s)) > fi+i(s). dg(s) = f fi(s) V ds V fi+i(s)y f/(s)fi+i(s) - f (s)fi+i(s) fi+i(s) = (n - i) - (n - i - l) fi(s)fi+2(s) fi+i(s) . (3.5) Hence f (t) = Di(K + ti ) Di+i(K + ti ) is an increasing and concave function on [0, +то). □ Theorem 3.2. Let K be symmetric positively definite matrix and I stands for the identity matrix. If 0 < i < n, then (n - i)Di+2(K)(Di+i(K)2 - Di(K)Di+2(K)) > (n - i - 2)Di(K)(Di+2(K)2 - Di+i(K)Di+3(K)). (3.6) n—i j Proof. Let /i(t) = Di(K + ti) for t > 0 and g(t) = /i(t/L , then /i+1(t) i+1 dg(s)_, л , . r/i(tJ/+2(t) f 2 I i+1 d_ = (n - i) - (n - i - 1) / t) . (3.7) By differentiating the both sides of (3.6) again, we have d2r=-<■• - • - D (n - i)/i+2(t)/i+1 (t) + (n - i - 2)/i(t)/i+1(t)/i+3(t) - 2(n - i - 1 )/i(t)/i+2(t) x /3+1 (t) . (3.8) From (3.8) and in view of g(s) being a concave function, we obtain (n - i)/i+2(t)/+1(t) + (n - i - 2)/i(t)/i+1(t)/i+3(t) - 2(n - i - 1)/i(t)/2+2(t) > 0, for t G (0, +то). This can be equivalently written in the form (n - i)/i+2(t)(/+1(t) - /i(t)/i+2(t)) > (n - i - 2)/i(t)(/i+ 2(t) - /i+1(t)/i+3(t)). (3.9) Hence (n - i)Di+i(K + ti) (Di+1(K + ti)2 - Di(K + ti)Di+i(K + ti)) > (n - i - 2)Di(K + ti) (Di+i(K)2 - Di+1 (K + ti)Di+3(K + ti)). (3.10) Notice that /i(t) is continuous function, letting t ^ 0+ in (3.10), (3.10) reduces to the inequality in Theorem 3.2. □ References [1] A. D. Aleksandrov, Zur Theorie der gemischten Volumina von konvexen Körpen, Die gemischten Diskriminanten und die gemischten Volumina (in Russian), Mat. Sbornik 3 (1938), 227252. [2] E. F. Bechenbach, R. Bellman, Inequalities, first ed., Springer-Verlag, Berlin-Gottingen, Heidelberg, 1961. [3] R. Bellman, Notes on matrix theory-IV: an inequality due to Bergstrom, Amer. Math. Monthly 62 (1955), 172-173. [4] H. Bergstrom, A triangle inequality for matrices. Den Elfte Skandinaviski Matematiker-kongress, Trondheim, first ed., 1949, Oslo: John Grundt Tanums Forlag, 1952. [5] A. Dembo, T. M. Cover and J. A. Thomas, Information theoretic inequalities, IEEE Trans. Infor. Theory 37 (1991), 1501-1518. [6] K. Fan, Some inequalities conerning positive-definite hermitian matrices, Proc. Cambridge Phil. Soc. 51 (1955), 414-421. [7] K. Fan, Problem 4786, Amer. Math. Monthly 65 (1958), 289. [8] M. Fradelizi, A. Giannopoulos and M. Meyer, Some inequalities about mixed volumes, Israel J. Math. 135 (2003), 157-179. [9] A. Giannopoulos, M. Hartzoulaki and G. Paouris, On a local version of the Aleksandrov-Fenchel inequality for the quermassintegrtals of a convex body, Proc. Amer. Math. Soc. 130 (2002), 2403-2412. [10] X. Y. Li and G. S. Leng, Some inequalities about dual mixed volumes of star bodies, Acta Math. Sci. 25 (2005), 505-510. [11] E. Lutwak, D. Yang and G. Zhang, A new affine invariant for polytopes and Schneider's projection problem, Trans. Amer. Math. Soc. 353 (2001), 1767-1779. [12] M. Marcus and M. Marcus, Inequalities for symmetric functions and Hermitian matrices, CanadJ. Math. 8 (1956), 524-531. [13] L. Mirsky, Maximum principles in matrix theory, Proc. Glasgow Math. Assoc. 4 (1958), 34-37. [14] P. Pranayanuntana, Elliptic Brunn-Minkowski Theory, first ed., UMT Dissertations Publishing, 2003. [15] R. Schneider, Convex Bodies: The Brunn-Minkowski Theory, first ed., Cambridge Univ. Press, 1993. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 267-277 Odd edge coloring of graphs Borut Lužar Faculty of Information Studies, 8000 Novo mesto, Slovenia and Institute of Mathematics, Physics and Mechanics, 1000 Ljubljana, Slovenia Mirko Petruševski Department of Mathematics and Informatics, Faculty of Mechanical Engineering, Skopje, Republic of Macedonia Received 24 November 2013, accepted 1 October 2014, published online 11 January 2015 An edge coloring of a graph G is said to be an odd edge coloring if for each vertex v of G and each color c, the vertex v uses the color c an odd number of times or does not use it at all. In [5], Pyber proved that 4 colors suffice for an odd edge coloring of any simple graph. Recently, some results on this type of colorings of (multi)graphs were successfully applied in solving a problem of facial parity edge coloring [3, 2]. In this paper we present additional results, namely we prove that 6 colors suffice for an odd edge coloring of any loopless connected (multi)graph, provide examples showing that this upper bound is sharp and characterize the family of loopless connected (multi)graphs for which the bound 6 is achieved. We also pose several open problems. Keywords: Edge coloring, odd subgraph, Shannon triangle. Math. Subj. Class.: 05C15 E-mail addresses: borut.luzar@gmail.com (Borut Lužar), mirko.petrushevski@gmail.com (Mirko Petruševski), skrekovski@gmail.com (Riste Skrekovski) Riste Škrekovski Institute of Mathematics, Physics and Mechanics, 1000 Ljubljana, Slovenia and Faculty of Information Studies, 8000 Novo mesto, Slovenia and University of Primorska, FAMNIT, 6000 Koper, Slovenia Abstract 1 Introduction Throughout the article we mainly follow the terminology and notation used in [1]. A graph is denoted by G = (V (G), E (G)), where V (G) is the vertex set and E (G) is the edge set. A graph G is always regarded as being finite (i.e. having finite number of vertices n(G), and finite number of edges m(G)) with loops and multiple edges allowed. The parameters n(G) and m(G) are usually called order and size of G, respectively. Whenever n(G) = 1 we say that G is trivial and whenever m(G) =0 we say that G is empty. For X С V (G) U E (G), the subgraph obtained by removing the vertices and edges from the set X is denoted by G - X. We refer to the vertices having even (resp. odd) degree as even (resp. odd) vertices. A graph is called even (resp. odd) whenever all of its vertices are even (resp. odd). An odd edge coloring of G is a (not necessarily proper) edge coloring such that each color class induces an odd subgraph of G. An odd edge coloring of G using at most k colors is referred to as an odd k-edge-coloring, and we say that G is odd k-edge-colorable. If G admits an odd edge coloring, the odd chromatic index x'O (G) is defined to be the minimum integer k for which G is odd k-edge-colorable. By definition, each loop at a vertex v colored with c contributes 2 to the count of appearances of c at v. Thus, it is obvious that a necessary and sufficient condition for the existence of an odd edge coloring of G is the absence of vertices incident only to loops. Apart from this, the presence of loops does not influence the existence nor changes the value of the index xO(G). Therefore, we choose to restrict our attention to loopless connected graphs throughout the article. As a notion, odd edge coloring was first introduced by Pyber in his survey on graph coverings [5] as an edge decomposition of a graph into (edge disjoint) odd subgraphs. Such decompositions represent a counterpart to decompositions into even subgraphs, which were mainly used while proving various flow problems (see e.g. [6]). In the mentioned work, Pyber considered simple graphs and proved the following result. Theorem 1.1 (Pyber, 1991). For every simple graph G, it holds that xO(G) < 4. He also remarked that the upper bound is realized by a wheel on four spokes W4 and asked whether there is an infinite family of connected graphs for which this bound is achieved. In 2006, Mätrai [4] presented such a construction, taking an even number of copies of W4 and an additional vertex v. Choosing an arbitrary edge from the wheel, he removed the same edge from every copy and connected its two end-vertices with v (see Fig. 1). From xO(W4) = 4 readily follows that the obtained graph has xO equal to 4. A generalization of Theorem 1.1 was successfully applied in solving a problem of facial parity edge colorings in [2], and its improvement in [3]. In this paper, an analogous result to Theorem 1.1 is proved for loopless graphs. Namely, in Theorem 11 we prove that 6 colors suffice for an odd edge coloring of any loopless connected graph, and characterize the family of loopless connected graphs needing 6 colors. 2 Preliminary results It is an easy matter to characterize the graphs having xO < 1: namely, xO(G) =0 if and only if G is empty, while xO(G) = 1 if and only if G is nonempty and the subgraph of G Figure 1: A graph with odd chromatic index equal to 4. induced by its non-isolated vertices is odd. The following result was initially stated in [5], but for the sake of completeness we present it here with a proof. Proposition 2.1. For every forest F, it holds that x'0(F ) < 2. Proof. It is enough to prove this for an arbitrary tree T. If T is trivial or odd, then x0 (T) < 1, so suppose this is not the case. We construct an odd 2-edge-coloring of T. Take an even vertex r as the root of T. To begin with, color the edges incident to r by using the color 1 for all but one such edge, and color this remaining edge by the color 2. Continue by coloring the incident edges to each vertex v which has one incident edge already colored as follows: • if v has even degree in T, then we complete the coloring of its incident edges by coloring them in the other color (if color 1 was used for the already colored edge, then we use color 2 for the remaining edges, and vice versa); • if v has odd degree in T, then we complete the coloring of its incident edges by coloring them with the same color as the already colored edge. Since there are no cycles in T, every vertex u = r eventually is in a position to have exactly one of its incident edges colored. Namely, consider the vertices along the unique ru-path in T and suppose the opposite, i.e. suppose there exist at least one vertex on this path that never gets in the stated position. Choose the first such vertex after r (denote it by w) on the tracing of this path. Thus, the predecessor of w gets in the stated position. But this implies that w also gets in position, a contradiction. Therefore, the above procedure produces an odd 2-edge-coloring of T. □ Let G be a graph and T be an even-sized subset of V(G). Following [1], a spanning subgraph H of G is said to be a T-join if dH (v) is odd for all v g T and even for all v g V(G)\T. For example, if P is an xy-path in G, the spanning subgraph of G with edge set E (P ) is an {x, y}-join. Note that by removing (resp. adding) a cycle (as an edge set) from (resp. to) a T -join we again obtain a T-join. Thus, whenever a T-join exists, there also exists such a forest (resp. coforest). A classical result about T-joins (see [1]) is that whenever G is connected, there exists a T -join for every even-sized subset T of V (G). Consider a connected graph G of even order, and let T be the set of its even vertices. The handshake lemma assures that T has even size, hence there exists a T-join H. By putting K := G - E (H ) we obtain an odd factor of G, i.e. a spanning odd subgraph. Note that if the T-join H was chosen to be a forest, then the obtained odd factor K satisfies the following statement. Proposition 2.2. Given a connected graph G of even order, there exists an odd factor K of G such that G — E (K ) is a forest. 3 Tight upper bound for x'o In this section, through a number of propositions we derive the main result of the paper, a general tight upper bound xO < 6 with a characterization of the loopless connected graphs for which the bound is achieved. Proposition 3.1. Given a loopless connected graph G of even order, it holds that xO(G) < 3. If furthermore G is even, then xO(G) < 2. Proof. By Proposition 2.2, we can take an odd factor K of G such that G — E (K ) is a forest denoted by F. From Proposition 2.1 we infer that xO(G) < x'o(F)+ x'o(K) < 2 + 1 = 3 . If in addition G is even, then F is a spanning odd forest of G, giving x'o(G) < x'o(F)+ xO(K) < 1 + 1 = 2, which completes the proof. □ Let G be a loopless graph. By a bouquet of parallel edges in G we refer to a subset of E(G) consisting of all the edges linking a pair of adjacent vertices. Figure 2: A loopless graph (left) and its reduction (right). The reduction red(G) of a loopless graph G is defined to be a spanning subgraph of G obtained by the following change at every bouquet of parallel edges: remove maximum possible even number of edges without altering the adjacency relation in V(G) (see Fig. 2 for an example). Obviously, up to isomorphism, each loopless graph has a unique reduction. We say that a loopless graph G is reduced whenever its multiplicity is at most 2, i.e. when G = red(G). It was already remarked in [2] that an odd k-edge-coloring of its reduction readily provides an odd k-edge-coloring of any loopless graph G, since the removed edges between any two adjacent vertices may all adopt one arbitrary color used on the remaining edges between them in red(G). Hence, xO(G) < xO(red(G)). (3.1) Remark 1. Regarding the inequality (3.1), suppose that a graph G satisfies xO(G) < xO(red(G)). In respect of the reduced graph red(G), assume G is minimal such graph in terms of size. Consider an arbitrary optimal odd edge coloring of G. Then, by the minimality of G, on each bouquet of parallel edges no color appears more than once, unless possibly on a bouquet of size 2. In other words, in every optimal odd edge coloring, on any bouquet of parallel edges in G that reduces in size for red(G), no color is repeated. Hence, whenever the inequality (3.1) is strict, it holds that xO(red(G)) > 4. A loopless graph G on three pairwise adjacent vertices is referred to as a Shannon triangle. Denote the sizes of its bouquets of parallel edges by p, q, r in non-increasing order. We say that the considered Shannon triangle is of type (p, q, r). In particular, whenever p, q, r are even, we speak of a Shannon triangle of even type. Observe that there are exactly four different types of reduced Shannon triangles and only one of them is of even type (depicted in Fig. 3). Next, we prove that (3.1) is always an equality for the case of Shannon triangles. Furthermore, we prove that the reduced Shannon triangles of different types attain odd edge chromatic indices for all integer values between 3 and 6. Proposition 3.2. Given a Shannon triangle G, let (p, q, r) be the type of red(G). Then xO(G) = xO (red(G)) = p + q + r. (3.2) Figure 3: The four types of Shannon triangles with odd chromatic indices 3, 4, 5, and 6, respectively. The last one is of even type. Proof. Observe that in an arbitrary odd edge coloring of a Shannon triangle, no color appears altogether an even number of times on any bouquet. Namely, denote the vertices of the triangle by u, v, and w and suppose that on the edges between u and v some color, say 1, appears in total an even number of times. This implies that 1 appears overall an odd number of times on each of the remaining two bouquets. Thus, 1 appears an even number of times at the vertex w, a contradiction. This clearly implies that no color is repeated in an arbitrary odd edge coloring of a reduced Shannon triangle. Hence the second equality in (3.2) is established. Next, we prove the first equality. Suppose there exists a Shannon triangle G for which x'o(G) < xO(red(G)). In respect of the reduced graph red(G), assume G is minimal such graph in terms of size. Then, Remark 1 and the above observation imply that no color is repeated in an optimal odd edge coloring of G. This readily gives xO(G) > xO(red(G)), a contradiction. □ In the following, we give several other propositions leading to the main theorem of this article, but first we introduce some additional notation. Let v be a vertex of a reduced graph G and Sv be the subgraph of G induced by the set of edges incident to v, i.e. Sv := G[E(v)]. Each pair of parallel edges from this subgraph is said to be a petal of v. Each edge of a petal is referred to as a petal edge of v. The other edges incident to v are called leaf edges of v. Denote by p(v) and l(v) the number of petals and number of leaf edges of v, respectively. Proposition 3.3. If a connected reduced graph G has a non-cutvertex v for which either p(v) is odd or l(v) = 0, then xO(G) < 5. Proof. By Proposition 3.1, we may assume n(G) is odd. Suppose v is a non-cutvertex of G such that either p(v) is odd or l(v) = 0. We consider the four possible cases according to the parities of p(v) and l(v): (i) p(v) is odd and l(v) is even. By Proposition 3.1, there exists an odd 3-edge-coloring ( of G - v with color set {1, 2, 3}. We extend ( to G by using two additional colors 4 and 5: color by 4 each leaf edge of v and precisely one petal edge from each petal of v; color by 5 the remaining petal edges of v. We obtain an odd 5-edge-coloring of G. (ii) p(v) is odd and l(v) is odd. Let e be a leaf edge of v. Let F be a forest in G - v (as in the proof of Proposition 3.1), and add v together with the leaf edge e to F. We obtain a forest F' in the subgraph G' := G - (E(v)\e). Since two colors suffice for an odd edge coloring of F' and the edge set E(G')\E(F') induces an odd subgraph of G', by using a third color for the edges from E(G')\E(F') we obtain an odd 3-edge-coloring ( of G'. We extend ( to an odd 5-edge-coloring of G as in (i). (iii) p(v) is even and l(v) is odd. In this case, we use a similar approach as in (ii), but take a petal edge of v as the edge e, instead. Again, we finish as in (i). (iv) p(v) is even and l(v) is even and positive. Let ( be an odd 3-edge-coloring of G — v with color set {1, 2, 3}. Extend ( to G by coloring with 4 one leaf edge of v and precisely one petal edge of each petal of v, and moreover, by coloring with 5 all the remaining uncolored edges incident to v. We obtain an odd 5-edge-coloring of G. □ Whenever l(v) = 0, we refer to Sv as the orchid at v. According to the parity of p(v), we distinguish between even and odd orchids. A graph obtained from a path of length k > 1 in which every edge is replaced by two parallel edges is called an open k-necklace. Observe that every open k-necklace is an even graph of order k +1. Similarly, a graph obtained from a cycle of length k > 2 in which every edge is replaced by two parallel edges is called a closed k-necklace. Every closed k-necklace is an even graph of order k. Proposition 3.4. The odd chromatic index of an open k-necklace G satisfies X0(G) 2 if k is odd, 4 if k is even. Proof. Assume k is odd. Since G needs at least two colors for an odd edge coloring, Proposition 3.1 implies xO(G) = 2. We construct a particular odd 2-edge-coloring of G that we will use in Proposition 8. Fix one of the two natural orders for the bouquets of parallel edges of G, and color as follows: 1) for the edges of the first, third, fifth,..., k-th such bouquet use each of the colors 1 and 2 once; 2) for the edges of the second, fourth, sixth,..., (k - 1)-st such bouquet use the color 1 twice. Assume now k is even. First we establish the inequality xO(G) < 4 by constructing an odd 4-edge-coloring of G. Again, by fixing one of the two natural orders for the bouquets of G: on the edges of the first (k - 1) bouquets apply the odd 2-edge-coloring of 1) and 2); for the edges of the k-th bouquet use each of the colors 3 and 4 once. Second we prove xO(G) = 4. Suppose xO(G) < 4 and consider an optimal odd edge coloring of G. Fix one of the two natural orders for the bouquets of G. Obviously, the first bouquet is dichromatic, i.e. two colors are used for its edges. Hence, xO(G) < 4 implies that the second bouquet is monochromatic. But then the third bouquet is dichromatic, etc. We deduce that the k-th bouquet is monochromatic. This is a contradiction, hence Proof. Assume k is even. Since G needs at least two colors for an odd edge coloring, Proposition 3.1 implies xO(G) = 2. Assume next k is odd and k > 5. First we prove xO(G) < 4 by constructing an odd 4-edge-coloring of G. Remove a vertex v from G to obtain the open (k — 2)-necklace G — v. Use the odd 2-edge-coloring of G — v constructed in Proposition 3.4 with a single change of color: for one edge from the last bouquet instead of using the color 2 use the color 3. Now color the edges of the orchid at v: • for the uncolored petal of v "neighboring" a bouquet from G — v colored by 1 and 2, use the color 2 for both edges; • for the remaining uncolored petal of v use the colors 2 and 4 once each. Now, suppose xO(G) < 4 and consider an optimal odd edge coloring of G. Obviously, at each vertex the orchid is dichromatic with three of its edges having the same color. Thus, any two consecutive bouquets in G are such that one is monochromatic and the other is xO(G) = 4. □ Proposition 3.5. The odd chromatic index of a closed k-necklace G satisfies dichromatic. Hence, precisely half of the k bouquets in G are monochromatic. This is a contradiction with the parity of k. Finally, note that for k = 3, the closed k-necklace G is the Shannon triangle of type (2,2, 2) and thus xO(G) = 6 (see Proposition 3.2). □ We are now in a position to prove our main result for loopless 2-connected graphs. Proposition 3.6. Let G be a loopless 2-connected graph which is not isomorphic to a Shannon triangle of even type. Then xO(G) < 5. Proof. By Proposition 3.1, inequality (3.1) and Proposition 3.3, we may assume that G has odd order n and is a 2-connected reduced graph with an even orchid at every vertex. Denote by v a vertex of maximum degree in G. If the orchid Sv has precisely two petals, then G is a closed n-necklace with n > 5 (namely, if n = 3 then G would be the Shannon triangle of type (2,2, 2)). Hence, in this case Proposition 3.5 implies xO(G) = 4. So, we may assume Sv has at least four petals. Consider the graph G - v and denote by u1 , u2,..., u2s the neighbors of v in G. By Proposition 3.1, we have xO(G - v) = 2. Consider an initial odd 2-edge-coloring of G - v with color set {1,2} such that the edges of the spanning odd forest F constructed in the proof of Proposition 3.1 are colored by 1. Denote by M the collection of maximal paths in F that have non-empty intersection with the set {u1,u2,... ,u2s}. Note that every member of M is non-trivial and has both end-vertices among the leaves of F. We distinguish between the following three cases: (i) At least one P e M does not have both end-vertices in the set {u1,u2,..., u2s}. Start a tracing of one such path P from an end-vertex not belonging to {u1,u2,... ,u2s}. Let u be the first vertex from V (P ) n {u1, u2,..., u2s} met on this tracing, and denote by P0 the traced subpath of P. Beginning at the edge incident to u, recolor the edges of P0 by the colors 3 and 4, alternatingly. Color the petal edges of v in the following way: use the color 4 for one vu-edge, while using the color 1 for all the remaining petal edges incident to v. Thus, we obtain an odd 4-edge-coloring of G. (ii) Each member of M has both end-vertices in the set {u1, u2,..., u2s} and there is such a path P e M of length at least 2. Denote by u and u j the end-vertices of P. We recolor the edges of P by using the colors 3,4 and 5 as follows: the edge incident to u is recolored by 3 and the edge incident to u j is recolored by 4. The other edges along P are recolored in such a way to obtain a proper edge coloring of P, which is clearly achievable. Color the petal edges of v in the following way: • use color 4 for both vuj-edges; • use color 4 for one vuj-edge; • use color 1 for the remaining petal edges of v. Thus, we obtain an odd 5-edge-coloring of G. (iii) Each member of M has both end-vertices in the set {u1, u2,..., u2s} and there is no such path of length at least 2. In this case, the edges of F incident to at least one of the vertices u1; u2,..., u2s form a matching on the set {u1; u2,..., u2s}. Without loss of generality, suppose this matches u2i-1 with u2i, for every i e {1,2,..., s}. From the initial odd 2-edge-coloring of G - v, we obtain an odd 5-edge-coloring of G as follows: • recolor the edge uiu2 in F with the color 5; • use both the colors 1 and 4 once for the petal vui-edges; • use both the colors 3 and 4 once for each of the petals vu2, vu3,..., vu2s-1; • use both the colors 3 and 5 once for the petal vu2s-edges (here we make use of the fact that s > 2). This establishes the inequality xO(G) < 5, which completes the proof of the statement. □ 3 2 Figure 4: An odd 4-edge-coloring. Next we prove that 5 colors suffice for an odd edge coloring of any loopless connected graph which is not a block. Proposition 3.7. If G is a loopless graph of connectivity к =1, then xO(G) < 5. Proof. We may restrict to reduced graphs of connectivity к = 1. Suppose the statement is not valid, and let G be a minimal counterexample in terms of the number of blocks. By Propositions 3.1 and 3.3, G has odd order and is a connected reduced graph with an even orchid at every non-cutvertex. Observe that n(G) > 4 (for otherwise к = 1 implies m(G) < 4). Choose an end-block B of G, with s being the relevant cut-vertex, such that the graph G' := G - V(B — s) satisfies the inequality xO(G') < 5. Namely, if the graph G has more than two blocks then we merely choose B to be an arbitrary end-block of G : since G' is of connectivity 1 and has one block less than G, the choice of G assures that xO (G') < 5. And, if G has only two blocks, Proposition 3.6 and Fig. 4 assure that at least one of these two blocks will do: G cannot be the odd 4-edge-colorable graph depicted in Fig. 4. Observe that B is an even graph having odd order. Namely, every v g V (B)\{s} has an even orchid Sv. Hence, the same is true for the vertex s in B, proving that B is an even graph. Regarding the order, an even n(B) would imply that n(G') is also even and hence the inequalities xO(B) < 2 and xO(G') < 3 would yield the bound xO(G) < 5, which is a contradiction. So, B — s is an even graph of even order. Take an odd factor K of B — s and color the edges from B — s by using the color 1 for E(K ) and the color 2 for E(B — s — E (K )). Extend this to an edge coloring of B by using the color 1 for each edge incident to s in B. Use an odd 5-edge-coloring of G' with color set {1, 2,..., 5} such that the color 1 is used for at least one edge incident to s in G'. These two edge colorings of B and G' together constitute an odd 5-edge-coloring of G. This is in contradiction with the choice of G, which completes the proof of the statement. □ By now, we have assembled all the parts for a proof of the main theorem in this article. It characterizes the Shannon triangles of even type as the only loopless connected graphs needing the maximum 6 colors for an odd edge coloring. Theorem 3.8. For every loopless connected graph G, it holds that x'o(G) < 6. Equality is achieved only for the Shannon triangles of even type. Proof. Straightforward from Propositions 3.6 and 3.7. □ 4 Concluding remarks and further work Based on several additional observations, we propose three conjectures. Conjecture 4.1. For every loopless graph G, it holds that Xo(G)= xO(red(G)). (4.1) From Remark 1, we infer that Conjecture 4.1 is true whenever xO(red(G)) < 3 and Theorem 3.8 assures its validity whenever xO(red(G)) = 6. Furthermore, Propositions 3.4 and 3.5 imply that whenever red(G) is a necklace (open or closed), (4.1) stands. Namely, it is readily deduced from Remark 1 that if xO(red(G)) < 4 and every bouquet of red(G) has size 2, then the equality (4.1) is fulfilled. In regard to Proposition 3.6, we propose the following Conjecture 4.2. For every loopless 2-connected graph G whose reduction is neither the Shannon triangle of type (2,2,1) nor of type (2,2,2), it holds that xO(G) < 4. Furthermore, we believe that an even stronger statement is true. Namely, we propose that the bound in Theorem 3.8 could be further reduced, by excluding the graphs whose reductions are isomorphic to the Shannon triangle of type (2,2,1). Conjecture 4.3. For every loopless connected graph G whose reduction is neither the Shannon triangle of type (2,2,1) nor of type (2,2,2), it holds that xO(G) < 4. Assuming the validity of Conjecture 3, note that any possible counterexample G to Conjecture 4.1 must satisfy xO(G) = 3 and xO(red(G)) = 4. Another direction for further work is to provide an answer to the following open problem. Problem 4.4. Characterize the loopless graphs which are odd 2-edge-colorable. Acknowledgements. We thank the anonymous referees for helpful comments. This work is partially supported by ARRS Program P1-0383. References [1] J. A. Bondy and U. S. R. Murty, Graph Theory, Graduate Texts in mathematics, Springer, New York 244, 2008. [2] J. Czap, S. Jendrol', F. Kardoš and R. Sotäk, Facial parity edge colouring of plane pseudo-graphs, Disc. Math. 312 (2012), 2735-2740. [3] B. Lužar and R. Skrekovski, Improved bound on facial parity edge coloring, Discrete Math. 313 (2013), 2218-2222. [4] T. Mätrai, Covering the edges of a graph by three odd subgraphs, J. Graph Theory 53 (2006), 75-82. [5] L. Pyber, Covering the edges of a graph by..., Graphs and Numbers, Colloq. Math. Soc. Janos Bolyai 60 (1991), 583-610. [6] P. D. Seymour, Nowhere-zero 6-flows, J. Comb. Theory Ser. B 30 (1981), 130-135. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 279-286 Edge looseness of plane graphs Julius Czap Department of Applied Mathematics and Business Informatics, Faculty of Economics Technical University ofKosice, Nemcovej 32, 040 01 Kosice, Slovakia Received 7 February 2014, accepted 15 April 2015, published online 8 July 2015 A face of an edge colored plane graph is called e-loose if the number of colors used on its edges is at least three. The e-looseness of a plane graph G is the minimum positive integer k such that any edge coloring of G with k colors involves an e-loose face. In this paper we determine tight lower and upper bounds for the e-looseness of connected plane graphs. These bounds are expressed by linear polynomials of the number of faces. Keywords: Plane graph, edge coloring. Math. Subj. Class.: 05C10, 05C15 1 Introduction We use the standard terminology according to Bondy and Murty [1]. All considered graphs are undirected, finite, loopless, multiple edges are allowed. Let G = (V, E, F) denote a connected plane graph with the vertex set V, the edge set E and the face set F. We say that two edges of G are face-independent if they are not incident with the same face. Two edges of G are face-adjacent if they are consecutive edges of a facial trail of some face. The medial graph M (G) of G is the simple graph obtained as follows. For each edge e of G insert a vertex m(e) in M (G). Join two vertices of M (G) if the corresponding edges are face-adjacent in G. The embedding of G induces the embedding of M (G). Edge colorings of graphs embedded on surfaces with face-constrains have recently drawn a substantial amount of attention, see [5, 6, 7, 12] and references therein. There are two questions derived from hypergraph colorings that one may ask in this setting: Question 1.1. What is the minimum number of colors needed to color the edges of a connected plane graph in such a way that each of its faces is incident with edges of at least two different colors? E-mail address: julius.czap@tuke.sk (Julius Czap) Abstract Question 1.2. What is the maximum number of colors that can be used in an edge coloring of a connected plane graph in such a way that each of its faces is incident with edges of at most two different colors? The following result gives the answer to Question 1.1. Theorem 1.3. Every connected plane graph (on at least two edges) has an edge coloring with at most 3 colors such that each of its faces is incident with edges of at least two different colors. Moreover, if G is simple, then 2 colors are sufficient. Proof. First we show that every connected plane graph without faces of size two has an edge coloring with 2 colors such that these two colors appear on every face. Let G be such a plane graph and let M (G) be its medial graph. Add edges to M (G) to obtain a plane triangulation T. By the Four Color Theorem, T has a proper vertex coloring which uses 3 or 4 colors. Combine the first color class with the second, and combine the third with the fourth. This yields a 2-coloring of the graph T. It is easy to check that each face of T is incident with vertices of two different colors. This vertex coloring induces a required edge coloring of G. Now assume that the claim does not hold for a connected plane graph with faces of size two. Let G be a counterexample on minimum number of edges. Let e be an edge which is incident with faces f and g, where f has size two. Let G - e be the graph obtained from G by removing the edge e. The graph G - e is not a counterexample because it has fewer edges than G. Consequently, it has a required coloring. Let h be the face in G - e corresponding to the faces f and g in G. Extend the coloring of G - e to a coloring of G in the following way. If two colors appear on h, then color e with the third color. Otherwise (three colors appear on h) we color e with a color which does not appear on the second edge of f. This means that the minimum counterexample does not exist. There are graphs which require three colors for such a coloring, for example the connected graph on two vertices and three edges. □ In this paper, we focus on Question 1.2. A face of an edge colored plane graph is called monochromatic or bichromatic if the number of colors used on its edges is one or two, respectively. A face which is neither monochromatic nor bichromatic is called edge loose (or shortly e-loose). The edge looseness (or shortly e-looseness) of a plane graph G is the minimum positive integer k such that any surjective k-edge-coloring involves an e-loose face. This parameter of G will be denoted by els(G). The e-looseness is well defined for all plane graphs having at least one face incident with at least three different edges. (Note that every connected plane graph on at least four vertices has such a face.) Throughout the paper, we will consider only such graphs. 2 Upper bounds 2.1 1-connected plane graphs Theorem 2.1. Let G = (V, E, F) be a connected plane graph. Then els(G) < |F| + 2. Moreover, this bound is sharp. Proof. Let ( be an edge coloring of G such that each of its faces is either monochromatic or bichromatic. First we show that ( uses at most |F| + 1 colors. Clearly, we can assume that ( uses at least two colors. This means that at least one face is bichromatic, say f1. Let /2 be a face of G adjacent to /1. There is at most one color which appears on /2 and does not appear on /1, since the faces /1 and /2 share an edge and at most two colors occur on /2. The graph G is connected, therefore we can number the faces /1, /2,..., /|F | of G such that for every i > 2 the face / is adjacent to a face /j for some j < i. Using the argument described above, there is at most one color which appears on /j and does not appear on /ь ..., /j-1. Consequently, ^ uses at most 2 + (|F| - 1) colors. This means that every edge coloring with at least |F | + 2 colors involves an e-loose face. Now we prove that this upper bound is tight. Let H be a connected plane graph on at least three vertices. We insert a new vertex into each face of H and join each new vertex by an edge inside the corresponding face to exactly one of its vertices. In this way we obtain a 1-connected plane graph H. Clearly, we added |F(H)| edges to H and these new edges are face-independent in H. If we color these face-independent edges with distinct colors and all other edges with the same color, then we obtain an edge coloring of H such that every face is bichromatic. Hence, els(H) > |F(H)| + 2 = |F(H)| +2. □ 2.2 2-connected plane graphs When G is 2-connected, the bound from Theorem 2.1 can be improved by one. Let us recall that the (geometric) dual G* = (V*,E*,F*) of a plane graph G = (V, E, F) can be defined as follows: Corresponding to each face / of G there is a vertex /* of G*, and corresponding to each edge e of G there is an edge e* of G*; two vertices /* and g* are joined by the edge e* in G* if and only if their corresponding faces / and g are separated by the edge e in G (an edge separates the faces incident with it). Theorem 2.2. Let G = (V, E, F ) be a 2-connected plane graph. Then els(G) < |F | + 1. Moreover, this bound is sharp. Proof. Let ^ be an edge coloring of G such that every face is either monochromatic or bichromatic. This coloring induces a coloring of the dual G* in a natural way. Observe that at most 2 colors appear at any vertex of G*. Let us choose one edge from each color class, and let the chosen edges induce the subgraph H of G*. Each component of H is either a path or a cycle, since the graph G* does not contain any loop. Therefore, 2|E(H )| = (H) degH (v) < 2| V (H)| < 2|V (G*)| = 2|F (G)|. Since the number of colors used by ^ equals |E(H) |, we deduce that ^ uses at most |F(G) | colors. To see that the bound is tight consider the plane embedding of the complete bipartite graph K2,n, see Figure 1. This plane graph has n faces and an n-edge-coloring such that every face is bichromatic. □ Figure 1: A plane drawing of K2, 2.3 3-connected plane graphs Lemma 2.3. Let G = (V, E, F) be a 3-connected plane triangulation. Let t denote the maximum number of disjoint cycles in its dual. Then els(G) = 1 |F | + 1 +1. Proof. Any M2-edge coloring (being an edge coloring of a graph such that at most 2 colors appear at any vertex, see [2, 3]) of the dual graph G* corresponds to an edge coloring of G with property that every face of G is incident with edges of at most two different colors and vice versa. The graph G* is cubic, since G is a plane triangulation. In [3], it was proved that the maximum number of colors that can be used in an M2-edge coloring of a cubic plane graph H is equal to ЩНИ +1, where t is the maximum number of disjoint cycles in H. Consequently, els(G) = +1 + 1 = ^Gi +1 +1. □ Theorem 2.4. Let G = (V, E, F) be a 3-connected plane triangulation. Let g denote the girth of its dual. Then • els(G) < § |F | + 1 if g > 3, • els(G) < 4 |F | + 1 if g > 4, • els(G) < 10|F| + 1if g > 5. Moreover, these bounds are sharp. Proof. It follows from Lemma 2.3, since the number of disjoint cycles in the dual G* is not greater than |V(^}| = JFM. By Lemma 2.3 we can easily prove that the bounds are attained on the duals of graphs shown in Figure 2. □ Figure 2: The duals of these graphs show that the bounds are tight. Conjecture 2.5. Let G = (V, E, F) be a 3-connected plane graph. Then els(G) < § |F | + 1. 3 Lower bounds A matching of G is a set of pairwise disjoint edges, and a maximum matching is one of maximum cardinality. Let c(H) denote the number of components of a graph H. Lemma 3.1. Let G be a connected plane graph and let G* be its dual. Let M * be a matching in G*. Then els(G) > |M* | + c(G* - M*) + 1. Moreover, this bound is sharp. Proof. We color the edges of the matching M * with distinct colors and use an additional new color for the edges of each component of G* — M *. Thus, we obtain an edge coloring of the dual graph G* such that at most 2 colors appear at any vertex of G*. This coloring induces a coloring of G in which every face is either monochromatic or bichromatic. To see that the bound is sharp, let G be a plane graph whose dual is a 2r-sided prism. The 2r-sided prism H2r, r > 2, consists of the vertex set V = {ul, u2,..., u2r, vb v2,..., v2r } and the edge set E = {uiui+1, vivi+1, uiv^i = 1,..., 2r}, where 2r + 1 := 1. The set of faces consists of two 2r-gonal faces f1 = [ub ..., u2r] and f2 = [vb ..., v2r] and 2r quadrangles [uiui+1 vi+1vi] for i = 1,2,..., 2r, see Figure 3 for illustration. Figure 3: A 2r-sided prism. Let M = {uiui+1, vivi+1\i = 1,3, 5,..., 2r - 1} be a matching of H2r. Observe that \M\ = 2r and c(H2r — M) = r. Therefore els(G) > 3r +1. On the other hand, the graph H2r is a simple 3-connected cubic plane graph, hence its dual (the graph G) is a simple triangulation. By Theorem 2.4 we have els(G) < | • 4r +1. □ Corollary 3.2. Let G be a connected plane graph and let G* be its dual. Let M * be a maximum matching in G*. Then els(G) > \M*\ + 2. Proof. It immediately follows from Lemma 3.1, since c(G* — M*) > 1. □ Since there are 2-connected (and 1-connected) graphs G with arbitrarily many faces which have els(G) < 4 (Take a plane drawing of the cycle C = v1v2... v3n. Add n vertices u1, u2,..., un to the inner part of C and join ui with v3i-2 and v3i as it is depicted in Figure 4. It is easy to see that the e-looseness of the obtained graph is four.), there is no nontrivial lower bound on els(G) expressed by a linear polynomial of \F\ if G is not 3-connected. Hence, in the remaining part of the paper we will investigate e-looseness of 3-connected plane graphs G. Since the dual of G is a simple plane graph, we may apply structural properties of planar graphs on the dual graph; in particular, we will use the following one. V5 Figure 4: A graph G with "many" faces and els(G) = 4. Theorem 3.3. [11] Let G be an n-vertex 3-connected planar graph with n > 78. Let S denote the minimum degree of G and let M be a maximum matching in G. Then, • |M | > n+4 if S > 3, • |M| > ^n+s if S > 4, • |M| > ^n+o if S > 5. |V (G)l 2 Note that, if a planar graph G is 4-connected, then it has a matching of size This immediately follows from Tutte's result [14]: every 4-connected planar graph contains a Hamiltonian cycle. Lemma 3.4. Let G = (V,E, F ) bea 3-connected plane graph. Then it has at least ^ + 2 faces. Moreover, this bound is tight. Proof. The minimum degree of G is at least 3, since it is 3-connected. Using the hand deg(v) > 3|V|. Consequently, |E| > 2 shaking lemma we have 2|E| = ^2v(lV deg(v) > 3|V |. Consequently, |E | > | |V |. Using this fact and the Euler's polyhedral formula | V| — |E| + |F| = 2 we obtain |F| > 2 + lV. This bound is attained for every 3-connected cubic plane graph. □ Using Corollary 3.2, Theorem 3.3 and Lemma 3.4 we can prove the following. Theorem 3.5. Let G = (V, E, F) be a 3-connected plane graph on at least 152 vertices. Let g be the girth of G. Then, • els(G) > Щ+4 +2 if g > 3, • els(G) > ^Fi8 + 2 if g > 4, • els(G) > ^i20 +2 if g > 5. Moreover, these bounds are sharp. Proof. The 3-connectedness of G implies that its dual G* is also 3-connected, hence we can use Theorem 3.3 for G*. By Lemma 3.4, |V(G* )| = |F(G)| > + 2 > 78. It is easy to check that n+4 < ^ni8 < for n > 2. As g(G) < S(G*), the result follows from Corollary 3.2 and Theorem 3.3. The sharpness of these bounds follows from Theorem 3.6. □ Theorem 3.6. For any integer n > 4 there exists a 3-connected cubic plane graph G = (V, E, F) with girth g such that • g = 3, |F| = 3n — 4 and els(G) = n + 2, • g = 4, |F| = 7n — 12 and els(G) = 3n — 2, • g = 5, |F| = 19n — 36 and els(G) = 9n — 14. Proof. Let T be a simple plane triangulation on n > 4 vertices. Let Gj be the graph obtained from T by inserting the configuration Hi, shown in Figure 5, into each of its faces, for i = 3,4, 5. Any plane triangulation on n vertices has 2n — 4 faces, therefore |F (G*)| = |V (G3)| = n + 2n — 4 = 3n — 4, |F(G*)| = |V(G4) = n + 3(2n - 4) = 7n - 12 and IF(g;s)| = |V(g5)| = n + 9(2n - 4) = 19n - 36. In [2] it was proved that K2(G3) = n + 1, K2(G4) = 3n - 3 and K2(G5) = 9n - 15, where K2(Gj) denotes the maximum number of colors used in an M2-edge coloring of Gj, i = 3,4,5. Consequently, for the duals of these graphs it holds els(G*) = n + 2, els(G4) = 3n - 2 and els(G*) = 9n - 14. Observe, that each minimum edge-cut of size g in Gj corresponds to a cycle in G* and vice versa, therefore, the edge connectivity of the graph Gj is equal to the girth of the dual graph G*. The edge connectivity of Gj is equal to i, for i = 3,4, 5, since the tetrahedron, octahedron and icosahedron have edge connectivities 3,4 and 5, respectively. □ Finally, note that the vertex version of Question 1.2 was investigated in [4], where it was proved that the vertex looseness of a connected plane graph G equals the maximum number of vertex disjoint cycles in the dual graph G* increased by 2. Vertex looseness of triangulations on closed surfaces was studied in [8, 9, 10, 13]. References [1] J.A. Bondy and U.S.R. Murty, Graph theory, Springer, New York, 2008. [2] K. Budajova and J. Czap, M2-edge coloring and maximum matching of graphs, Int. J. Pure Appl. Math. 88 (2013), 161-167. [3] J. Czap, Mi-edge colorings of graphs, Appl. Math. Sciences 5 (2011), 2437-2442. [4] J. Czap, S. Jendrol', F. Kardos and J. Miskuf, Looseness of plane graphs, Graphs Combin. 27 (2011), 73-85. [5] S. Jendrol', J. Miskuf, R. Sotak and E. Skrabul'akova, Rainbow faces in edge-colored plane graphs, J. Graph Theory 62 (2009), 84-99. [6] B. Luzar, M. Mockovciakova, R. Sotäk, R. Skrekovski and P. Sugerek, ^-facial edge colorings of graphs, Discrete Appl. Math. 181 (2015), 193-200. [7] B. Luzar and R. Skrekovski, Improved bound on facial parity edge coloring, Discrete Math. 313 (2013), 2218-2222. [8] K. Matsuoka, Looseness width of 5-connected triangulations on the torus, Electron. Notes Discrete Math. 31 (2008), 105-108. [9] S. Negami, Looseness ranges of triangulations on closed surfaces, Discrete Math. 303 (2005), 167-174. [10] S. Negami and T. Midorikawa Loosely-tightness of triangulations of closed surfaces, Sci. Rep. Yokohama Nat. Univ., Sect. I, Math. Phys. Chem. 43 (1996), 25-41. [11] T. Nishizeki and I. Baybars, Lower bounds on the cardinality of the maximum matchings of planar graphs, Discrete Math. 28 (1979), 255-267. [12] J. Przybylo, On the facial Thue choice index via entropy compression, J. Graph Theory 77 (2014), 180-189. [13] T. Tanuma, One-loosely tight triangulations of closed surfaces, Yokohama Math. J. 47 (1999), 203-211. [14] W.T. Tutte, A theorem on planar graphs, Trans. Am. Math. Soc. 82 (1956), 99-116. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 287-300 Levels in bargraphs Aubrey Blecher, Charlotte Brennan *, Arnold Knopfmacher t University of the Witwatersrand, The John Knopfmacher Centre for Applicable Analysis and Number Theory, Johannesburg, South Africa Received 20 January 2014, accepted 8 February 2015, published online 8 July 2015 Abstract Bargraphs are lattice paths in N0, which start at the origin and terminate immediately upon return to the x-axis. The allowed steps are the up step (0,1), the down step (0, -1) and the horizontal step (1,0). The first step is an up step and the horizontal steps must all lie above the x-axis. An up step cannot follow a down step and vice versa. In this paper we consider levels, which are maximal sequences of two or more adjacent horizontal steps. We find the generating functions that count the total number of levels, the leftmost x-coordinate and the height of the first level and obtain the generating function for the mean of these parameters. Finally, we obtain the asymptotics of these means as the length of the path tends to infinity. Keywords: Bargraphs, levels, generating functions, asymptotics. Math. Subj. Class.: 05A15, 05A16 1 Introduction Bargraphs are lattice paths in N0, starting at the origin and ending upon first return to the x-axis. The allowed steps are the up step, u = (0,1), the down step, d = (0, -1) and the horizontal step, h = (1,0). The first step has to be an up step and the horizontal steps must all lie above the x-axis. An up step cannot follow a down step and vice versa. It is clear that the number of down steps must equal the number of up steps. Related lattice paths such as Dyck paths and Motzkin paths have been studied extensively (see [4, 9]) whereas until now bargraphs which are fundamental combinatorial structures, have not attracted the same amount of interest. * Corresponding author. This material is based upon work supported by the National Research Foundation under grant number 86329 t This material is based upon work supported by the National Research Foundation under grant number 81021 E-mail addresses: Aubrey.Blecher@wits.ac.za (Aubrey Blecher), Charlotte.Brennan@wits.ac.za (Charlotte Brennan), Arnold.Knopfmacher@wits.ac.za (Arnold Knopfmacher) Bousquet-Melou and Rechnitzer in [2] and Geraschenko in [8] have studied bargraphs which were named skylines in the latter, and wall polyominoes as per the study of Feretic, in [6]. Bargraphs models arise frequently in statistical physics, see for example [3, 5, 10, 12, 15, 17]. In addition, bargraphs are commonly used in probability theory to represent frequency diagrams and are also related to compositions of integers [11]. In this paper, we consider levels, which are maximal sequences of two or more adjacent horizontal steps. We find different generating functions in each of the following sections where x counts the horizontal steps, y counts the up vertical steps and w counts one of the following parameters: the total number of levels and the horizontal position or the height of the first level. To facilitate these computations, we also find the generating function for paths with no levels. The study of levels in bargraphs is related to the modelling of tethered polymers under pulling forces, see [13, 14]. These pulling forces have vertical and horizontal components and tend to be resisted by what is known as the stiffness of the polymers. The polymers undergo phase changes, called the stretched (adsorption) phase, where the polymer is stretched vertically. The free (desorbed) phase occurs only when the vertical force is zero. In the bargraph models of polymers positive or negative energy is added to points in levels on the bargraph (called stiffness sites), they tend to keep the polymer horizontal or cause it to bend. As an example of a bargraph we have Figure 1: A bargraph with 12 up steps, 13 horizontal steps and 4 levels Often in the lattice walk and polygon literature, "bargraphs" refer to polygon structures (which would be obtained from the objects considered here by joining the first and last vertices with horizontal steps). The objects discussed here are sometimes called "partially directed walks above a wall" depending on the context (in polymer modelling work for example). The main tool for elucidating the statistics of interest in this study is a decomposition of bargraphs which is based on the first return to level one. This was described initially by Prellberg and Brak in [16] and more recently in [2], where it is called the wasp-waist decomposition. The present authors have also discussed it in [1]. It follows from the wasp-waist decomposition that the generating function B(x,y) which counts all bargraphs is 6. 1 2 3 4 5 6 7 8 9 1o 11121з 14 B := B(x, y) 1 — x — y — xy — \J (1 — x — y — xy)2 — 4x2y (1.1) 2x □-□•cO О Figure 2: Wasp-waist decomposition of bargraphs Here x counts the number of horizontal steps and y counts the number of up steps (see Theorem 1 in [1]) or [2, 7]). The series expansion, B(x, y) begins x(y + y2 + y3 + y4) + x2(y + 3y2 + 5y3 + 7y4) + x3(y + 6 y2 + 16y3 + 31y4) + x4(y + 10y2 + 40y3 + 105y4 + 219y4). The bold coefficient of x4y2 is illustrated below with the full set of 10 bargraphs with 4 horizontal steps and 2 vertical up steps. + + + 1 2 3 4 5 In [1, 2] the authors found an asymptotic expression for B(z, z), where z marks the semiperimeter of the bargraphs. This is known as the generating function for the isotropic case. The dominant singularity p is the positive root of D := 1 - 4z + 2z2 + z4 = 0, (1.2) given by 1 / 4 x 22/3 3 V - (13 + 3v33)1/3 We have B(z, z)--(1 - p)1/2 as z ^ p. Hence 1 / 4X 22/3 / ,_\ i/3\ p = 1 (-1 - (13 + 3vW/3 + ('2<13 + :W5>) ) =».295598-. (13) - p - 3 [zn]B(z, z) ~ v JL^ p-n. (1.4) 2 у n pn3 The following definitions will be used: A level in a bargraph is a maximal sequence of two or more adjacent horizontal steps denoted by hr where r > 2. It is preceded and followed by either an up step or a down step. The length of the level is the number r of horizontal steps in the sequence. The height of a level is the y-coordinate of the horizontal steps in the sequence. Thus, the graph in Figure 1 has four levels, three of length 2 and one of length 3. In all the generating functions of the following sections, the horizontal steps are counted by x, the vertical up steps are counted by y and the parameter that is under investigation by w. In each section, we use G(x, y, w) or F(x, y, w) for the generating function where the definition of G or F applies only to the section under consideration. 2 Total number of levels 2.1 Generating function for the number of levels A level is a sequence of two or more adjacent horizontal steps as defined in the previous section. Let F(x,y,w) be the generating function where w marks the total number of levels. Using the wasp-waist decomposition in Figure 2, we have F := F(x,y,w)= ^ + ^ + ■yF^ + xyF + FF2 (2.1) 1 2 3 4 5 The numbers below the terms refer to the cases in the wasp-waist decomposition. This will be done throughout the paper. The generating function F2 := F2 (x, y, w) is the analogous function restricted to case 2. We use the following symbolic decomposition for F2 tQ 4P) F ' m пП F R Figure 4: Decomposition for F2 where FR is the generating function for bargraphs in which the first column is of height 2 or more. The function FR is easily obtained by considering all bargraphs except those starting with a column of height one. Thus FR = F - xy - F2. (2.2) From Figure 4, we get 2R F2 = x(F — xy — xF ) + wx y + wx F (2.3) + + So, combining equations (2.1), (2.2) and (2.3), we find 1 ( 2 2 F -^-^ 1 - x - y - xy + 2xy - 2wx y- 2(x - x2 + wx2) у J4(-x + x2 - wx2)(xy - x2y + wx2y) + (1 - x - y - xy + 2x2y - 2wx2y)2^ . (2.4) In order to find the generating function for the total number of levels in bargraphs, we differentiate F with respect to w and then put w = 1 to obtain dF (1 - x)(1 - у) (1 - x - y - xy - J(1 - x - y - xy)2 - 4x2yj FLevels • dw 2^/(1 - x - y - xy)2 - 4x2y w=1 - x - У where z marks the semiperimeter. The series expansion begins x2 (y + y2 + y3 + y4) + x3(y + 5y2 + 9y3 + 13y4) + x4(y + 12y2 + 38y3 + 79y4). There are in total 12 levels in our example in Figure 3. This is shown in bold in the series expansion. 2.2 Asymptotics in the isotropic case We consider bargraphs with respect to the semiperimeter by substituting z for x and y in F to obtain (1 - z)2(1 - 2z - z2 - v1 - 4z + 2z2 + z4) (1 - z 2 2 FLevels (z? z) V1 - 4z + 2z2 + z4 ' In order to compute the asymptotics for the coefficients, we use singularity analysis as described in [7]. Let p be as in (1.2) and (1.3). We find that as z —у p 1 - 4p + 4p2 - p4 F Levels 4 Jp(1 - P - P' By singularity analysis we have rzni F 1 - 4p + 4p2 - p4 [z JFLevels ~ --p . Vnn^Jp (1 - p - p3) Then after dividing by the asymptotic expression for the total number of bargraphs found in (1.4), we get the following result: Theorem 2.1. The average number of levels in bargraphs of semiperimeter n is asymptotic to 1 - 4p + 4p2 - p4 2(1 - p - p3) as n —>• oo where C = 0.117516 • • • . ■ n = C n, 3 Bargraphs with no levels 3.1 Generating function for the number of graphs with no level Because we require it later, we begin by enumerating a special class of bargraphs, namely one in which an adjacent sequence of horizontal steps does not occur (i.e. the only sequences of horizontals are single). This is denoted by Fo := F(x, y, 0) where F is the generating function (2.4) from the previous section. We use the wasp-waist decomposition in Figure 2 to obtain Fo = ^ + Fo^ + yFb + yFox; + Fo^. (3.1) 1 2 3 4 5 Case 2 is explained below in Figure 5. □ □ - □ Figure 5: Explanation for case 2, decomposition of F0 2 Thus which leads to Fo,2 = x(Fo - xy - Fo,2), Fo o,2 x Fo - xy 1 + x (3.2) The exclusions in case 2 are because we are not allowing adjacent horizontal steps. Hence, from (3.1) and (3.2), we have: x(Fo - xy) Fox(Fo - xy) Fo = xy +--—--+ yFo + yFox + - 1 + x 1 + x 1 Solving this for Fo, we obtain Fo 1 - y - 2xy - /1 - y^J 1 - y - 4xy - 4x2y 2x The series expansion for Fo begins (3.3) x(y + y2 + y3 + y4) + x2(2y2 + 4y3 + 6y4) + x3(y2 + 7y3 + 18y4) + x4 (6y3 + 32y4 + 92y5). Our example in Figure 3, shows that indeed there are no bargraphs having 4 horizontal and 2 up steps and no levels, which is confirmed by the lack of x4y2 term. 3.2 Asymptotics in the isotropic case As before we substitute z for x and y in F0 and obtain N 1 - z - 2z2 - ^/T—z /1 - z - 4z2 - 4 z3 Fo(z,z) =-2z-• Let t be the dominant root of 1 - z - 4z2 - 4z3 = 0, its value is t = -1 (-4 + (224 - 24/87)1/3 + 2(28 + 3/87)1/3) = 0.34781 • • • . Using singularity analysis we have as z — t „ , , /1-7/т(1 + 8t + 12t2)t fo(z,z)---2T-• Extracting coefficients will yield the asymptotic number of bargraphs with no levels. /Г-Г^т (1 + 8t + 12t 2) _„ [zn]Fo(z, z) ~ 4л/n n3 as n —у то. For n = 100, there are 3.20775 x 1042 bargraphs whereas the asymptotics give 3.24376 x 1042. 4 Horizontal position of the first level 4.1 Generating function for the mean Now we derive a generating function Gx for bargraphs in which the leftmost x-coordinate of the first level is counted by w. In the case where the bargraph has no level, we define the horizontal position to be 0. In Figure 6, the start of the first level is the point with coordinates (2, 5) and therefore the x-coordinate of the start of the first level here is 2. By the wasp-waist decomposition we have Gx = ^xy^ + JF^ + yG + yGxX + 1 2 3 4 5 (4.1) To calculate the generating function for case 2, we use Figure 7 below. The part labelled L in Figure 7 indicates a bargraph with at least one level. L □ Q L m nQ Figure 7: Decomposition for FL,2 + + Note that FL,2 is the generating function for case 2, (paths which have at least 1 level). The generating function for the graph labelled "a" in Figure 7 is therefore Gx - F0, since F0 is the generating function for graphs with no levels from Section 3. Thus, using Figure 7, we have: fl,2 := fl,2(x, y, w) = wx(Gx - Fo - Fl,2) + x2y + x2B where B is the generating function for all bargraphs from equation (1.1). Hence, wxGx - wxFo + x2y + x2B FL,2 = --, (4.2) 1 + wx and from (3.2) I . l-y-2xy+Vl-W 1-y-4xy-4x2y x -xy +--2x- F0.2 = —— (Fo - xy) = 1 + x 1 + x So, for case 2 F2 = Fl,2 + Fo,2. Thus finally, the decomposition for case 5 requires Figure 8 below: D I a |_в_i Figure 8: Case 5 For case 5, we have the concatenation of two bargraphs labelled a and ß. There are three cases depending on whether the graphs a and ß have levels or not. i. Graph a has levels with generating function y(Gx -F0), in which case the generating function for в is —. 1 y ii. Neither graph has levels, thus the generating function is F0F0,2 where F0,2 is as in (3.2) or iii. Graph a has no levels but graph в has, so the generating function is F0(xw, y)F2 where F0(w) := F0(xw, y) indicates that x has been replaced by xw in F0(x, y). Thus Gx = ^ + ^ + yG* + yGxx + ((Gx - Fo)xB + F0F0,2 + Fo(xw, y^) 1 2 3 4 5 where in all but one case, the parameters have been omitted. We solve for Gx, leading to Gx (x,y,w) Bx2F0 +1 + BxFo - -Bx+T - F0F0 2 - x2yF+(1w) + wx(f°1)2 + - Fo 2 - wx+1 1 0 — X+1 0 0,2 -x+1 1 — X+1 1 — X + 1 0,2 -x + 1 xy where Fo(w) = Bx + -xF0(-) I -x I xy I y _ 1 Bx + -x + 1 + -x + 1 + xy + y 1 1 — y — 2 wxy — v1 — y\/1 — у — 4 wxy — 4 w2x2y 2wx (4.3) Remark: We note that from (3.3) F0(w)|w=1 = F0. Now, in order to find the mean horizontal position, we calculate: dGx w=1 dw (Bx(x + 1) + F0x + (x + 1 )2y - 1)2 Fr x {Fo((x + 1)Fq + Fo + 1)(Fo,2 + xy + y - 1) + Fq^x + ад + Fq + 1) +B2x2(x(-Fq + Fq + 1) - Fq) + Bx(2x2y(-F' + Fq + 1) - (y - 1)F^ +x(F02 - 3yFg + Fq + Foy + Fq + y)) + xy(x2y(-F' + Fq + 1) - yF^ + 2F^ +x(F02 - 2yFg + 2F^ + Foy + Fq + y) + Fq + 1)} where dFo (w) о dw y (-2xVy-T + J(2x + 1)2y - 1 - Vy-r) - J(2x +1)2y - 1 + Vv-1 2x^J(2x + 1)2y - 1 ' (4.4) The series expansion of dw-1 = , begins x dw I w =1 3 (2y2 + 4y3 + 6y4) + x4 (5y2 + 25y3 + 60y4) . In our example in Figure 3, the sum of the horizontal positions of the first levels is 5. 4.2 Asymptotics in the isotropic case Using singularity analysis and computer algebra we find that dGx dw where p is as in (1.3) and 1 - P -2 ci(p)Jp (1 - p - p3)(^1 - p) \ 1/2 z 4 c1(p) ((-1 + P)P2 + v-TTPJY(P)) Jy(p) x (J-1+ P(1 - P - 12p2 - 4p3 + 13p4 + 27p5 + 18p6 + 18p7 + 4p8) +(-1 + p + 4p2 + 8p3 - 5p4 + p5 - 6p6 - 2p7) JY^) as z ^ p and Y(p) = -1 + p + 4p2 + 4p3 The coefficient is [zn]dGx c1(pWp (1 - p - p3) p-n w=1 Vnn3 dw After dividing by the asymptotic number of bargraphs we get x Theorem 4.1. The average horizontal position of the first level in bargraphs is asymptotic to the constant 2 p ci (p) = 2.38298, as n ^ то. For n = 200, the exact average is 2.35787 • • •. 5 Height of the first level 5.1 Generating function for the mean Let Gy (x, y, w) be the generating function for the y-coordinate of the first level for bar-graphs where w marks this coordinate. If there are no levels then there is no w, so we have a contribution to w0. As in the previous section, the first level in Figure 9 begins at the point (2, 5), with y-coordinate 5. Height of first level 7 6 5 4 3' 2 1+ ru 1 2 3 4 5 6 7 8 9 10 11 12 13 14 Figure 9: Height of the first level Using the wasp-waist decomposition, this yields: Gy = xy^ + ^ + xF3 + (5.1) 1 2 3 4 5 Considering case 2 separately, we have for F2 : tO. □(□<,-£))+ш + пО Figure 10: Decomposition for F2 Thus F2 = x(Gy — xy — F2) + x2yw + x2wB. So p x(Gy — xy + xyw + xwB) ^ 2 1 + x and F3 = yw(Gy — Fq) + yFo (5.3) where the first and second terms distinguish between the cases where there are levels (which are therefore multiplied by w) and no levels. Also separately, for the last case F5 we can use Figure 8. If a has levels, then the generating functions for a and ß are w(Gy — F0) and xB(x, y) respectively. On the other hand, if a has no levels, the generating functions are yF0 and F2. Thus F5 = w(Gy — Fo)xB(x, y) + yFoF2. (5.4) Substituting (5.2), (5.3), and (5.4) in (5.1) and solving for Gy, we obtain T y = Bwx + Fx + w(x + 1)y + ^ — 1 (5.5) where BF0wx2 Bwx2 F0wx2y , . T =--—--+ BFowx--—---—--+ Fo w(x + 1)y x +1 x +1 x + 1 Fox2y wx2y x2y + —л" — Fo(x + 1)y--tt + т~г — xy. x + 1 x + 1 x + 1 The generating function for the sum of the heights of the first levels is obtained from the derivative of Gy with respect to w and then setting w = 1. Using the following substitutions f X (x,y) = —1 + (1 + 2x)2y, (56) \ Y(x, y) = ( — 1 + y)( —1 + x2( —1 + y) + y + 2x(1 + y)), ( . ) we have dGy dw (-1 + x + y - xy + JY(x,y)) (x2(1 - y) + xjY(x,y) - JX(x,y)Vy-T + JY(x,y)) X (4x2(y - 1)y + x (-2 jX(x,y)Jy - 1 y + JX(x,y)Jy - 1 + 4y2 - 3y - l) +y (-JX (x,y)jy-1 + y - 1)) . (5.7) The series expansion of ^wf- \w=i begins x2 (y + 2y2 + 3y3 + 4y4) +x3 (y + 8y2 + 21y3 + 40y4) + x4 (y + 15y2 + 71y3 + 198y4) . Figure 3 illustrates that the sum of the heights of the first levels is 15 as shown in bold above. 5.2 Asymptotics in the isotropic case Substituting z for both x and y in the above equation (5.7) and using X(z, z) := X(z) = -1 + z + 4z2 + 4z3 and Y(z, z) := Y(z) = 1 - 4z + 2z2 + z4, we obtain dGy dw (-1 + 2z - z2 + /Y(ž)) ((1 - z)z2 - yz-IyX(Z) + /Y(Z) + z/Y(Ž)) X |^4z3(z - 1) + z (-1 + z - Vz - 1 /X(z)) +z (-1 - 3z + 4z2 + Vz - 1 /X(z) - 2zVz - 1 /ХЩ)] --2 c2(p)/p(1 - p - P3^1 - p) / by using computer algebra as z ^ p, where (-2 + 2p + p2 - p3 + У-Г+P/Xp)) (1 + P - 2p3 + pV-1TPу/ХЩ c2 (P) = 2p (p2(-1 + p) + v-1+p /xjp)3 Hence <*Шp(1 - p - ^p-n as n [z nl Щ1 [z ] dw where w=1 ynn \fn n3 Thus after dividing by the asymptotic number of bargraphs we obtain Theorem 5.1. The average height of the first level in bargraphs is asymptotic to the constant 2 pc2(p) « 6.15883 ••• , as n ^ <. For n = 300, the exact average is 6.00066 • • •. References [1] A. Blecher, C. Brennan and A. Knopfmacher, Combinatorial parameters in bargraphs (2015), submitted. [2] M. Bousquet-Melou and A. Rechnitzer, The site perimeter of bargraphs, Adv. Appl. Math. 31 (2003), 86-112. [3] M. Bousquet-Melou and R. Brak, Exactly solved models of polyominoes and polygons, Chapter 3 of Polygons, Polyominoes and Polycubes, Lecture notes in physics, volume 775, Springer, Berlin, Heidelberg, 2009, 43-78. [4] E. Deutsch, Dyck path enumeration, Discrete Math. 204 1-3 (1999), 167-202. [5] P. Duchon, q-grammars and wall polyominoes, Ann. Comb. 3 (1999), 311-321. w = 1 [6] S. Feretić, A perimeter enumeration of column-convex polyominoes, Discrete Math. Theor. Comput. Sci. 9 (2007), 57-84. [7] P. Flajolet and R. Sedgewick, Analytic Combinatorics, Cambridge University Press, 2009. [8] A. Geraschenko, An investigation of skyline polynomials. http://people.brandeis. edu/~gessel/4 7a/geraschenko.pdf [9] K. Humphreys, A history and a survey of lattice path enumeration, J. Statist. Plann. Inference 140:8 (2010), 2237-2254. [10] E. J. Janse van Rensburg and P. Rechnitzer, Exchange symmetries in Motzkin path and bargraph models of copolymer adsorption. Electron. J. Comb. 9 (2002), R20. [11] D. Merlini, F. Uncini and M. C. Verri, A unified approach to the study of general and palindromic compositions, Integers 4 (2004), #A23. [12] J. Osborn and T. Prellberg, Forcing adsorption of a tethered polymer by pulling, J. Stat. Mech-TheoryE. (2010), P09018. [13] A. Owczarek, Exact solution for semi-flexible partially directed walks at an adsorbing wall, J. Stat. Mech.: Theor. and Exp. (2009), P11002. [14] A. Owczarek, Effect of stiffness on the pulling of an adsorbing polymer from a wall: an exact solution of apartially directed walk model, J. Phys. A: Math. Theor. 43 (2010), 225002 (16pp). [15] A. Owczarek and T. Prellberg, Exact Solution of the Discrete (1+1)-dimensional SOS Model with Field and Surface Interactions, J. Stat. Phys. 70:5/6 (1993), 1175-1194. [16] T. Prellberg and R. Brak, Critical exponents from nonlinear functional equations for partially directed cluster models. J. Stat. Phys. 78 (1995), 701-730. [17] C. Richard, I. Jensen and A. J. Guttmann, Scaling Function for Self-Avoiding Polygons, Proceedings TH2002 Supplement, Birkhauser Verlag, Basel, 2003, 267-277. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 301-320 What can Wikipedia and Google tell us about stock prices under different market regimes? In less than five years a surprisingly high level of attention has built up in the possible connection between internet search data and stock prices. It is the main aim of this paper to point out how this connection may depend heavily on different regimes of the market, i.e. the bear market vs. the bull market. We consider three types of internet search data (relative Google search frequencies of company tickers, relative Google search frequencies of company names and page visits of Wikipedia articles about individual companies) and a substantial sample of companies which are members of the S&P 500 index. We discover two inverse patterns in stock prices: in the bear market what we propose to term a "merry frown" and in bull market a "sour smile", both clearly seen especially for the Wikipedia data. We propose market neutral strategies that exploit these new patterns and yield up to 17% in average annual return during our sample period from 2008 to 2013. Keywords: Stock returns, internet search data, market regimes, trading strategies. Math. Subj. Class.: 91G70, 97K80 1 Introduction A byproduct of the increasingly widespread use of the internet is the data on internet activity of individual users. While most of this data is retained by the website owners and unavailable to the public either due to privacy or business reasons there are some exceptions. One such example is the Google Trends service which enables users to view the * Operation part financed by the European Union, European Social Fund. t Supported by a grant from the Slovenian Research Agency - ARRS. E-mail addresses: boris.cergol@xlab.si (Boris Cergol), matjaz@omladic.net (Matjaž Omladic) Boris Cergol * XLAB, Ljubljana, Slovenia Matjaž Omladic f Department for automation, biocybernetics and robotics, Jozef Stefan Institute, Ljubljana * Received 21 October 2013, accepted 19 December 2013, published online 10 July 2015 Abstract relative frequencies of search queries entered into Google's search engine. Since becoming publicly available in 2006 Google Trends have attracted attention of researchers in various fields. In [12] the authors show that analysis of health-related search queries can lead to accurate estimates of influenza epidemics with a reporting lag of only one day which is almost two weeks sooner than traditional surveillance systems. Choi and Varian [6] apply a similar approach to estimating a number of economic indicators such as automobile sales or unemployment claims. The relevance of internet search data for financial data analysis was first explored by Da, Engelberg and Gao [6] who considered the relative search frequencies of company tickers and names as proxies for investor attention in the US stock market. They show that search frequencies outperform existing measures of investor attention and that an increase in a company ticker's search frequency predicts a higher stock price in the following two weeks. In [2, 14] the authors obtain similar results in terms of future returns and additionally observe that an increase in a company name's search frequency is associated with a rise in trading activity and stock liquidity. The prevailing explanation for positive correlation between future stock returns and company-related relative search frequencies is based on the theory of Barber and Odean [3]. They suggest that attention-grabbing stocks experience short-term buying pressure from individual investors. This might simply be due to the fact that a single investor faces a difficult decision when deciding which of the thousands of available stocks to buy, while the decision of which stock to sell is much easier since it is usually limited to the few stocks that are part of his existing portfolio. Google trends data has also been used in assessing investor sentiment. In [8] the authors construct the index which is a sum of relative search frequencies of economy-related terms associated with negative sentiment. This new index is able to predict values of existing investor sentiment indicators and has a perceptible impact on short-term future stock prices. In [18] a number of stock market index strategies are tested that profit from fluctuations of relative search frequencies of individual economy-related terms. Strategies of the same type are further explored in a related work [17] where Google trends data is replaced by the numbers of page visits to economy-related Wikipedia pages. The main contribution of our paper is the addition of market regimes into the study of the connection between stock returns and internet search. If the reason for positive correlation between future returns and search frequencies is in fact in the cognitive bias of individual investors then we would expect that the effect would be even stronger in periods when investors face greater uncertainty and are even more prone to irrational decisions. We present a two-state hidden Markov model for the returns of the S&P 500 index. The model parameters are estimated by the Baum-Welch algorithm after which the most likely sequence of hidden states is found by the Viterbi algorithm. The first of the two states is characterized by low returns and high volatility and corresponds to what is commonly refered to as the "bear market" regime by investors. Conversely, the second state is characterized by high returns and low volatility and we label it the "bull market" regime. We choose a sample of stocks that are members of the S&P 500 index and study the relation between their future short term returns and three different types of internet search data: the page visits of company-related articles on Wikipedia, the relative frequency of Google searches for company tickers and the relative frequency of Google searches for company names. To the best of our knowledge ours is the first study of this kind that takes Wikipedia data for individual companies into consideration. We also perform our analysis on daily data which is in contrast to most of the existing literature where financial applications of internet search is studied using weekly data. We perform a number of cross-sectional Fama-MacBeth regressions where future stock returns are the explained variable and a single internet search variable is the regressor. This regressions are performed on a subsample of observations that belong to either the bear regime or the bull regime as well as on the entire sample. Our main result is that the market regime indeed has a strong influence on the relation between future stock returns and internet search data. In all three cases of internet search variables the future returns are higher in the bear regime compared to the bull regime given the same increase of the chosen internet search variable. This effect remains evident even after controlling for the factors of the Carhart four-factor model [5]. After controlling for the Carhart factors the Wikipedia page visits variable emerges as the one with the greatest influence on future stock returns. In fact, both of the Google search variables prove to be statistically insignificant. To our surprise, we also find very little evidence supporting the theory that an increase of investor attention to a given stock translates into a short term rise of the stock's price due to increased buying pressure. Instead we observe two different price patterns for which we propose the terms "merry frown" and "sour smile". A merry frown is a pattern of positive correlation between future stock returns and Wikipedia page visits that is observed only during the bear market. A sour smile is a pattern of negative correlation between future stock returns and Wikipedia page visits that is observed only during the bull market. Both patterns might be explained as a corrective investor counter-reaction to initial overpessimism in the bear market and to initial overoptimism in the bull market. Economic significance of the merry frown and the sour smile is explored by constructing a market neutral strategy with long positions in stocks that are in the highest decile and short positions in stocks that are in the lowest decile with regard to Wikipedia page visits during the bear market. In the bull market the positions are reversed. We backtest the strategy for different trading frequencies (from 1 to 10 days) and observe that they generate positive returns which decrease with the length of trading frequency. The returns of the strategies are compared to random market neutral strategies generated by a Monte Carlo simulation and their statistical significance is established. We also find that returns of the strategies strongly increase if they are restricted to a subsample of stocks that are preferred by individual investors such as high volatility stocks, low market capitalization stocks or low price-to-book ratio stocks. In the best case, a trading strategy with daily trading frequency that is restricted to stocks with higher than median volatility yields an average annual return of 17% in our sample period. The paper is organized as follows: Section 2 describes the data. The market regime model is presented in Section 3. In Section 4 we discuss the results of the Fama-MacBeth regressions. The trading strategies and backtest results are presented in Section 5. Section 6 concludes the paper. 2 Data description The most important choice in the beginning of every statistical research is the choice of statistical population. We decided to limit our study to a sample of stocks that are included in the index S&P 500. More precisely, the stocks that were members of this index on June 7, 2013. Our choice is primarily motivated by the fact that the publicly accessible and freely available data on individual stocks is of highest quality for stocks listed on the US stock market. Of course, the extension of our study to stocks listed on some less common stock markets remains a challenge for future investigation. We choose a sampling period from October 1, 2007, up to June 30, 2013. We restrict ourselves to this specific sampling period primarily because of the availability of website search data. Fortunately, this period includes very diverse market conditions including one of the greatest market crashes in history and the following rebounding growth. This gives us confidence that our findings would easily extend to future periods. For every stock in our sample we obtain the daily dividend and split adjusted closing prices from the Yahoo Finance website1. We additionally remove all stocks for which data is not available for the entire sample period. This mostly includes stocks that were members of the S&P 500 index on June 7, 2013, but were not yet publicly traded at the begining of our sample period. We get the daily closing values of the index S&P 500 in the economic data section of the website of Federal Reserve Bank of St. Louis2. This data is obtained for a longer period from January 1, 2000, up to June 30, 2013, as required by the regime-switching model described in Section 3. There are various choices for website search data that are worth testing for possible relations with fluctuations in stock prices. One of the possibilities is Google search data which is publicly available via the Google Trends Service3. In related studies, authors most commonly use relative search frequencies of stock tickers [7, 14] while some also consider relative search frequencies of company names [2]. The majority of studies rely on weekly data for these frequencies. This might be based on availability problems with Google data. When one requests a search frequency time series for a certain term the format of the returned series depends on the length of the period. For periods no longer than 3 months one gets the daily data, but for longer periods only weekly data is returned. An additional feature of the data so obtained is that it is normalized within the series to have a maximum value of 100. This has some advantages but also makes it difficult to compare values of series in different periods. This may have been the reason for most authors to restrict their studies to weekly data. In order to overcome this difficulty, we acquire the three-month-period data every two months. From the data for the overlapping month we compute the quotient between the normalized factors of the two consecutive periods thus enabling us to concatenate the short period time series into one long period time series with the daily data. An additional problem arises with company names. Namely, one would need to know which name people are using for the company when searching for information about it. For instance, it is unlikely that most people would search for American Express Company by typing its full name into the search window but would instead just type American Express or simply AmEx. Accordingly, we replace company names in our sample with suitable abbreviations. By following the concatenating procedure described above we obtain a daily time series of relative search frequencies for both a company ticker and an abbreviated company name for each company in our sample. The sample period for this data is chosen to be from January 1,2008, up to May 31, 2013. Another source of internet search data that has recently been studied related to financial data is Wikipedia. For every article on Wikipedia, a time series of daily unique page visits 1 http://www.finance.yahoo.com 2http://research.stlouisfed.org/fred2 3http://www.google.com.au/trends can be obtained from the website Stats.Grok.Se4. This source of information has not gained as much attention as the Google Trends Data and the reason for this may lie in the fact that the available time series only span from January 2008. For every company in our sample, we find the Wikipedia article associated to the company and obtain a time series of unique daily page visits to this article in a sample period from January 1, 2008, up to May 31, 2013. There are some comments we have to make that relate to data preprocessing of both Google search and Wikipedia page visits data. We first note that this data is not available throughout the chosen period for each company in our sampling period. Therefore, we have to exclude the companies with too much missing data. Our rule is to allow no more than 10% of missing data, while for the missing values we apply an imputation procedure that takes into account the weekly seasonality. Next we perform a detrendization of data using a sort of "longitudinal normalization", i.e. we divide the number for a given day by the average of the numbers for the last 56 days. The choice of length for this normalization period is similar to choices made by authors in related studies, for example in [7] where the length of the normalization period is 8 weeks. We also have to take care of the outliers. Their influence is reduced by taking a logarithm transformation of our data. Additionally, for each stock in our sample and each of the three variables, we perform a winsorization of the corresponding time series by limiting the range of data to its first percentile from below and to its 99th percentile from above. There are strong seasonal effects on the weekly basis in both Google and Wikipedia data. We want to make data collected on different days of the week comparable by introducing a seasonal adjustment in the following way. We regress the data to the days-of-the-week (but one) as dummy variables to get average differences between the different days of the week which we add to the data of this particular day. 3 Market regimes We intend to study the influence of market regimes on the relation between internet search data and stock returns. A market regime may be considered as a phase of persistent attributes observed in financial time series. This concept is most commonly used by investors when classifying the market into two phases: the bear market characterized by low returns and high volatility and the bull market characterized by high returns and low volatility. This dichotomous approach also lies at the core of our view on market dynamics. A systematic regime-switching time series model was first proposed by Hamilton [13] and the variants of this model are still being overwhelmingly used to study the regimes of economic and financial time series. However, recently, the problem of parameter estimation in financial regime-switching models has also been tackled by the Baum-Welch algorithm [19] which has previously been mostly used in engineering applications. We also decided to base our market regime estimation model on the Baum-Welch algorithm and refer the reader to [15] for a discussion of its advantages over Hamilton's algorithm. We will determine the actual switching of the regimes only on the cumulative level and base it on the returns data of the S&P 500 index. The background of the model is a hidden Markov chain (denoted by Q = (qt)f=1, where T is the length of the period) so that the market regimes are seen as states of this chain, the bear market becomes say state i = 1 and the bull market state i = 2. As a consequence the model has a 2 x 2 4http://stats.grok.se transition matrix A defined in the usual way: aij = p(qt+1 = j | qt = i). Given an initial distribution of the states П = (n1, n2 ) we then have the Markov chain uniquely determined. Furthermore, we assume that the observations form a random sequence denoted by O = (ot)T=1 where each ot is the index return at time t and is determined randomly following a normal distribution N (pi, ai), for i = 1,2, where the two parameters of this distribution depend on the state of the hidden Markov chain, i.e. on the market regime. We will call them the observation distributions and denote the corresponding sequence of distributions by B. The entire model will be denoted by M := {A, П, B}. We first present the forward algorithm which helps computing the so called forward variable Kt(i) = p(o102 ...Ot,qt = i | M), Here, к is defined recursively: K1(i) = nibi(o1), for i = 1, 2, where bi(o1) is the density of N(pi,ai) at the point o1. Furthermore, Kt+1(j) J2Kt(i)a i= 1 bj(0t+1),. for j = 1, 2 and t = 1,2,... T - 1 and bj (ot+1) is defined by analogy with the above. Likelihoodp(O|M) can be computed using the forward variable in the following manner: 2 p(O|M) = E кг (i). i=1 In the Baum-Welch algorithm we also need the backward variable et(i) = p(ot+1ot+2 ...ot | qt = i, M) which we compute recursively using the backward algorithm. We first initialize eT(i) = 1 for i = 1 , 2, and then let 2 et(i) = E aij bj (ot+1)et+1(j) j=1 for i = 1, 2 and t = T - 1, T - 2,..., 1. Likelihoodp(O|M) can be computed using the backward variable in the following manner: p(O|M) = E nie1(i)bi(o1). i=1 ij To initialize the Baum-Welch algorithm we choose a starting estimate of the model denoted by Mo, and then we compute the likelihoodp(O | Mo) using the forward variable. At this stage we start the iterative procedure consisting of four steps. The first step is to compute the forward variable Kt(i) and the backward variable et (i) based on the estimate of the model obtained on the previous iteration step Mk. The final result of this step is the likelihood of transition from the state i to the state j given the model Mk and the observations O when time goes from t to t + 1: iat(i,j) = p(qt = i,qt+1 = j |O, Mk ) _ p(qt = i,qt+i = j, O\Mk ) = p(O\Mk ) = Kt(i)aij bj (ot+i)ßt+i(j ) = p(o\Mk ) . The first equation above is just the definition, in the second one we use the conditional formula and in the third one we express the likelihood with (the estimates of) the forward and backward variable. Next, we express the denominator of the last fraction above also using (the estimates of) the forward and backward variable: 2 2 p(O \Mk) = 253 Kt(iRj bj (ot+i)dt+i(j)-i=i j=i On the second step of the iteration procedure we need to estimate another likelihood, i.e. R Fi(i)= p(qt = i \O, M) = 2 ^t(i,j). j=i Using the estimates of the first step i>t(i,j) we compute the next estimate of the model Mk+i. First, we compute the elements of the transition matrix A a. = Et=i Mi,j) 'ij x^T-i : Et=i rt(i) followed by the initial distribution П ai = Ti(i) and finally the observation distributions B aj (s) = ET=irt (j)'. £T=i rt(i) Here we understand rt(j)' given ot = s. On the third step of the iteration procedure we compute the likelihood using the new model Mk+i. The fourth step is decisive: we compare the estimates of these likelihoods on the last two steps. If they are close enough, we stop the algorithm. If not, we proceed with another iteration starting with step one. So, the final result of this algorithm is an estimate of the hidden Markov model M := {A, П, B}. Based on this estimate, we want to give a prediction of the most probable state (bull or bear regime) for each point of time in the period. This will be done using the Viterbi algorithm. We first introduce, for i = 1,2, the Viterbi variable 6t(i) = max p(qi,q2, ...,qt-i,qt = i, oi, 02, ..., ot \M) 9i,92,...,9t-i which means the conditional likelihood of the most likely path of length t ending in state i given the model, where ot are the the actual observed values of the index under consideration. We also need the value of the last but one state in this optimal path that ends in state i which we denote by Tt(i). We initialize by letting Ji(i) = nibi(oi) and Ti(i) = 0 (an "empty state" on which nothing really depends) for i = 1,2. The inductive steps of the algorithm go for t = 2,3,... ,T. The dynamic programming approach yields St(i) = bi(ot) max(6t-i(i)aij) i=1,2 together with Tt (j) = argmax(Jt-i(i)aij ). i=1,2 At the end of the algorithm we terminate with the final optimal regime qT = arg -1(i)aij) and then backtrack the whole optimal path qt = Tt+i(qJ+i) for t = T - 1,T - 2,..., 1. Both the Baum-Welch algorithm and the Viterbi algorithm are not what we usually call online algorithms that would process the input data in the sequence they would be fed to the algorithm. This is a shortcoming since we are looking for a way to determining the market regime as a stopping time in the sense of martingale theory, i.e. the decision about a certain point in time can be made only based on the data of previous points in time. We are overcoming this obstacle by implementing it in an expanding window approach. The starting window in our approach will be the period from January 2, 2000 to January 2, 2008 (because January 2 is the first trading day in a year). On each step we expand the window by one trading day until we reach May 31, 2013 where the period that we are interested in ends. In each of these windows we run the Baum-Welch and the Viterbi algorithm and retain only the final optimal regime qT. This way we determine the optimal market regime for each trading day of the period we are interested in a stopping time manner. In Figure 1 we present the results of the algorithm described above. In this figure the daily values of the S&P 500 index are superimposed over two backgrounds - the red one corresponds to days of the bear market and the blue to days of the bull market according to our estimation. It is clear that our model is able to recognize quite well the bear market of 2008/2009 as well as the more pronounced market corrections in the following years. However it does perform less admirably when recognizing the beginning of the bull market in 2009. This is not unexpected since this period was characterized by extremely high returns as well as high, albeit decreasing, volatility. Such conditions are not well aligned with our model which assumes only two market regimes - the one with high returns and low volatility and the one with low returns and high volatility. An obvious solution to this problem would be extending our model to 4 regimes. However, fitting such a model would require a much larger data sample than we have available. For example, the authors of [16] fit a 4-regime model on 123 years of data. Due to this limitation, we decided for the more parsimonous 2-regime model. 2008 200Э 2010 2011 2012 2013 Figure 1: S&P 500 index during the bear regime (red background) and the bull regime (blue background) 4 Linear regression In this section we study the connection between the stock returns and the data described in Section 2. Our main statistical tool will be the Fama-Macbeth cross-sectional linear regression [9]. This is a two-step procedure where a cross-sectional regression is performed for each time unit and then the time-series average of the estimated regression coefficients is calculated. We first present the results of an analysis in which the explained variables are the cumulative future returns, while the search data is used as the regressor. For each k running from 1 to 15 we perform a Fama-Macbeth regression for the cumulative return from the time T = t + 1 to the time T = t + k as the explained variable which we denote rijt+k where i is the index spanning our entire selection of stocks. This way we allow for different periods of time that may be of interest ranging from 1 trading day to roughly 1 trading month. We perform an additional regression to test the contemporary return, i.e. the return observed on T = t. As described before we use three types of search data: Wikipedia page visits (denoted by wikiijt), Google search queries for company tickers (denoted by goog_tickersit) and Google search queries for company names (denoted by goog_namesit). All this data is taken at time T = t. The cross-sectional regressions performed for each time unit t and each k = 0,1,..., 15 are described by the following equations: Ti,t+k = ai,t + ^wiki.t wikij,t + £i,t+k, ri,t+k = «i,t + egoog_tickers,t goog_tickerSi,t + £i,t+k, ri,t+k ai,t + egoog_names,t goog_nameSi,t + e^t+k. There is an additional regressor we want to test for its influence which has a form of a dummy variable, i.e. the market regime. So, in practice we perform three actual regressions for each case of interest, one for the bear markets, one for the bull markets and one for the joint data independent of the regime. The results are presented in Table 1. To make the presentation clearer we standardized the search data on every fixed date under consideration so that the regression coefficient has a simple interpretation. It gives an increase in the average return (positive or negative) given that the average internet search variable increases by one standard deviation. When we want to present the relation of these data to the length of the period we run into another difficulty, namely that the cumulative returns computed for different periods are not immediately comparable since their magnitude trivially depends on the period, so we decided to annualize them. There are three graphs in Figure 2. The first one presents dependence of annualized returns (based on regression coefficients) for the Wikipedia page visits, where the red line presents the data of the bear market, the blue one the data of the bull market and the purple one the joint data. A similar graph is presented for the Google search data for company tickers and the third one for the Google search data for company names. Figure 2: Changes in annualized future returns over k days after we observe a one standard deviation increase in individual search variables. We first observe that the regression results in the case when market regimes are not taken under consideration differ substantially from the results when we do take them into account. Actually, in all the three cases of internet search variables we observe that a raise in the internet search variable is associated with higher future returns in the bear market compared to future returns in the bull market. This is confirming our starting hypothesis that market regimes have a strong influence on the connection between internet search data and stock returns. We also observe that there is a substantial difference between the importance of distinct internet search variables. Based on the analysis performed so far it seems that the possible influence of search data on stock returns is statistically the strongest for Google company tickers, followed by Google company names and finally Wikipedia page visits. It is also evident that the influence of search data on future returns is mostly short term with the largest absolute values of annualized returns (based on regression coefficients) attained for cases where k < 10. Wikipedia Google tickers Google names k Both Bear Bull Both Bear Bull Both Bear Bull 0 0.018* 0.010 0.027*** (0.006) (0.009) (0.007) 0.017** 0.041** 0.003 (0.008) (0.017) (0.008) 0.006 0.019 0.004 (0.008) (0.015) (0.007) 1 0.002 0.005 -0.003 (0.004) (0.008) (0.004) 0.019** 0.022 0.011 (0.009) (0.020) (0.008) 0.000 0.014 -0.014* (0.008) (0.014) (0.007) 2 0.009 0.031** -0.018*** (0.008) (0.016) (0.007) 0.042*** 0.061* 0.016 (0.016) (0.032) (0.015) 0.010 0.048** -0.026** (0.013) (0.021) (0.012) 3 0.015 0.044** -0.022** (0.012) (0.022) (0.010) 0.082*** 0.125*** 0.040* (0.024) (0.044) (0.022) 0.028 0.089*** -0.030* (0.018) (0.030) (0.017) 4 0.020 0.057** -0.026** (0.014) (0.026) (0.012) 0.131*** 0.205*** 0.056** (0.032) (0.058) (0.026) 0.046** 0.121*** -0.022 (0.022) (0.037) (0.021) 5 0.018 0.053* -0.027* (0.017) (0.031) (0.014) 0.156*** 0.235*** 0.079** (0.039) (0.069) (0.031) 0.059** 0.148*** -0.021 (0.027) (0.047) (0.025) 6 0.018 0.046 -0.021 (0.020) (0.035) (0.016) 0.186*** 0.280*** 0.098*** (0.047) (0.083) (0.037) 0.075** 0.184*** -0.019 (0.033) (0.056) (0.029) 7 0.017 0.042 -0.021 (0.024) (0.040) (0.019) 0.221*** 0.320*** 0.128*** (0.055) (0.095) (0.045) 0.078** 0.191*** -0.018 (0.037) (0.061) (0.033) 8 0.027 0.056 -0.018 (0.026) (0.044) (0.022) 0.251*** 0.363*** 0.152*** (0.062) (0.104) (0.052) 0.077* 0.198*** -0.024 (0.040) (0.066) (0.036) 9 0.032 0.063 -0.018 (0.029) (0.048) (0.023) 0.277*** 0.387*** 0.181*** (0.066) (0.112) (0.057) 0.077* 0.205*** -0.028 (0.044) (0.070) (0.040) 10 0.040 0.076 -0.015 (0.031) (0.051) (0.025) 0.292*** 0.398*** 0.198*** (0.072) (0.122) (0.061) 0.075 0.208*** -0.031 (0.046) (0.075) (0.043) 11 0.047 0.091* -0.018 (0.033) (0.054) (0.027) 0.312*** 0.414*** 0.223*** (0.076) (0.128) (0.066) 0.074 0.206** -0.030 (0.049) (0.080) (0.046) 12 0.052 0.100* -0.021 (0.035) (0.058) (0.029) 0.314*** 0.392*** 0.243*** (0.079) (0.135) (0.070) 0.059 0.169** -0.025 (0.052) (0.085) (0.048) 13 0.053 0.099 -0.020 (0.037) (0.061) (0.031) 0.325*** 0.371*** 0.278*** (0.084) (0.143) (0.076) 0.054 0.168* -0.035 (0.054) (0.089) (0.051) 14 0.056 0.107* -0.022 (0.039) (0.063) (0.033) 0.334*** 0.372** 0.290*** (0.088) (0.150) (0.081) 0.042 0.148 -0.042 (0.056) (0.092) (0.053) 15 0.061 0.109 -0.016 (0.042) (0.068) (0.035) 0.353*** 0.394** 0.306*** (0.094) (0.159) (0.087) 0.034 0.137 -0.049 (0.059) (0.097) (0.056) Table 1: Regression coefficients of internet search variables in Fama-MacBeth regressions where cumulative future stock returns are the explained variable. Table columns correspond to different regressor-regime combinations and table rows correspond to different horizons of future returns. Standard errors for regression coefficients are given in parentheses. Statistical significance at levels of 10%, 5% and 1% is denoted by *,**, and ***, respectively. Additionally, statistically significant results are printed in bold. We observe another phenomenon which is best seen in the case of Wikipedia page visits. In the bear market the values of returns first go up and then go back down so that they form a shape of a frown. During the bull market, on the other hand, we observe a mirror shape, i.e. a shape of a smile. Now, the interpretation of these shapes is in some sense the opposite of the usual meaning conveyed by these shapes. While the frown noticed means good news in bear times, the smile means bad news in bull times. So, we propose the two shapes to be called the "merry frown" and the "sour smile". These shapes are not so easy to interpret. A possible explanation (taking into account also some other details of the Wiki shape) is that in bear markets investors are pessimistic and their overpessimistic reaction after increased attention perceived via the number of Wikipedia page visits on the first day, results in a counter-reaction in the days to follow and creates the merry frown. In the bull markets though investors are optimistic and their overoptimistic immediate reaction on the increased attention overturns into a sour smile. In the next step we investigate whether the observed connection between internet search data and stock returns can be explained by including additional factors into our model. We replace our initial explained variable ri t (future cumulative returns) by the so-called abnormal cumulative returns arijt. These returns are obtained as residuals in a variant of the Carhart [5] four factor asset pricing model which is an extension of the well known Fama-French model [10]. The model is defined by the following equation: n,t+k = rrtf + ßi,i(rmkt - rf ) + ß2,i HMLt + ßs,i SMBt + ß4,i UMDt + arM+k, where rf is the risk-free rate of return (approximated by the daily rate of one month U.S. Treasury bills), rtnkt — rr is the excess return of the entire stock market over the risk-free return, HMLt is the return difference between a portfolio of stocks with high and low book-to-market stocks, SMBt is the return difference between a portfolio of small and big stocks in terms of their market capitalization and UMDt is the return difference between a portfolio of stocks with high and low returns in the past year. The betas are estimated on a daily basis, using a rolling window of 120 days. We repeat the cumulative return regressions described above for the case of abnormal cumulative returns and report the results in Table 2. The period dependencies of annualized abnormal returns are displayed in Figure 3. We see that results in the case of Wikipedia page visits variable are quite similar to those obtained before accounting for the Carhart factors. However the influence of Google search queries for company tickers and company names is greatly diminished. In fact, no statistically significant results at the 5% level are obtained for the company tickers variable regardless of the bear or bull market. This is in contrast to previous research which was performed on samples taken from earlier periods. Furthermore, we find little evidence that a rise in internet search variables corresponding to individual companies might directly translate into short-term buying pressure and consequently higher stock prices. The differences between the bear market and the bull market remain clearly visible, especially in the case of Wikipedia page visits. While our results show that a statistically significant dependence between future stock returns and internet search variables exists, we do note that the explanatory power of all the tested regressions as measured by the R2 statistic is very low and only rises above 1% in a few cases. This is not unexpected if we take into account the fact that the regressions are predictive, that we are only using a single explanatory variable and that future stock returns are notoriously hard to predict. The questions whether our observations can nevertheless be used to obtain economic gains will be explored in Section 5. Wikipedia Google tickers Google names k Both Bear Bull Both Bear Bull Both Bear Bull 0 0.014*** 0.006 0.022*** (0.004) (0.007) (0.006) 0.011* 0.010 0.015** (0.006) (0.012) (0.006) -0.000 -0.006 0.006 (0.005) (0.008) (0.006) 1 0.002 0.006 -0.003 (0.003) (0.006) (0.003) 0.002 0.004 0.003 (0.006) (0.011) (0.006) -0.010** -0.012 -0.009 (0.005) (0.008) (0.005) 2 0.005 0.018* -0.013** (0.006) (0.011) (0.006) 0.005 0.006 0.003 (0.010) (0.019) (0.010) -0.016* -0.009 -0.022** (0.009) (0.014) (0.009) 3 0.012 0.034** -0.015** (0.008) (0.015) (0.008) 0.003 0.001 0.005 (0.015) (0.027) (0.013) -0.024** -0.018 -0.028** (0.012) (0.020) (0.013) 4 0.018* 0.048** -0.016* (0.011) (0.019) (0.009) 0.017 0.030 0.006 (0.019) (0.035) (0.016) -0.022 -0.009 -0.031* (0.015) (0.026) (0.016) 5 0.019 0.048** -0.017 (0.013) (0.021) (0.011) 0.019 0.034 0.005 (0.023) (0.042) (0.019) -0.022 0.000 -0.038** (0.019) (0.031) (0.019) 6 0.020 0.046* -0.013 (0.014) (0.024) (0.013) 0.024 0.042 0.009 (0.026) (0.050) (0.023) -0.024 0.004 -0.044** (0.021) (0.036) (0.022) 7 0.024 0.054** -0.015 (0.016) (0.027) (0.014) 0.03 8 0.064 0.018 (0.030) (0.056) (0.027) -0.021 0.011 -0.043* (0.024) (0.040) (0.025) 8 0.027 0.059** -0.015 (0.018) (0.029) (0.016) 0.046 0.072 0.027 (0.033) (0.062) (0.032) -0.025 0.012 -0.052* (0.027) (0.045) (0.028) 9 0.029 0.062** -0.014 (0.019) (0.031) (0.017) 0.059 0.084 0.044 (0.038) (0.070) (0.036) -0.032 0.007 -0.062** (0.029) (0.048) (0.030) 10 0.031 0.068** -0.016 (0.020) (0.033) (0.019) 0.057 0.076 0.050 (0.041) (0.077) (0.039) -0.043 -0.006 -0.070** (0.032) (0.052) (0.032) 11 0.034 0.074** -0.016 (0.022) (0.036) (0.021) 0.066 0.089 0.058 (0.045) (0.084) (0.043) -0.051 -0.017 -0.075** (0.034) (0.056) (0.034) 12 0.036 0.077** -0.017 (0.023) (0.039) (0.022) 0.066 0.084 0.061 (0.048) (0.087) (0.046) -0.058 -0.029 -0.079** (0.037) (0.061) (0.036) 13 0.039 0.081* -0.013 (0.025) (0.042) (0.023) 0.070 0.081 0.072 (0.051) (0.094) (0.048) -0.067* -0.041 -0.085** (0.039) (0.064) (0.039) 14 0.040 0.086* -0.018 (0.027) (0.044) (0.024) 0.070 0.076 0.078 (0.054) (0.101) (0.050) -0.077* -0.060 -0.087** (0.041) (0.067) (0.040) 15 0.039 0.084* -0.018 (0.029) (0.047) (0.026) 0.072 0.074 0.085 (0.057) (0.107) (0.053) -0.085* -0.075 -0.088** (0.044) (0.070) (0.043) Table 2: Regression coefficients of internet search variables in Fama-MacBeth regressions where cumulative abnormal future stock returns are the explained variable. Table columns correspond to different regressor-regime combinations and table rows correspond to different horizons of future returns. Standard errors for regression coefficients are given in parentheses. Statistical significance at levels of 10%, 5% and 1% is denoted by *,**, and ***, respectively. Additionally, statistically significant results are printed in bold. Figure 3: Changes in abnormal annualized future returns over k days after we observe a one standard deviation increase in individual search variables. Based on these tests we conclude that the influence of the kind of attention noticed by Google search queries (either for company tickers or names) can be perceived also by other data that are more commonly applied by financial practitioners. On the other hand, it shows that Wikipedia page visits do indeed provide new information about the behavior of stock prices. Also, we perceive that the merry frown and sour smile effects persist for Wikipedia page visits even after controlling for the most commonly used asset pricing factors. 5 Trading strategies In this section we want to verify how the results of Section 4 can be used, if at all, in forming trading strategies. In other words, we want to either statistically prove or disprove that internet data can increase our profits in financial markets. The evidence for the influence of internet search data on future stock returns is most compelling in the case of Wikipedia data, as shown in Section 4. Since we were also not able to find any examples in existing literature of this type of data being used in construction of trading strategies based on individual stocks, we decided to limit our analysis in this chapter only to Wikipedia page visits. Our results show that in bear markets higher Wikipedia page visits are positively correlated with short term future return while in bull markets the corresponding correlation is negative. So we propose the strategy for bear markets to enter long position at the end of the trading day for stocks in the upper decile with respect to the most recent available data on Wikipedia page visits; and similarly to enter the short position at the end of the trading day for stocks in the lower decile with respect to these visits. In the bull market, the strategy is to do exactly the opposite. We propose that all the long positions and all the short positions are entered using the same weights with respect to the wealth that we are prepared to invest into this strategy. Since the data on Wikipedia page visits for any given day is only made available the following day we lag our Wikipedia variable for one day to ensure that the data would have been available at the time of our trading decision. Of course there is a problem of determining the actual frequency of trading, this means for how long we should hold our positions. We know that we are talking about a short term effect, but what does short really mean in this particular context? To make this dilemma as clear as possible we are making a number of tests using some alternatives. Let f be the number of trading days between two consecutive trading decisions. For f = 1,2,..., 8 we are testing the f th strategy and give the result for three options. The first option is that we allow only trading in the bear market, the second one is that we allow trading only in the bull market and the third one is that we allow trading during both markets. In Table 3 we present the results obtained in percentage points of the annual return. It is clear that f = 1 is the best of the proposed strategies in all the three cases. It is also clear that the results are getting smaller with f increasing in the case of combined strategy and the bull-only strategy. However in the case of the bear strategy f = 2 and f = 3 are slightly better than f = 1. For f big enough the results of the strategies seem to become more or less random. The best of the three options tested is the combined application of both bear and bull strategy. It is also clear that the results of the bear strategy are better than the results of the bull strategy. Trading frequency (days) 1 2 3 4 5 6 7 8 Bear Bull Joint Bear & Bull 5.28 5.50 6.10 4.02 2.30 0.94 -0.55 2.21 2.76 1.33 -0.15 0.91 0.59 1.02 1.08 -2.53 8.19 6.91 5.94 4.97 2.91 1.97 0.52 -0.37 Table 3: Average annual returns (in percentage points) of proposed trading strategies in relation to the trading frequency. We also want to compare our strategies to suitable benchmark strategies. However, as we believe, the most usual benchmarks such as various indices are long-only strategies and the comparative testing with our strategies which include both long and short positions would not be fair. So we decided to compare it with random strategies using a Monte Carlo approach. Our control strategy is to choose in a uniformly random way 10 % stocks to be put in a long position and 10 % to be put in a short position. We created 1000 strategies of this type and computed the average yearly return for each of them. This produces a random sample of possible average yearly returns which we compare statistically to the average return of each of the strategies under consideration. As usual in this kind of situation, we perform a one-sample one-way Student t-test where we test the null hypothesis that the mean yearly return for the population of random strategies is equal to the return of our strategies against the alternative hypothesis which states that the mean yearly return for the population of random strategies is lower to the return of our strategies. As can be seen from results given in Table 4 we can reject this hypothesis for our joint bear and bull strategies for most of the trading frequencies considered. In Figure 4 we want to present a slightly different view on the results of our strategies compared to the random approach. Assume we invest a certain equity in the strategies above to be compared; and that we invest the same amount into each of the random strategies described in the previous paragraph. We compare the average of the randomly invested equity to the equity gained via the strategy under consideration for each day of our sample period. More interesting than the averages as such are the bands created around the averages using the daily standard deviation and its small multiples. We can see that the equity invested in our joint bear and bull strategy mostly stays in the area that is beyond the band which is three standard deviations above the average equity of random strategies. Trading frequency (days) 1 2 3 4 5 6 7 8 ßö 8.19 6.91 5.94 4.97 2.91 1.97 0.52 -0.37 ß -0.25 -0.25 -0.25 -0.25 -0.25 -0.25 -0.25 -0.25 t value -64.55 -54.78 -47.34 -39.91 -24.18 -16.96 -5.89 0.92 p value 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.82 Table 4: The results of a one-way Student t-test for testing the null hypothesis that the mean yearly return (p) for the population of random strategies is equal to the return (p0) for our joint bear and bull strategy against the alternative hypothesis p < p0. The p values are given in percentage points and rounded to two decimals. Figure 4: Equity curve of joint bear and bull strategy with trading frequency of one day compared to equities of random strategies represented by standard deviation bands around the mean equity. In most cases the Wikipedia pages on individual S&P 500 companies contains only the most basic information. It is therefore safe to assume that this information source will mostly be utilized by individual investors since institutional investors have access to more sophisticated tools offering greater depth of information. Our hypothesis is that the influence of Wikipedia page visits on future stock returns will be higher for stocks that are likely to attract a higher proportion of individual investors. According to Barber and Odean [4] the individual investors generally have a tendency to tilt their stock investments towards high-beta, small and value stocks. In light of this result we construct three additional strategies based on our joint bear and bull strategy. In all these strategies we restrict our trading decision to a subsample of stocks that fall above or below the median of one of the following variables: volatility, market capitalization and price-to-book ratio. In the first strategy we choose a subsample of high volatility stocks, in the second one we choose a subsam-ple of low market capitalization stocks and in the last one we chose a subsample of low price-to-book ratio. Volatility is calculated in a 20 trading day rolling window approach. The market capitalization and price-to-book ratio variables are obtained from the ADVFN service5. We present the results in Table 5. The results strongly support our hypothesis since all three subsample strategies outperform the full sample joint bear and bull strategy in cases of the most relevant trading frequencies (1 < f < 5). Trading frequency (days) 1 2 3 4 5 6 7 8 Full sample 8.19 6.91 5.94 4.97 2.91 1.97 0.52 -0.37 High volatility subsample Low market cap subsample Low price to book subsample 15.35 16.40 10.13 8.53 4.12 7.15 -0.78 1.32 17.43 13.41 9.07 6.86 4.53 5.49 0.31 1.12 13.57 9.99 7.16 8.15 5.87 0.81 -0.64 3.30 Table 5: Average yearly returns (in percentage points) of subsample strategies compared to average yearly return of full sample joint bear and bull strategy. All of the strategies presented up to this point have included both long and short positions. Since many investors face restrictions with respect to opening short positions in stocks, the question naturally arises whether our strategies can be adapted to be long-only. Let us consider the simplest possible adaptation which is the strategy where the investment rule during the bear market is to enter long positions at the end of each trading day for stocks in the upper decile with respect to the most recent available data on Wikipedia page visits. During the bull market, the strategy enters long positions for stocks which are in the lower decile with respect to the Wikipedia page visits. The annualized return of such a strategy in our sample period is 20.36%. Since the strategy is long-only, it is reasonable to compare its performance to that of the S&P 500 index whose annualized return during our sample period is merely 4.36%. The equity curves obtained by investing the same amount of wealth in both our adapted long-only strategy and the S&P 500 index are shown in Figure 5. The backtesting results favor the conclusion that even those investors who are restricted to only opening long positions might benefit from including the information about Wikipedia page visits in their investment decisions. 6 Conclusion The key point of our paper is that it is essential to incorporate information about the market regime when studying the influence of internet search data on stock returns. This is clearly true for all the search variables considered since all show markedly higher correlations with future stock returns in the bear regime than in the bull regime. However, the distinction between the two regimes is especially singnificant in the case of the Wikipedia variable where we observe two inverse price patterns - a merry frown in the bear regime and a sour smile in the bull regime. Our regime estimation method is based on a hidden Markov model that only accounts for information revealed to us through the price fluctuations of the S&P 500 index. We suspect that even more interesting results might be obtained if search data were somehow included into the regime switching model itself, perhaps by building upon existing research into estimation of investor sentiment by internet search data such as [8]. After controlling for the Carhart factors the Wikipedia page visits variable emerges 5http://www.advfn.com Figure 5: Equity curve of adapted long-only joint bear and bull strategy with trading frequency of one day compared to equity curve of S&P 500 index. as the one with the most significant influence on future stock returns. Until recently this data set has been largely overlooked by researchers however we believe that it holds great potential for future applications. In a surprising turn, both of the Google search variables prove to be statistically insignificant for most periods of future return for stocks in our sample. This result is at odds with previous studies performed on earlier sample periods and warrants further research that would explain this discrepancy. We suggest that this might be caused by arbitrageurs already taking advantage of the effect of company-related Google search frequencies in line with the weak-form market-efficiency hypothesis. We would like to make an additional point about Google Trends data with regard to future research. We noticed that previous studies have almost exclusively focused on relative search frequencies which is most likely due to the fact that individual time series obtained from the Google Trends service are normalized within series so that their values always span the interval from 0 to 100. In Section 2 we describe a straightforward approach that enables us to obtain full sample daily trends data regardless of normalization. A quite similar approach might be used to obtain data where non-relative search frequencies of two different terms can be compared. It would be interesting to know whether such data provides us with an even better proxy for investors' attention. We also believe that the results presented in our paper may be of benefit to financial practitioners in at least two ways. Firstly, we show that Wikipedia can provide investors with insights into a stock's risk profile that are overlooked by existing asset pricing models such as the Carhart four-factor model. Secondly, the trading strategies presented in Section 5 may be of interest to speculative investors who are comfortable executing trading strategies with target investment holding periods of less than a week. References [1] A. Ang and A. Timmermann, Regime Changes and Financial Markets, Annual Review of Financial Economics, 4 (2012), 313-337. [2] M. Bank, M. Larch and G. Peter, Google search volume and its influence on liquidity and returns of German stocks, Financial Markets and Portfolio Management, 25 (2011), 239-264. [3] B. M. Barber and T. Odean, All that glitters: The effect of attention and news on the buying behavior of individual and institutional investors, Review of Financial Studies, 21 (2008), 785818. [4] B. M. Barber and T. Odean, Trading Is Hazardous to Your Wealth: The Common Stock Investment Performance of Individual Investors, Journal of Finance, 55 (2000), 773-806. [5] M. M. Carhart, On the persistence in mutual fund performance, Journal of Finance, 52 (1997), 57-82. [6] H. Choi and H. Varian, Predicting the Present with Google Trends, Economic Record, 88 (2012), 2-9. [7] Z. Da, J. Engelberg and P. Gao, In Search of Attention, Journal of Finance, 66 (2010), 14611499. [8] Z. Da, J. Engelberg and P. Gao, The Sum of all FEARS: Investor Sentiment and Asset Prices, Working Paper, (2010). [9] E. F. Fama and J. MacBeth, Risk, return and equilibrium: Empirical tests, Journal of Political Economy, 81 (1973), 607-636, [10] E. F. Fama and K. R. French, Common risk factors in the returns on stocks and bonds, Journal of Financial Economics, 33, 3-56. [11] M. S. Drake, D. T. Roulstone and J. R. Thornock, Investor Information Demand: Evidence from Google Searches Around Earnings Announcements, Journal of Accounting Research, 50 (2012), 1001-1040. [12] J. Ginsberg, M. H. Mohebbi, R. S. Patel, L. Brammer, M. S. Smolinski and L. Brilliant, Detecting influenza epidemics using search engine query data, Nature, 457 (2009), 1012-1014. [13] J. D. Hamilton, A New Approach to the Economic Analysis of Nonstationary Time Series and the Business Cycle, Econometrica, 57 (1989), 357-384. [14] K. Joseph, M. B. Wintoki, Z. Zhang, Forecasting abnormal stock returns and trading volume using investor sentiment: Evidence from online search, International Journal of Forecasting, 27 (2011), 1116-1127. [15] S. Mitra, P. Date, Regime switching volatility calibration by the Baum-Welch method, Journal of Computational and Applied Mathematics, 234 (2010), 3243-3260. [16] J. M. Maheu, T. H. McCurdy, Y. Song, Extracting Bull and Bear Markets from Stock Returns, Working Paper, (2009). [17] H. S. Moat, C. Curme, A. Avakian, D. Y. Kenett, H. E. Stanley, and T. Preis, Quantifying Wikipedia usage patterns before stock market moves, Scientific Reports, 3 (2013) [18] T. Preis, H. S. Moat and H. E. Stanley, Quantifying Trading Behavior in Financial Markets Using Google Trends, Scientific Reports, 3 (2013) [19] L. R. Rabiner, A tutorial on hidden markov models and selected applications in speech recognition, Proceedings of the IEEE, 77 (1989), 257-285. [20] N. Vlastakis and R. N. Markellos, Information demand and stock market volatility, Journal of Banking & Finance, 6 (2012), 1808-1821. ARS MATHEMATICA CONTEMPORANEA Also available at http://amc-journal.eu ISSN 1855-3966 (printed edn.), ISSN 1855-3974 (electronic edn.) ARS MATHEMATICA CONTEMPORANEA 9 (2015) 321-344 Subdivision into i-packings and S-packing chromatic number of some lattices Nicolas Gastineau * LE2I, UMR CNRS 6306, Université de Bourgogne, 21078 Dijon, France Hamamache Kheddouci Université de Lyon, CNRS, Université Lyon 1, LIRIS, UMR5205, F-69622, France Olivier Togni LE2I, UMR CNRS 6306, Université de Bourgogne, 21078 Dijon, France Received 30 January 2013, accepted 3 June 2015, published online 7 August 2015 An i-packing in a graph G is a set of vertices at pairwise distance greater than i. For a nondecreasing sequence of integers S = (sb s2,...), the S-packing chromatic number of a graph G is the least integer k such that there exists a coloring of G into k colors where each set of vertices colored i, i = 1,..., k, is an sj-packing. This paper describes various subdivisions of an i-packing into j-packings (j > i) for the hexagonal, square and triangular lattices. These results allow us to bound the S-packing chromatic number for these graphs, with more precise bounds and exact values for sequences S = (sj,i G N*), Sj = d + L(i - 1)/nJ. Keywords: Packing chromatic number, i-packing, hexagonal lattice, square lattice, triangular lattice, distance coloring. Math. Subj. Class.: 05C15, 05C63, 05C70. 1 Introduction Let G = (V, E) be a (finite or infinite) graph and let N (u) = {v g V (G) | uv G E (G)} be the set of neighbors of vertex u. A set Xj С V (G) is an i-packing if for any distinct pair u, v g Xj, d(u, v) > i, where d(u, v) denotes the usual shortest path distance between u and * Partially supported by the Burgundy Council and the Rhone-Alpes Council. E-mail addresses: Nicolas.Gastineau@u-bourgogne.fr (Nicolas Gastineau), hamamache.kheddouci@univ-lyon1.fr (Hamamache Kheddouci), Olivier.Togni@u-bourgogne.fr (Olivier Togni) Abstract v. We will use Xj to refer to an i-packing in a graph G. A k-coloring c of G is a map from V(G) to {1,..., k} such that for every pair of adjacent vertices (u, v), we have c(u) = c(v). For a graph G and a k-coloring c of G, let cj be {u G V (G) | c(u) = i}. The smallest integer k such that there exists a k-coloring of G for which for every i, with 1 < i < k, cj is a i-packing, is called the packing chromatic number of G and is denoted by xP(G). This concept was introduced by Goddard et al. [7] under the name of broadcast chromatic number. More generally, for a nondecreasing sequence of integers S = (s^ s2,...), an S -packing k-coloring is a k-coloring c of V (G) such that for every i, with 1 < i < k, cj is a sj-packing. Such a coloring will also simply be called an (s1,..., sk)-coloring. The S -packing chromatic number of G denoted by xS (G), is the smallest k such that G admits an S -packing k-coloring. For the sequences S = (sj, i G N*), with sj = d + [(i - 1)/nJ, we call xSp (G) the (d, n)-packing chromatic number and denote it by xp'n(G). For any connected graph G such that |V(G)| > d +1, xdP'n(G) > d +1 and xp'1(G) = Xp(G). For every bipartite graph G, xi'2 (G) = 2 (a bipartite graph is 2-colorable). Moreover, the smallest n such that xp'n(G) = n corresponds to the d-distant chromatic number [12], i.e. the minimum number of d-packings that form a partition of the vertices. Let denote the two-way infinite path, let Z2 = denote the planar square lattice (where □ is the Cartesian product), T denote the planar triangular lattice and H denote the planar hexagonal lattice. In this article, for an (s1, s2,...)-coloring of a graph, we prefer to map vertices to the color multiset {s1, s2,...} even if two colors can then be denoted by the same number. This notation allows the reader to directly see to which type of packing the vertex belong depending on its color. When needed, we will denote colors of vertices in different i-packings by ia, ib,.... 1.1 Motivation and related work Packing colorings in graphs are inspired from frequency planning in wireless systems. The concept of S-packing coloring emphasizes the fact that signals can have different powers similar to the packing coloring but enables the presence of several signals with the same power, providing a more realistic model for the frequency assignment problem. The packing chromatic number of lattices has been studied by several authors: Soukal and Holub [13] proved that Xp(Z2) < 17, Ekstein et al. [1] that 12 < Xp(Z2); Fiala et al. [4] showed that Xp(H) < 7, Xp(Z2^) = то and Xp(H^Pe) = то and Finbow and Rall [5] proved that xp(T) = то. S-packing colorings with sequences S other than (1,2,..., k) first appear in [7, 3]. Goddard and Xu [8] have recently studied S-packing colorings for the infinite path and for square and triangular lattices, determining conditions on the first elements of the sequence for which the graph is or is not S-packing-colorable. Regarding the complexity, Goddard et al. [7] proved that the problem of (1,2,3)-packing coloring is polynomial while (1, 2, 3,4)-packing coloring and (1, 2, 2)-packing coloring are NP-complete. Fiala and Golovach [3] showed that the problem of (1,2,..., k)-coloring is NP-complete for trees. The NP-completeness of (1,1, 2)-coloring was proved by Goddard and Xu [9] and afterward by Gastineau [6]. While the packing coloring corresponds to an S-packing coloring with a strictly increasing sequence and the d-distant chromatic number corresponds to a constant one, the sequence in the (d, n) -packing coloring also tends to infinity, but the parameter n allows us to control its growth. d\n 1 2 3 4 5 6 1 7 [4, 11] 2 2 2 2 2 2 TO 5-8 5 4 4 4 3 TO 15-35 9- 13 8 - 10 7-8 6 4 TO 61 - ? 20-58 15-27 13-21 12- 18 5 TO TO 37-? 25-? 21 - ? 19- ? 8 TO TO TO ? ? ? 11 TO TO TO TO ? ? 13 TO TO TO TO TO ? 16 TO TO TO TO TO TO Table 1: Bounds for (d, n)-packing chromatic numbers of the hexagonal lattice. Moreover, one can note that all the S-packing colorings of square and hexagonal lattices published so far have the property that the s i-packing is maximum and the other si -packings are obtained by subdivisions of s 1-packings (and are not always maximum). Therefore, we find it interesting to study subdivision of an i-packing into j -packings, j > i, in lattices. These subdivisions can in turn be used to describe patterns to obtain an S-packing coloring of a lattice. However, determining the families of graphs G for which for any S such that G is S-colorable, the S-coloring satisfies the above property remains an open question. Recently Goddard and Xu [8] proved that there exist nondecreasing sequences S such that P» is S-colorable and in any (s1,..., xSp (P»))-packing coloring of P», the s 1 -packing is not maximum, showing that for P», there are sequences S for which the above property is not satisfied. 1.2 Our results The second section introduces some definitions and results related to density. The third section introduces some subdivision of the lattices into i-packings. The fourth and fifth sections give lower bounds resulting from Section 2 and upper bounds resulting from Section 3 for the S -packing chromatic number and the (d, n)-packing chromatic number of the lattices H, Z2 and T. Tables 1, 2 and 3 summarize the values obtained in this paper for the (d, n)-packing chromatic number, giving an idea of our results. The emphasized numbers are exact values and all pairs of values are lower and upper bounds. Lower bounds have been calculated from Proposition 2.2 and Propositions 2.5, 2.7 and 2.9. Some of the results for square and triangular lattice have been found independently by Goddard and Xu [8]. 2 Density of i-packings 2.1 Density of an i-packing in a lattice Let G = (V, E) be a graph, finite or infinite and let n be a positive integer. For a vertex x of G, the ball of radius n centered at x is the set Bn(x) = {v g V(G)|dG(x, v) < n} and the sphere of radius n centered at x is the set dBn(x) = {v g V (G) |dG(x, v) = n}. The density of a set of vertices X С V (G) is d(X ) = limsupmax{ ^B^oi ^ }. The notion of k-area was introduced by Fiala et al. [4]. We propose here a slightly modified d\n 1 2 3 4 5 6 1 12-17 [1, 13] 2 2 2 2 2 2 TO 11-20 7-8 6 [8] 5 [8] 5 3 TO 57 - ? 16-33 12-20 10-17 10- 14 4 TO TO 44- ? 25-56 20-34 18-28 5 TO TO 199- ? 50- ? 35-? 29-? 6 TO TO TO ? ? ? 8 TO TO TO TO ? ? 10 TO TO TO TO TO ? 12 TO TO TO TO TO TO Table 2: Bounds for (d, n)-packing chromatic numbers of the square lattice. d\n 1 2 3 4 5 6 1 to [5] 5 - 6 [8] 3 3 3 3 2 TO 127-? 14- ? 10-16 9 - 13 8 - 10 3 TO TO 81 - ? 28-72 20-38 17-26 4 TO TO TO 104- ? 49 - ? 36- ? 5 TO TO TO TO ? ? 7 TO TO TO TO TO ? 8 TO TO TO TO TO TO Table 3: Bounds for (d, n)-packing chromatic numbers of the triangular lattice. A(2)=7 A(1)=3 Figure 1: Examples of k-area in T. definition: Definition 2.1. Let G be a graph, x e V (G), and let k be a positive integer. The k-area A(x, k) assigned to G is defined by : |Bk/2(x)| for k even; A{x,k)={ |BLfc/2J(x)| + E N (")nBik/2j(x)i+|N (u)ndBrk/2i(x)|/2 for k odd. deg(u) For vertex-transitive graphs, the k-areas are the same for all vertices, hence we denote it by A(k). The motivation for our modification of the notion of k-area with the introduction of the set of neighbors inside the sphere is to have sharper density bounds than the ones obtained by the initial notion of k-area. For the square and the hexagonal lattice the notion coincide as the relation is empty. However, for the triangular lattice, the density bound is smaller: the definition of Fiala et al. [4] gives A(1) = 2 whereas A(1) = 3 in our case since there are adjacent vertices in the sphere (as for every u e dB\(x), |N(u) n dBi(x)|/2 = 1, then it adds one to the initial definition of k-area). Figure 1 illustrates this example giving a coverage of the triangular lattice by balls of radius 1. In one case (on the left) the balls are disjoint and in the second case (on the right) each sphere can be shared by several balls. Observe that in the second case, each vertex u in a sphere centered at x has two neighbors, and hence N (u) n dB1(x)|/2 = 1. Proposition 2.2. Let G be a vertex-transitive graph with finite degree, and i be a positive integer. If Xi is an i-packing in G, then d(Xi) < 1 A(i) Proof. Observe that for arbitrary vertices x and y of an i-packing Xi, the sets B[i/2J (x) and B[i/2J (у) are disjoint, since the vertices x and y are at distance greater than i. Then d(Xi) < 1/|B[i/2j (x)|. Assume that i is an odd number and let u be a vertex at distance [i/2] from x, then u has deg(u) neighbors. If among these deg(u) neighbors, k neighbors are in B[i/2J (x) then u can be at distance [i/2] from only k/deg(u) vertices in Xi. Hence d(Xi) <1/(|BLi/2j(x)| + E |N(u)nB'k/2'(x)| uedB rfc/2i(x) deg(u) |). Moreover u andaneighbor v of u in Blì/2j(x) cannot be both at distance [i/2] from more than 2 vertices in Xit therefore uv can only belong to two spheres of radius \i/2] centered at a vertex in Xi. Hence it follows that d(Xi) < 1 /A(i). □ Corollary 2.3. Let G be a vertex-transitive graph with finite degree, and i be a positive integer. If G has a finite S-packing chromatic number, then ж i § AW — L Corollary 2.4. Let G be a vertex-transitive graph with finite degree, and i be a positive integer. If G has a finite (d, n) -packing chromatic number, then ж En Ж) - 1 i=d v ' An i-packing Xi is called a maximized i-packing if for any other i-packing Xi, d(Xi ) — d(Xi ). 2.2 Density of an i-packing in the hexagonal lattice Proposition 2.5. Let H be the hexagonal lattice, x bea vertex in V (H) and n bea positive integer. Then 1. |dBn(x)| = 3n; 2. |B„(x)| = f n2 + f n +1. Proof. 1. As the set dBn(x) always contains three more vertices than dBn-i(x), then |öB„(x)| = 3n. 2. The graph H is 3-regular and so |B1 (x) | = 4. Suppose the statement is true for n, then |Bn+i(x)| = |Bn(x)| + |0Bn+i(x)| = f n2 + f n + 1 + 3(n +1) = ( 2 n2 + f + 3n) + ( f n + f ) + 1 = f (n + 1)2 + 2 (n + 1) + 1 and the result follows by induction. □ Proposition 2.6. Let H be the hexagonal lattice and k be a positive integer. Then 1. A(2k) = f к2 + § к + 1; 2. A(4k + 1) = 6k2 + 6к + 2; 3. A(4k + 3) = 6k2 + 12k + 6. Proof. 1. The first property results easily from Proposition 2.5. 2. If n = 4k +1, then Proposition 2.5 gives |B2k(x)| = |(2k)2 + f (2k) + 1 = 6k2 + 3k +1. For every vertex y in dB2fc+1(x), y has no neighbor in dB2fc+1(x) other than itself, so ^(y) n dB2fc+1(x)| = 0. We have to distinguish two kinds of vertices: 3k vertices have two neighbors in B2k(x) and |dB2k+1(x)| - 3k = 3k + 3 vertices have one neighbor in B2k(x). Therefore, |A(4k + 1)| = 6k2 + 3k + 1 + f + fk+f = 6k2 + 6k + 2. 3. If n = 4k + 3, then Proposition 2.5 gives |B2k+1(x)| = f (2k + 1)2 + f (2k + 1) + 1 = 6k2 + 9k+4. For every vertex y in dB2k+2(x), y has no neighbor in dB2k+2(x)) other than itself, so |N (y)ndB2k+2 (x) | = 0. We have to distinguish two kinds of vertices: 3k vertices have two neighbors in B2k+2 (x) and |dB2k+2 (x) | —3k = 3k+6 vertices have one neighbor in B2k(x). Hence, we have |A(4k + 3)| = 6k2 +9k + 4+ + fk+6 = 6k2 + 12k + 6. □ Note that this result appeared in the article of Goddard and Xu [8]. 2.3 Density of an i-packing in the square lattice Proposition 2.7. Let Z2 be the square lattice, x be a vertex in V(Z2) and n be a positive integer. Then 1. \dBn(x)| = 4n; 2. \Bn(x)\ = 2n2 +2n + 1. Proof. 1. As the set dBn(x) always contains four more vertices than dBn-1(x), then \dBn(x)\ =4n. 2. The graph Z2 is 4-regular and so \B1 (x) \ = 5. Suppose the statement is true for n, then \Bn+i(x)\ = \Bn(x)\ + \dBn+i(x)\ = 2n2+2n+1+4n+4 = 2(n+1)2+6n+5-4n-2 = 2(n + 1)2 + 2(n + 1) + 1 and the result follows by induction. □ Proposition 2.8. Let Z2 be the square lattice and k be a positive integer. Then 1. A(2k) = 2k2 + 2k + 1; 2. A(2k + 1) = 2k2 +4k + 2. Proof. 1. The first property results easily from Proposition 2.7. 2. If n = 2k + 1, then Proposition 2.7 gives \Bk(x)\ = k2 + 2k +1. For every vertex y in dBfc+1(x), y has no neighbor in dBfc+1(x) other than itself, so \N(y) П dBfc+1 (x)\ = 0. We have to distinguish two kinds of vertices: 4k vertices have two neighbors in Bk (x) and 4 vertices have one neighbor in Bk(x). Hence, we have \A(2k + 1)\ = 2k2 + 2k + 1 + 2 f + 4 =2k2 +4k + 2. □ Note that this result appeared implicitly in the article of Fiala et al. [4]. 2.4 Density of an i-packing in the triangular lattice Proposition 2.9. Let T be the triangular lattice, x be a vertex in V (T) and n be a positive integer. Then 1. \dBn (x)\ =6n; 2. \Bn(x)\ = 3n2 +3n + 1. Proof. 1. As the set dBn(x) always contains six more vertices than dBn-1(x), then \dBn(x)\ = 6n. 2. The graph T is 6-regular and so \B1(x)\ = 7. Suppose the statement is true for n, then \Bn+1(x)\ = \Bn(x)\ + \dBn+1 (x)\ = 3n2 +3n+1+6n+6 = 3(n2 + 1)+3n+1 + 6-3 = 3(n2 + 1) + 3(n + 1) + 1 and the result follows by induction. □ Proposition 2.10. Let T be the triangular lattice and k be a positive integer. Then 1. A(2k) = 3k2 + 3k + 1; 2. A(2k + 1) = 3k2 + 6k + 3. Г^^П с 1 3 1 4 1 3 1 ^ 1_ 1 4 1 3 1 4 3 1 1 Figure 2: The sets X2 (2-packing), X3 (3-packing) and X4 (4-packing) in H. Proof. 1. The first property result easily from Proposition 2.9. 2. If n = 2k + 1, then Proposition 2.9 gives |Bk (x)| = 3k2 +3k + 1. For every vertex y in dBk+ i(x), y has two neighbors in dBk+i(x)) other than itself, so |N(y) П dB2k+1(x)| = 2. We have to distinguish two kinds of vertices: 6k vertices have two neighbors in Bk (x) and six vertices have one neighbor in Bk (x). Hence, we have |A(2k + 1)| = 3k2 + 3k + "" " "" " = 3k2 +6k + 3. □ -i i 6k+6 i о bk i 1 + 6 + 2 6 + 3 Subdivision of an i-packing in H, Z2 and T 3.1 Subdivision of a 2-packing in H Let X2 be the (unique) maximized 2-packing in H represented in Figure 2. Note that d(X2) = 1/A(2) = 1/4 and remark that four 2-packings form a partition of H if we translate X2 three times. The hexagonal lattice can be seen as a subgraph of the square lattice. In fact in Figure 2, H is represented as subgraph of the usual representation of the square lattice. In the square lattice, we can choose one vertex as the origin and all the other vertices can be nominated by a Cartesian coordinate. In every description of H, our origin (0,0) will be a vertex in the packing that we want to describe such that there is no edge between (0,0) and (0,1). In fact we illustrate packings with a figure in this subsection but it will not be the case after; we will use Cartesian coordinates in order to describe a packing. For example, X2 from Figure 2 is the set of vertices: X2 = {(2x + 4y, x)| x G Z, y G Z}. In Appendix A, we recall a proposition about distance in the hexagonal lattice from Jacko and Jendrol [10]. This proposition is useful to verify that a set is an i-packing. These verifications are left to the reader in the remaining propositions. Proposition 3.1. Let k > 0 and m > 0 be integers. There exist: 1. k2 (3k — 1)-packings that form a partition of X2; 2. 2k2 (4k — 1)-packings that form a partition of X2; 3. two (3 X 2k — 1)-packings that form a partition of a (4k — 1)-packings from Point 2; 4. m2 (3mk — 1)-packings that form a partition of a (3k — 1) -packing from Point 1; 5. m2 (4mk — 1)-packings that form a partition of a (4k — 1) -packing from Point 2. 3 3 2 2 2 Proof. 1. Let Ak be the (3k — 1)-packing defined by Ak = {(2kx + 4ky, kx) | x G Z, y G Z}. Let F = {(2i + 4j, i) |i, j G {0,..., k — 1} be a family of k2 vectors. Make k2 copies 13b 13b 13b 3a 3a 3a ; i , 1 13b 13b 13b 3a 3a 3a 1 , 1 13b 13b 13b ■ 3a 3a -- 3a -- 0 1 ršd T5C T5d |5] p^ Г5] I ptfVvV^i I |5e ~ , 5d j I 5e L5'a I I ^^^ -5a- Figure 3: Two 3-packings forming a partition of X2 (on the left) and four 5-packings forming a partition of X2 (on the right). of the set Ak and translate each one by a vector from F to obtain a partition of X2. 2. Let Bk be the (4k - 1)-packing defined by Bk = {(4kx, 2ky)| x e Z, y e Z}. Let F = {(4i + 2а, 2j + a)|i, j e {0,..., k - 1}, a e {0,1}} be a family of 2k2 vectors. Make 2k2 copies of the set Bk and translate each one by a vector from F to obtain a partition of X2 . 3. Note that A2k C Bk and if A2k is A2k translated by the vector (0,2k), then A2k U A2k = Bk. 4. Note that Amk C Ak. Let F = {(2mki + 4mkj, mki)|i, j e {0,..., m - 1}} be a family of m2 vectors. Make m2 copies of the set Amk and translate each one by a vector from F to obtain a partition of Ak . 5. Note that Bmk C Bk .Let F = {(4mki, 2mkj)|i, j e {0,..., m - 1}} be a family of m2 vectors. Make m2 copies of the set Bmk and translate each one by a vector from F to obtain a partition of Bk. □ Figure 3 illustrates a partition of X2 from Points 1 and 2 for k = 1. In the remaining of the section, the proofs of decomposition of a set X will be resumed in a table and the proofs of properties similar from those from Points 3, 4 and 5 will be left to the reader. 3.2 Subdivision of a 3-packing in H Let X3 = {(3x + 6y, x) | x e Z, y e Z} be the maximized 3-packing in H from Figure 2. Note that d(X3) = 1/A(3) = 1/6 and that six 3-packings form a partition of H if we translate X3 five times. Proposition 3.2. Let k > 0 and m > 0 be integers. There exist: 1. k2 pi,k -packings, pi,k = (4k - 1), that form a partition of X3; 2. 3k2 p2k -packings, p2,k = (6k - 1), that form a partition of X3; 3. 8k2 p3,k -packings, p3,k = (10k - 1), that form a partition of X3; 4. 24k2 p4,k -packings, p4,k = (18k - 1), that form a partition of X3; 5. m2 p j mk -packings that form a partition of a p j k -packing from Point j, for j e {1,..:, 4}; 6. three (4 x 3 k — 1)-packings that form a partition of a (6 k — l)-packing from Point 2; 7. two (4 x 4k — 1)-packings that form a partition of a (10k — 1)-packing from Point 3; 8. four 17-packings and six 23-packings that form a partition of every 5-packing from Point 2. Proof. The proof is resumed in Table B.4, this table contains: in which i-packing X will be decomposed (Column 1), the number of i-packings needed to form a partition of X (Column 2), the description of an i-packing with Cartesian coordinates (assuming x and y are integers) (Column 3) and the family of translation vectors (Column 4). We assume that if we do copies of this i-packing and we translate each one by one of these vectors. Afterward, we obtain a partition of X in i-packings. □ 3.3 Subdivision of a 4-packing in H Let X4 = {(3x + 7y, 2x + y)| x G Z, y g Z} be the 4-packing in H from Figure 2. Note that d(X4) = 1/11 and that 1/A(4) = 1/10. However, we claim that a 4-packing with density 1/10 does not exist. Note that eleven 4-packings form a partition of H if we translate X4 ten times. Proposition 3.3. Let k > 0 and m > 0 be integers. There exist: 1. k2 pi,k-packings, pi,k = (5k — 1), that form a partition of X4; 2. 2k2 p2,k -packings, p2,k = (6k — 1), that form a partition of X4; 3. 3k2 p3,k -packings, p3,k = (8k — 1), that form a partition of X4; 4. 6k2 p4,k -packings, p4,k = (11k — 1), that form a partition of X4; 5. m2 pjmk -packings that form a partition of a p j k -packing from Point j, for j G {1,..:, 4}; 6. two (5 x 2k — 1)-packings that form a partition of a (6k — 1) -packing from Point 2; 7. two (6k — 1)-packings that form a partition of a (5k — 1)-packing from Point 1; 8. three (5 x 3k — 1)-packings that form a partition of a (8k — 1)-packing from Point 2; 9. three (8k — 1)-packings that form a partition of a (5k — 1)-packing from Point 1. Proof. See Table B.5. □ 3.4 Subdivision of a 2-packing in Z2 In the square lattice, we can choose one vertex as the origin and all the other vertices will be nominated by Cartesian coordinates. In all our representations our origin (0,0) will be in the packing that we want to describe. Let X2 = {(2x + y, x + 3y)| x G Z, y G Z} be the maximized 2-packing in Z2 from Figure 4. Note that d(X2 ) = 1/A(2) = 1/5 and that five 2-packings form a partition of Z2 if we translate X2 four times. - O 2 СЧ L 3 - -n c> L 3 - -J 4> -0- - - i Figure 4: The sets X2 (2-packing), X3 (3-packing) and X4 (4-packing) in Z2. 3 2 3 4 Proposition 3.4. Let k > 0, m > 0 be integers. There exist: 1. k2 (3k — l)-packings that form a partition of X2; 2. 2k2 (4k — l)-packings that form a partition of X2; 3. two (3 x 2k — l)-packings that form a partition of a (4k — 1)-packing from Point 2; 4. two (4k — l)-packings that form a partition of a (3k — 1)-packing from Point 1; 5. m2 (3 mk — l)-packings that form a partition of a (3k — 1)-packing from Point 1; 6. m2 (4mk — 1)-packings that form a partition of a (4k — 1)-packing from Point 2. Proof. See Table B.6. □ 3.5 Subdivision of a 3-packing in Z2 Let X3 = {(2x + 4y, 2y) \ x G Z, y G Z} be the maximized 3-packing in Z2 from Figure 4. Note that d(X3) = 1/A(3) = 1/8 and that eight 3-packings form a partition of Z2 if we translate X3 seven times. Proposition 3.5. Let k > 0 and m > 0 be integers. There exist: 1. k2 (4k — 1)-packings that form a partition of X3; 2. m2 (4 mk — 1)-packings that form a partition of a (4k — 1)-packing from Point 1. Proof. See Table B.6. □ 3.6 Subdivision of a 4-packing in Z2 Let X4 = {(3x + 8y, 2x + y)\ x G Z, y G Z} be the maximized 4-packing in Z2 from Figure 4. Note that d(X4) = 1/A(4) = 1/13 and that thirteen 4-packings form a partition of Z2 if we translate X4 twelve times. Proposition 3.6. Let k > 0, m > 0 be integers. There exist: 1. k2 (5k — 1)-packings that form a partition of X4; 2. 2k2 (6k — 1)-packings that form a partition of X4; 3. two (5 X 2k — 1)-packings that form a partition of a (6k — 1)-packing from Point 2; \ \ \ \ \ 2 \ \ \ \ \ \ \ \ > \ \ \ \ \ \ \ 2 l_ Figure 5: The sets X1 (1-packing), X2 (2-packing) and X3 (3-packing) in T. 4. two (6 k — 1)-packings that form a partition of a (5k — l)-packing from Point 1; 5. m2 (5 mk — 1)-packings that form a partition of a (5k — 1)-packing from Point 1; 6. m2 (6 mk — 1)-packings that form a partition of a (6k — 1)-packing from Point 2. Proof. See Table B.6. □ 3.7 Subdivision of an independent set in T The square lattice can be seen as a subgraph of the triangular lattice. In fact in Figure 5, the triangular lattice is represented as a supergraph of the square lattice. Therefore, we can choose one vertex as the origin and all the other vertices will be nominated by Cartesian coordinates. In all our representations our origin (0,0) will be a vertex such that (0,0) has (1,0), (0,1), ( —1,0), (0, —1), ( —1,1) and (1, —1) as neighbors. Let Xi = {(x + 3y, x) | x G Z, y G Z} be the (unique) maximized independent set (1-packing) in T from Figure 5. Note that d(X1) = 1/A(1) = 1/3 and that three independent sets form a partition of T if we translate X1 two times. Proposition 3.7. Let k > 0 and m > 0 be integers. There exist: 1. k2 (2k — 1)-packings that form a partition of X1; 2. 3k2 (3k — 1)-packings that form a partition of X1; 3. three (3k — 1)-packings that form a partition of a (2k — 1)-packing from Point 1; 4. m2 (2mk — 1)-packings that form a partition of a (2k — 1) -packing from Point 1; 5. m2 (3mk — 1)-packings that form a partition of a (3k — 1) -packing from Point 2. Proof. See Table B.7. □ 3.8 Subdivision of a 2-packing in T Let X2 = {(2x — y, x + 3y ) | x G Z, y G Z} be the maximized 2-packing in T from Figure 5. Note that d(X2) = 1/A(2) = 1/7 and that seven 2-packings form a partition of T if we translate X2 six times. Proposition 3.8. Let k > 0 and m > 0 be integers. There exist: 1. k2 (3k — 1)-packings that form a partition of X2; 2. m2 (3mk — 1)-packings that form a partition of a (3k — 1) -packing from Point 1. Proof. See Table B.7. □ 3.9 Subdivision of a 3-packing in T Let X3 = {(2x + 6y, 2x) \ x G Z, y G Z} be the maximized 3-packing in T from Figure 5. Note that d(X3) = 1 / A(3) = 1/12 and that twelve 3-packings form a partition of T if we translate X3 eleven times. Proposition 3.9. Let k > 0 and m > 0 be integers. There exist: 1. k2 (4k — 1)-packings that form a partition of X3; 2. 3k2 (6k — 1)-packings that form a partition of X3; 3. three (4 x 3k — 1)-packings that form a partition of a (6k — 1)-packing from Point 1; 4. m2 (4mk — 1)-packings that form a partition of a (4k — 1)-packing from Point 1; 5. m2 (6 mk — 1)-packings that form a partition of a (6k — 1)-packing from Point 2. Proof. See Table B.7. □ 4 S-packing chromatic number 4.1 General properties In the previous section, we obtained several properties of subdivision of an i-packings in a lattice. This section illustrates general properties obtained on the S-packing chromatic number using only a small part of these properties. For a given sequence S, one can find other colorings of a lattice using properties from the previous section. Corollary 4.1. Let a0 = 1. If s1 = 2 and there exist three integers 1 < a1 < . .. < a3 and three integers k1,... ,k3 such that sai < 3ki — 1 and ai — ai-1 > k2 or sai < 4ki — 1 and ai — aj-1 > 2 k2 for i G {1,..., 3} then xS (H) < a3. This corollary can be useful to find upper bounds for a given sequence. For example, if S = (2,2,2,2,...), then taking a1 = 2, a2 = 3 and a3 = 4, Corollary 4.1 gives us XSp (H) < 4 (this result is in fact treated in next subsection). Similarly, for the sequence 5 = (2, 3, 3, 5, 5, 5, 5, 7, 7, 7, 7, 7, 7, 7, 7,...) , then taking a1 = 3, a2 =7 and a3 = 15, Corollary 4.1 gives us x'S (H) < 15. There are similar results for s1 = 3 or s1 =4 using Propositions 3.3 and 3.4. For the two remaining lattices, the two following properties are given for Z2 with s1 = 2 and for T with s1 = 1. There exist similar properties for Z2 with s1 = 3 or 4 using Propositions 3.5 and 3.6 and for T with s1 = 2 or 3 using Propositions 3.8 and 3.9. Corollary 4.2. Let a0 = 1. If s1 = 2 and there exist four integers 1 < a1 < ... < a4 and four integers k1,... ,k4 such that sai < 3ki — 1 and ai — ai-1 > k2 or sai < 4ki — 1 and ai — ai-1 > 2k2 for i G {1,..., 4} then xS(Z2) < a4. Corollary 4.3. Let a0 = 1. If s1 = 1 and there exist two integers 1 < a1 < a2 and two integers k1 and k2 such that sai < 2ki — 1 and ai — ai-1 > k2 or sai < 3ki — 1 and ai — ai-1 > 3k2 for i G {1,..., 2} then xS(T) < a2. 4.2 S-packing chromatic number and distance coloring Jacko and Jendrol [10], Fertin et al. [2] and Sevcikovä [14] have studied distance colorings of H, Z2 and T respectively. The following propositions comes from their work and can be translated in S-packing coloring: Proposition 4.4 ([10]). Let n and d be integers. The minimum n such that s1 = d, sn = d and xS (H) = n is given by \3(d +1)21 for d odd; \8(d + 4/3)2! for d even. Proposition 4.5 ([2]). Let n and d be integers. The minimum n such that s1 = d, sn = d and xS (Z2 ) = n is given by 1fJ i 1 42 , 2 (d + 1)2 for d odd; n =i f ((d +1)2 + 1) for d even. Proposition 4.6 ([14]). Let n and d be integers. The minimum n such that s1 = d, sn = d and xS (T) = n is given by n= 4(d +1)2 5 (d, n)-packing chromatic number 5.1 Hexagonal lattice Proposition 5.1. x2'1(H) = rc, xp'2(H) = rc, xp'3(H) = rc, xpM(H) = rc, XP3'5 (H)= rc and x16'6(H) = rc. Proof. Let H be the hexagonal lattice and k be an integer, k > 16. k ^ n ^ n ^ k ^ k ^ k ^ E ЛЩ = E A(2i) + E A(4i+1) + E A(4i+3) = E з42+зi+1 + E 6i2 + 6i+2 + i=1 i=1 i=0 i=0 i=1 2 2 i=0 k E 6i2+12i+6 < 125У1Бп tanh(6n^IŠ) + 1 tanh(6n^Š) + 36n2 - 1 < 1.494. i=0 + + Therefore: t АЩ < 1.494 - АЩ < °."4 < 1, E АЩ < 2(1.494 - E A1) ) < i—2 i—5 i—1 k 7 k 10 0.955 < 1 E A3) < 3(1.494-E Ä^) < о.935 < 1 E A4) < 4(1.494-E ) < i=8 i=1 i=11 i=1 k 12 k 0.925 < 1 and E Ä5) < 5(1.494 - E ащ) < 0.986 < 1, E Ащ < 6(1.494 - i=13 i=1 i=16 15 E ащ) < 0.968 < 1. Corollary 2.3 allows us to conclude. □ i=1 Proposition 5.2. xp'2(H) < 8, xp'3(H) < 5 and Vn > 4, xp'n(H) = 4. Proof. Using Proposition 3.1, we define a (2, n)-packing coloring of H for each n = 2,3 and n > 4. H can be partitioned into four 2-packings, the first two ones can be colored by color 2, the third one by two colors 3 and the last one by four colors under 5, it will be two 4 and two 5, to conclude xp'2(H) < 8. H can be partitioned into four 2-packings, the first three ones can be colored by colors 2 and the third one by two colors 3, to conclude Xp'3 (H) < 5. H can be partitioned into four 2-packings, hence Vn > 4, xp'n (H) =4. □ The following table summarizes the colorings defined in the above proof. The symbol P in the table refers to the packings we use and how we subdivide them into i-packings (Ai is an i-packing) and the symbol C refers to the associated colors we use for each i-packing. By k x Ai we mean we use k i-packings, and by k x i we mean we use k colors i. In the rest of the paper, similar proofs will be only described by a table using the same format than this one. (2,2)-packing P 2xXp Xp Xp 2xA3 4x A5 C 2x2 2x3 2x4,2x5 (2,3)-packing P 3xXp Xp 2xA3 C 3x2 2x3 Proposition 5.3. Xp 'P(H) < 35, xp '3(H) < 13, xp '4(H) < 10, xp '5(H) < 8 and Vn > 6, Xp ' n(H) = 6. Proof. Using Proposition 3.2, we define a (3, n)-packing coloring of H for each n = 2.3.4.5 and n > 6. H can be partitioned into six 3-packings, hence Vn > 6, xp ' n (H) = 6. The other colorings are described in Table C.8. □ Proposition 5.4. X4 ' 3(H) < 58, xp ' 4(H) < 27, xp ' 5(H) < 21, x4 ' 6(H) < 18 and from [10] Vn > 11, xp'n(H) = 11. Proof. Using Proposition 3.3, we define a (4, n)-packing coloring of H for each n = 3.4.5.6 and n > 11. H can be partitioned into eleven 4-packings. The other colorings are described in Table C.9. □ 5.2 Square lattice Proposition 5.5. xp '1 (ZP) = TO, xp 'P(ZP) = то, xp '3(ZP) = то, xp '4(ZP) = то, Xp0'5(ZP) = то and xpP'6(ZP) = то. Proof. Let ZP be the square lattice and k be an integer, k > 12. S A(i) = S A(Pi) + S A(Pi+1) = S Pi2+Pi+1 + Л Pi2+4i+P < 2П tanh( in) + i=1 i=1 i=0 i=1 i=0 YPnP - 1 < 1.264. Therefore: Y. A® < 1.264 - аЬ < 01764 < 1, Ž Ap) < 2(1.264 - Ž A1) ) < i=P i=4 i=1 0.877 < 1, Ž A3) < 3(1.264 - Ž АЩ ) < 01917 < 1, Ž A4) < 4(1.264 - è A1) ) < i=6 i=1 i=8 i=1 0.938 < 1, ŽC A5iy < 5(1.264 - £ ^ ) < 0.951 < 1 and £ А|у < 6(1.264 - i=10 i=1 i=1P 11 Ž А(7) ) < 0.959 < 1. Corollary 2.3 allows us to conclude. □ i=1 Proposition 5.6. Xp'P(ZP) < 20, xp'3(ZP) < 8, xp'4(ZP) < 6 and Vn > 5, xp'n(ZP) = 5. 1 2 1 3 1 2 1 10 1 4 1 7 15 16 13 12 13 13 12 14 17 15 1 4 19 13 12 13 16 1 2 1 15 1 5 1 11 1 2 1 6 1 3 1 2 1 3 1 4 1 14 1 5 1 4 1 16 1 2 1 3 1 3 12 13 16 15 17 1 7 1 10 1 2 1 3 1 2 1 2 13 15 14 18 13 14 12 13 12 19 1 3 1 6 1 13 1 7 1 3 1 2 12 13 12 15 14 1 8 15 14 13 12 13 13 1 2 1 11 1 6 1 10 1 4 17 13 12 13 15 1 2 1 17 1 5 1 4 1 2 1 51312131718 1 10 1 4 1 9 1 2 1 3 1 3 1 2 1 3 1 12 1 5 1 4 16 15 12 13 12 1 2 13 17 14 16 13 1 11 1 2 1 3 1 2 1 13 1 3 14 18 15 13 12 9 1 2 1 3 1 2 1 5 1 4 1 14 1 8 1 5 1 4 1 3 1 2 1 3 1 2 1 3 1 2 1 11 1 6 1 10 1 2 1 4 1 7 1 3 1 2 1 3 1 5 1 3 1 2 1 17 1 5 1 4 1 2 1 3 1 5 1 3 1 2 1 3 1 7 1 8 1 2 1 10 1 4 1 13 1 2 1 3 1 2 1 3 1 2 1 3 1 9 1 5 1 4 1 4 1 6 1 5 1 2 1 3 1 2 1 11 1 2 1 3 1 7 1 4 1 6 1 3 1 5 111 1 2 1 3 1 2 1 12 1 5 1 3 1 4 1 8 1 5 1 3 1 2 1 15 1 2 1 3 1 2 1 10 1 4 1 9 1 7 1 5 1 6 1 3 1 2 1 3 1 2 1 3 1 2 1 4 1 7 1 5 1 2 1 4 1 9 1 3 1 2 1 3 1 6 1 3 1 2 1 14 1 5 1 11 1 2 1 3 1 6 1 3 1 2 1 3 1 4 1 15 1 2 1 5 1 4 1 16 1 2 1 3 1 2 1 3 1 2 1 3 1 6 1 5 1 7 1 11 1 7 1 10 1 2 1 3 1 2 1 4 1 2 1 3 1 5 1 4 1 8 1 3 1 5 1 4 1 2 1 3 1 2 1 9 1 5 1 3 1 6 1 12 1 7 1 3 1 2 1 Figure 6: A 24 x 24 pattern [13]. Proof. Using Proposition 3.4, we define a (2, n)-packing coloring of Z2 for each n = 2,3,4 and n > 5. Z2 can be partitioned into five 2-packings, hence Vn > 5, xp'n(Z2) = 5. The other colorings are described in Table C.10. □ Soukal andHolub [13] have proven that xp'1(Z2) < 17, and proposed a 24 x 24 pattern in order to color the square lattice. Their pattern is recalled in Figure 6. Proposition 5.7. xp'3(Z2) < 33. Proof. In the pattern of Figure 6, Bi denotes the set of vertices colored by i. Note that B2 and B3 are both 3-packings. It can be seen that B16 U B17 form a 11-packing and that four 7-packings form a partition of B2 or B3. In order to color Z2 starting with 3, we partition 17 B1 into sixteen i-packings, 2 < i < 17 (since B1 is U Bj translated by the vector (1,0)). i=2 Let Bj denote a copy of Bi translated by (1,0). We use two colors 3 to color B2 and B3, and one color i in order to color Bi for i g [4,8]. We color Bj by one color i, for i g [3, 8] and B2 that is a 3-packing is colored by one color 4, one color 5, one color 6 and one color 7. We use the remaining color 8 to color B9. We use two colors 9 in order to color B16, B16, B17 and B17. The remaining color 9 is used to color B'g. We use two colors i in order to color Bi and Bj for i g [10,13]. The remaining colors 10, 11 ,12 and 13 are used to color B14, B14, B15 and B15. □ Proposition 5.8. x3p'4(Z2) < 20, xp'5(Z2) < 17, xp'6(Z2) < 14 andVn > 8, xp'n(Z2) = 8. Proof. Using Proposition 3.5, we define a (3, n)-packing coloring of Z2 for each n = 4,5,6 and n > 8. Z2 can be partitioned into eight 3-packings, hence Vn > 8, xp'n(Z2) = 8. The other colorings are described in Table C.11. □ Proposition 5.9. xp'4(Z2) < 56, xp'5(Z2) < 34, xp'6 (Z2) < 28 and Vn > 13, xp'n(Z2) = 13. Proof. Using Proposition 3.6, we define a (4, n)-packing coloring of Z2 for each n = 4,5,6 and n > 13. Z2 can be partitioned into thirteen 4-packings, hence Vn > 13 Xp'n(Z2) = 13. The other colorings are described in Table C.12. □ 5.3 Triangular lattice Proposition 5.10. xp'1(T) = то, x3'2(T) = то, x4'3(T) = то, xp'4(T) = то, xp'5(T) = то andxp'6(T) = то. Proof. Let T be the triangular lattice and k be an integer, k > 8. k k k k k A(i) = ^ A(2i) + 2 A(2i+1) = 2 3i2+3i+1 + 2 3i2 + 6i+3 < 3 1 i=1 i=0 < tanh( 6n\/3) + i=1 i=1 i=0 n2 - 1 < 0.854. Therefore: t Aj) < 0.854 < 1, £ A® < 2(0.854 - £ a1) ) < 0.755 < 1, £ < i=1 i=3 i=1 i=4 3 k 4 k 3(0.854 - £ Ai) ) < 0.883 < 1, £ a4) < 4(0.854 - £ A^) < 0.966 < 1, £ ^ < i=1 i=5 i=1 i=7 5(0.854 - £ A1) ) < 0.887 < 1 and £ Ащ < 6(0.854 - £ ) < 0.940 < 1. i=1 i=8 i=1 Corollary 2.3 allows us to conclude. □ Proposition 5.11. xp'2 (T) < 6 and Vn > 3, xp'n(T) = 3. Proof. Using Proposition 3.7, we define a (1, n)-packing coloring of T for each n = 2 and n > 3. T can be partitioned into three independent sets, hence Vn > 3, xp'n(T) = 3. The other coloring is described in the following table. (1,2)-packing P 2xX1 X1 4xA3 C 2x1 2x2, 2x3 □ Proposition 5.12. xp'4(T) < 16, xp'5(T) < 13, xp'6(T) < 10 and Vn > 7, xp'n(T) = 7. Proof. Using Proposition 3.8, we define a (2, n)-packing coloring of T for each n = 4, 5,6 and n > 7. T can be partitioned into seven 2-packings, hence Vn > 7, xp'n(T) = 7. The other colorings are described in Table C.13. □ Proposition 5.13. x3'4(T) < 72, x;3'5(T) < 38, xp'6(T) < 26 and Vn > 12, x;3'n(T) = 12. Proof. Using Proposition 3.9, we define a (3, n)-packing coloring of T for each n = 4,5,6 and n > 12. T can be partitioned into twelve 3-packings, hence Vn > 12, x3P'n(T) = 12. The other colorings are described in Table C.14. □ 6 Conclusion We have determined or bounded the (d, n)-packing chromatic number of three lattices H, Z2 and T for small values of d and n. Further studies can be done with other values of d and n or improving existing values. The (d, n)-packing chromatic number can also be investigated for other lattices. As an example, we can prove, using color patterns defined in [15] for distance graphs, that for the octagonal lattice O, i.e the strong product of two infinite path (which is a supergraph of T), xp'2(O) < 58. For other finite or infinite graphs, like k-regular infinite trees, the method has to be adapted or changed since a maximized packing cannot be described as easily as those considered in this paper. Also, for each of three lattices studied, finding a sequence S such that xSP = k and there is no S-packing k coloring where the s i-packing is maximized could be an interesting result. References [1] J. Ekstein, J. Fiala, P. Holub and B. Lidicky, The packing chromatic number of the square lattice is at least 12, arXiv:1003.2291v1 (2010). [2] G. Fertin, E. Godard and A. Raspaud, Acyclic and k-distance coloring of the grid, Information Processing Letters 87 (2003), 51-58. [3] J. Fiala and P. A. Golovach, Complexity of the packing coloring problem for trees, Discrete Applied Mathematics 158 (2010), 771-778. [4] J. Fiala, S. Klavžar and B. Lidicky, The packing chromatic number of infinite product graphs, Europan Journal of Combinatorics 30 (2009), 1101-1113. [5] A. S. Finbow and D. F. Rall, On the packing chromatic number of some lattices, Discrete Applied Mathematics 158 (2010), 1224-1228. [6] N. Gastineau, On dichotomies among the instance of the S-coloring problem, Discrete Mathematics 338 (2015), 1029-1041. [7] W. Goddard, S. M. Hedetniemi, S. T. Hedetniemi, J. M. Harris and D. F. Rall, Broadcast Chromatic Numbers of Graphs, Ars Combinatoria 86 (2008), 33-49. [8] W. Goddard and H. Xu, A note on S-packing colorings of lattices, Discrete Applied Mathematics 166 (2014), 255-262. [9] W. Goddard, H. Xu, The S-packing chromatic number of a graph, Discussiones Mathematicae Graph Theory 32 (2012), 795-806. [10] P. Jacko and S. Jendrol, Distance Coloring of the Hexagonal Lattice, Discussiones Mathematicae Graph Theory 25 (2005), 151-166. [11] D. Korže and A. Vesel, On the packing chromatic number of square and hexagonal lattice, Ars Mathematica Contemporanea 7 (2014), 13-22. [12] F. Kramer and H. Kramer, A survey on the distance-colouring of graphs, Discrete Mathematics 308 (2008), 422-426. [13] R. Soukal and P. Holub, A Note on Packing Chromatic Number of the Square Lattice, The Electronic Journal of Combinatorics 17 (2010), 447-468. [14] A. Sevcikovä, Distant chromatic number of the planar graphs, Manuscript (2001). [15] O. Togni, On packing Colorings of Distance Graphs, Discrete Applied Mathematics 167 (2014), 280-289. A Distances in the three lattices Definition A.1 ([10]). Let v = (a, b) be a vertex in the hexagonal lattice. Then the type of v is T (v) = a + b +1 (mod 2). As H = V0 U V1 is a bipartite graph, the type of a vertex v corresponds to the index of the set V to which v belongs. Proposition A.2 ([10]). Let v1 = (a1, b1), v2 = (a2, b2) be two vertices of the hexagonal lattice and assume that b1 > b2. Then the distance between v1 and v2 is d(v v ) f \a1 - a2\ + \b1 - b2\ if \a1 - a2\ > \b1 - b2\; d(v1,v2) = j 2\b1 - b2\- t(v1)+ t(v2) if \a1 - a2\ < \b1 - b2i. Example A.3. The set X2 from Figure 2 is a 2-packing in H. Proof. Let x and y be integers, then d((2(x + 1) + 4y, x + 1), (2x + 4y,x)) = \2x + 4y + 2 - 2x - 4y\ + \x + 1 - x\ = 3 > 2 and d((2x + 4(y + 1), x), (2x + 4y, x)) = 4 > 2; let i and j be integers, then d((2(x + i) + 4(y + j), x + i), (2x + 4y, x)) > min(d((2(x + 1) + 4y,x + 1), (2x + 4y, x)),d((2x + 4(y + 1),x), (2x + 4y,x))) = 3, hence X2 is a 2-packing. □ Claim A.4. Let v1 = (a1, b1) and v2 = (a2, b2) be two vertices of the square lattice. Then the distance between v1 and v2 is d(v1, v2) = \a1 - a2 \ + \b1 - b2 \. Example A.5. The set X2 from Figure 4 is a 2-packing in Z2. Proof. Let x and y be integers, then d((2(x+1)+y,x+1+3y), (2x+y,x+3y)) = \2x+y+2-2x-y\ + \x+1+3y-x-3y\ = 3 > 2 and d((2x + y + 1, x - 1 + 3(y + 1), (2x + y, x + 3y)) =4 > 2, to conclude X2 is a 2-packing. □ Claim A.6. Let v1 = (a1, b1) and v2 = (a2, b2) be two vertices of the triangular lattice. Then the distance between v1 and v2 is d( )=/ max(\a1 - a2\, \b1 - b2 \ ) if ((a1>a2)A(b1b2)); (v1,v2) \ \a1 - a2\ + \b1 - b2\ otherwise. Example A.7. The set X1 from Figure 5 is an independent set in T. Proof. Let x and y be integers, then, d((x + 1 + 3y, x + 1), (x + 3y, x)) = \x + 1 + 3y - x - 3y\ + \x + 1 - x\ = 2 > 1 and d((x + 3(y + 1), x), (x + 3y, x)) = 3 > 1, to conclude X1 is an independent set. □ B Decomposition of an i-packing in the three lattices i Number of i-packings Description of a i-packing Family of translation vectors 4k - 1 k2 {3kx + 6ky, kx)} (3i + 6j, i) i,j G{0,...,k-1} 6k - 1 3k2 {3kx + 6ky, 3kx)} (3i + 6j, 3i + 2a) i,j G {0,. .., k - 1}, a G {0, 1, 2} 10k - 1 8k2 {6kx +12ky,4kx)} (6i + 12j + 3b, 4i + 2a + b) i,j G{0,...,k - 1}, a G {0,1, 2, 3}, b G {0, 1} 18k - 1 24k2 {12kx + 24ky, 6kx)} (12i + 24j + 3b, 6i + 2a + b) i,j G {0,. .., k - 1}, a G {0, ..., 5}, b G {0,1, 2, 3} Table B.4: Decomposition of X3 in H into i-packings. i Number of i-packings Description of a i-packing Family of translation vectors 5k - 1 k2 {3kx - ky, 2kx + 3ky)} (3i - j, 2i + 3j) i,j G {0, ..., k - 1} 6k - 1 2k2 {7kx - ky, kx + 3ky)} (7i + 3a - j, i + 2a + 3j) i,j G {0,. .., k - 1}, a G {0,1} 8k - 1 3 й-i: {7kx + 2ky, kx + 5ky)} (7i + 2j + 3a, i + 5j + 2a) i,j G {0, ..., k - 1}, a G {0, 1, 2} 11k - 1 6k2 {-2kx + 11ky, 6kx)} (-2i + 11j + 7a, 6i + a) i, j G {0,. .., k - 1}, a G {0, ..., 5} Table B.5: Decomposition of X4 in H into i-packings. i Number of i-packings Description of a i-packing Family of translation vectors X2 3k - 1 k2 {2kx - ky, kx + 2ky)} (2i - j,i + 2j) i, j G{0,...,k - 1} 4k - 1 2 й-1: {4kx + ky, 2kx + 3ky)} (4i + 2a + j, 2i + 2a + 3j) i,j € {0, ...,k - 1}, a G {0,1} X3 4k - 1 k2 {2kx + 4ky, 2kx)} (2i + 4j, 2i) i,j G {0,. .., k - 1} X4 5k - 1 k2 {3kx - 2ky, 2kx + 3ky)} (3i - 2j, 2i + 3j) i,j G {0,. .., k - 1} 6k - 1 2 й- к: {6kx + ky, 4kx + 5ky)} (6i + j + 3a, 4i + 5j + 2a) i,j G {0, .. ., k - 1}, a G {0,1} Table B.6: Decomposition of X2, X3 and X4 in Z2 into i-packings. i Number of i-packings Description of a i-packing Family of translation vectors Xi 2k - 1 k2 {kx + 3ky, kx)} (i + 3j,i) i,j G {0,. .., k - 1} 3k - 1 3 й-i: {3kx + 3ky, 3kx)} (3i + 3j + a, 3i + a) i,j G {0,. .., k - 1}, a G {0, 1, 2} X2 3k - 1 k2 {2kx + 7ky, kx)} (2i + 7j,i) i,j G {0,. .., k - 1} X3 4k - 1 k2 {2kx + 6ky, 2kx)} (2i + 6j, 2i) i,j G {0,. .., k - 1} 6k - 1 3 к: {6kx + 6ky,6kx)} (6i + 6j + 2a, 6i + 2a) i,j G {0,. .., k - 1}, a G {0, 1, 2} Table B.7: Decomposition of Xb X2 and X3 in T into i-packings. C Decomposition and associated colors X3 X3 X3 X3 P 2xX3 3xX5 4xX7 4xX9, 8xX15 X5,3xX11, (3,2)-packing 4xX17, 6xX23 C 2x3 2x4,5 2x6, 2x7 2x8, 2x9, 2x12,2x13, 2x14, 2x15 5,2x10,2x11, 2x16,2x17,2x18, 2x19, 20 P 3xX3 X3 X3 X3 (3,3)-packing 3xX5 3xX5 4xX7 C 3x3 3x4 3x5 3x6,7 P 4xX3 X3 X3 (3,4)-packing 3xX5 3xX5 C 4x3 3x4 4,2x5 P 5xX3 X3 (3,5)-packing 3xX5 C 5x3 3x4 Table C.8: Decomposition of H into 3-packings and associated colors. (4,3)-packing P 3xX4 X4 2xX4 2xX4 X4 2xA5 6xA7 A5, 6xAg 6xA11,4xA19 C 3x4 2x5 3x6, 3x7 5, 3x8, 3x9 3x10, 3x11, 18, 3x19 P X4 X4 9xAM 11xA19, 10xA23 C 3x12, 3x13, 3x14 3x15, 3x16, 3x17, 2x18, 3x20, 3x21 3x22, 23 (4,4)-packing P 4xX4 2xX4 2xX4 2xX4 X4 4xA5 6xA7 8xA9 2xA7, 3xA14 C 4x4 4x5 4x6, 2x7 4x8,4x9 2x7, 3x10 (4,5)-packing P 5xX4 2xX4 3xX4 X4 4x A5 9xAr A5,2xA9 C 5x4 4x5 5x6,4x7 5, 7,8 (4,6)-packing P 6xX4 3xX4 2xX4 6x A5 6xA7 C 6x4 6x5 6x6 Table C.9: Decomposition of H into 4-packings and associated colors. (2,2)-packing P 2xX2 X2 X2 X2 2xA3 4x A5 6xA8, 6xA11 C 2x2 2x3 2x4,2x5 2x6,2x7,2x8 2x9,2x10,2x11 (2,3)-packing P 3xX2 X2 X2 2xA3 A3,2xA5 C 3x2 2x3 3,2x4 (2,4)-packing P 4xX2 X2 2xA3 C 4x2 2x3 Table C.10: Decomposition of Z2 into 2-packings and associated colors. (3,4)-packing P 4xX3 4xX3 16xA7 C 4x3 4x4, 4x5, 4x6, 4x7 (3,5)-packing P 5xX3 3xX3 12xA7 C 5x3 5x4, 5x5, 2x6 (3,6)-packing P 6xX3 2xX3 8xA7 C 6x3 6x4,2x5 Table C.11: Decomposition of Z2 into 3-packings and associated colors. P 4xX4 2xX4 4xX4 X4 X4 X4 (4,4)-packing 4xA5 16x A9 8xA11 9xA14 3xA14, 12xA17 C 4x4 4x5 4x6, 4x7, 4x10, 4x12 3x14,4x15, 4x8,4x9 4x11 4x13, 14 4x16,4x17 5xX4 2xX4 5xX4 X4 P 4xA5 A5, 18xA9 2xA9, (4,5)-packing 4xA11 C 5x4 4x5 5, 5x6, 5x7 2x9, 5x8, 3x9 4x10 P 6xX4 3xX4 4xX4 (4,6)-packing 6xA5 16xA9 C 6x4 6x5 6x6,6x7,4x8 Table C.12: Decomposition of Z2 into 4-packings and associated colors. (2,4)-packing P 4xX2 3xX2 12xA5 C 4x2 4x3,4x4,4x5 (2,5)-packing P 5xX2 2xX2 8Xa5 C 5x2 5x3, 3x4 (2,6)-packing P 6xX2 x2 4x a5 C 6x2 4x3 Table C.13: Decomposition of T into 2-packings and associated colors. (3,4)-packing P 4xXf 2xXf 2xXf 2xXf 6XA5 8XA7 2xA5, 12xA11 C 4x3 4x4,2x5 4x6, 4x7 2x5,4x9,4x10,4x11 P Xf Xf 16XA15 4xA11,20xA23 C 4x12,4x13 4x14,4x15 4x8,4x16,4x17, 4x18,4x19,4x20 (3,5)-packing P 5xXf 3xXf 2xXf 2xXf 9xA5 8XA7 18XA11 C 5x3 5x4,4x5 5x6, 3x7 5, 2x7, 5x8, 5x9, 3x10 (3,6)-packing P 6xXf 4xXf 2xXf 12xA5 8xA7 C 6x3 6x4, 6x5 6x6, 2x7 Table C.14: Decomposition of T into 3-packings and associated colors. ARS MATHEMATICA CONTEMPORANEA Author Guidelines Papers should be prepared in LTeX and submitted as a PDF file. Articles which are accepted for publication have to be prepared in LTeX using class file amcjou.cls and bst file amcjou.bst (if you use BibTEX). These files and an example of how to use the class file can be found at http://amc-journal.eu/index.php/amc/about/submissions#authorGuidelines If this is not possible, please use the default LTeX article style, but note that this may delay the publication process. Title page. The title page of the submissions must contain: • Title. The title must be concise and informative. • Author names and affiliations. For each author add his/her affiliation which should include the full postal address and the country name. If available, specify the e-mail address of each author. Clearly indicate who is the corresponding author of the paper. • Abstract. A concise abstract is required. The abstract should state the problem studied and the principal results proven. • Keywords. Please specify 2 to 6 keywords separated by commas. • Mathematics Subject Classification. Include one or more Math. Subj. Class. 2010 codes - see http://www.ams.org/msc. References. References should be listed in alphabetical order by the first author's last name and formatted as it is shown below: [1] First A. Author, Second B. Author and Third C. Author, Article title, Journal Title 121 (1982), 1-100. [2] First A. Author, Book title, third ed., Publisher, New York, 1982. [3] First A. Author and Second B. Author, Chapter in an edited book, in: First Editor, Second Editor (eds.), Book Title, Publisher, Amsterdam, 1999, 232-345. Illustrations. Any illustrations included in the paper must be provided in PDF or EPS format. Make sure that you use uniform lettering and sizing of the text. ARS MATHEMATICA CONTEMPORANEA Subscription Yearly subscription: 150 EUR Any author or editor that subscribes to the printed edition will receive a complimentary copy of Ars Mathematica Contemporanea. Subscription Order Form Name: .............................. E-mail: ............................. Postal Address: ...................... I would like to subscribe to receive ...... copies of each issue of Ars Mathematica Contemporanea in the year 2015. I want to renew the order for each subsequent year if not cancelled by e-mail: □ Yes □ No Signature: Please send the order by mail, by fax or by e-mail. By mail: Ars Mathematica Contemporanea UP FAMNIT Glagoljaška 8 SI-6000 Koper Slovenia By fax: +386 5 611 75 71 By e-mail: info@famnit.upr.si ARS MATHEMATICA CONTEMPORANEA HERE'S TO JACK E. GRAVER ON HIS 80th BIRTHDAY Brigitte Servatius Jack the educator. The Syracuse Mathematics Department is housed in Carnegy Library. Imagine heavy snowfall and a slim, tall figure approaching the Carnegy building, propping the heavy door open with one knee while his hands are busy shaking the snow off the Daily Orange (SU student newspaper) and the figure seemingly freezing in this position for as long as it takes to read the front page article. For SU students this is a familiar image of Jack E. Graver, always interested in students and always finding time to read their news. Jack the administrator. A surprisingly deep dark voice resonating from the chairman's office inside Carnegie singing "Nobody knows the trouble I have seen, nobody knows my sorrows" is another image of Jack E. Graver. Jack the researcher. The mathematician Graver is described on MathSciNet by publication in the areas of biology and other natural sciences, combinatorics, convex and discrete geometry, game theory, economics, social and behavioral sciences, geometry manifolds and cell complexes, operations research, mathematical programming, probability theory and stochastic processes. Graver received his Ph.D. from Indiana University in 1964 under Andrew Hugh Wallace. His dissertation's title was "An Analytic Triangulation of an Arbitrary Real Analytic Variety", the field algebraic topology. It is remarkable that the 1964 paper [3] of the same title is cited more than once in this millennium. In 1966 he cashed an Erdos check for results in [8], a paper that may be called his first major work (cited more than 10 times in this millennium!). Because of this paper, written with Jim Yackel at Dartmouth (John Wesley Young Research Instructor), Jack Graver is known as a Ramsey Theorist. From 1966 to present, he teaches at Syracuse University. Why Syracuse? Because I felt at home here, he says. In 1975 he published On the foundations of linear and integer linear programming I [4], a paper that turned his name into an adjective. On Wickipedia you can read up on Graver bases, but if you want a more reliable source, try [9], where the relationship of Grobner bases to Hilbert bases and Graver bases is presented. If you look for On the foundations of linear and integer linear programming II, you need to read Amir Fou-rudi's thesis, one of Jack Graver's 9 (so far) Ph.D. students - Jack himself was handicapped by his 1977-1994 chairmanship. However, despite administrative duties, his collaboration with Mark Watkins turned him to Graph Theory, with [6, 10, 7] as major contributions to the field. A fortunate diversion into Architecture produced not only a novel and unusual text [1], but awoke Jack's interest in rigidity of frameworks and he promoted matroids as a major tool in a colloquium talk that changed my life. I became his Ph.D. student, one of four simultaneously supervised by an acting chair. We did not fight over mathematical issues as Jack had all four of us work on totally disjoint topics, we merely fought for being next in line to enter the chairman's office. Combinatorial rigidity [2] (100 citations) is to this very day the focus of my research. ARS MATHEMATICA CONTEMPORANEA Meanwhile Jack has moved on to new Ph.D. students and new topics, namely, Fullerenes [5]. When asked about retirement he justifies the non-existence of plans by stating with a smile: "Normally people retire around 65, but next year will be only my 50th year at SU, so I have some time to think about that." There's more to life than math—there is family, scouting, Shakespeare, gardening, fine dining, stories. Whenever you tell him a story, Jack will tell you a funnier and better one than you have ever heard. We have told but a small part of an interesting life to which we may look up to for future inspiration. References [1] J. A. Baglivo and J. E. Graver, Incidence and symmetry in design and architecture, volume 7 of Cambridge Urban & Architectural Studies, Cambridge University Press, Cambridge, 1983. [2] J. Graver, B. Servatius and H. Servatius, Combinatorial rigidity, volume 2 of Graduate Studies in Mathematics, American Mathematical Society, Providence, RI, 1993, doi:10.1090/gsm/002. [3] J. E. Graver, An analytic triangulation of an arbitrary real analytic variety, J. Math. Mech. 13 (1964), 10211036. [4] J. E. Graver, On the foundations of linear and integer linear programming. I, Math. Programming 9 (1975), 207-226. [5] J. E. Graver and C. M. Graves, Fullerene patches. I, Ars Math. Contemp. 3 (2010), 109-120. [6] J. E. Graver and M. E. Watkins, Combinatorics with emphasis on the theory of graphs, Springer-Verlag, New York-Berlin, 1977, graduate Texts in Mathematics, Vol. 54. [7] J. E. Graver and M. E. Watkins, Locally finite, planar, edge-transitive graphs, Mem. Amer. Math. Soc. 126 (1997), vi+75, doi:10.1090/memo/0601. [8] J. E. Graver and J. Yackel, An upper bound for Ramsey numbers, Bull. Amer. Math. Soc. 72 (1966), 10761079. [9] B. Sturmfels, Algebraic recipes for integer programming, in: Trends in optimization, Amer. Math. Soc., Providence, RI, volume 61 of Proc. Sympos. Appl. Math., pp. 99-113, 2004, doi:10.1090/psapm/061/ 2104733. [10] M. E. Watkins and J. E. Graver, A characterization of infinite planar primitive graphs, J. Combin. Theory Ser. B 91 (2004), 87-104, doi:10.1016/j.jctb.2003.10.005. Mathematical Sciences, Worcester Polytechnic Institute, Worcester MA 01609-2280 E-mail address: bservat@wpi.edu Printed in Slovenia by Birografika Bori d.o.o. Ljubljana