{"id":2290,"date":"2025-10-24T01:43:11","date_gmt":"2025-10-23T23:43:11","guid":{"rendered":"https:\/\/www.pauljorion.com\/blog_en\/?p=2290"},"modified":"2025-10-24T01:43:11","modified_gmt":"2025-10-23T23:43:11","slug":"comparison-between-google-s2r-pribors-combinatorial-magic-and-pribors-che-contextual-hyper-embedding","status":"publish","type":"post","link":"https:\/\/www.pauljorion.com\/blog_en\/2025\/10\/24\/comparison-between-google-s2r-pribors-combinatorial-magic-and-pribors-che-contextual-hyper-embedding\/","title":{"rendered":"<b>Comparison between Google S2R, Pribor\u2019s Combinatorial Magic and Pribor\u2019s CHE (Contextual Hyper-Embedding)<\/b>"},"content":{"rendered":"<p class=\"p1\"><b>Comparison between Google S2R, Pribor\u2019s Combinatorial Magic and Pribor\u2019s CHE (Contextual Hyper-Embedding)<\/b><\/p>\n<p class=\"p1\">This document presents the characteristics, divergences and synergies between three approaches: Google S2R, Pribor\u2019s Combinatorial Magic and Pribor\u2019s CHE (Contextual Hyper-Embedding).<\/p>\n<p class=\"p1\"><b>1. Google\u2019s S2R *<\/b><\/p>\n<p class=\"p1\">\u201cS2R\u201d means Speech-to-Retrieval. It is a recent voice search architecture that Google is deploying, which bypasses the explicit speech \u2192 text transcription step to try to directly establish a match between the spoken audio and the information sought. The model relies on a dual encoder: one processes the audio, the other the candidate texts, in order to bring their vector representations closer together in the same semantic space.<\/p>\n<p class=\"p1\"><b>2. <a href=\"https:\/\/www.pauljorion.com\/blog_en\/2025\/10\/03\/combinatorial-magic-logic-proof-of-concept-2\/\" target=\"_blank\" rel=\"noopener\">Pribor\u2019s Combinatorial Magic<\/a><\/b><\/p>\n<p class=\"p1\">Combinatorial Magic is a bijective, lossless and fixed-dimensional encoding of simple sentences into 4D or 5D vectors: three symbolic components (Subject, Verb, Object) plus a \u201cmeta\u201d register of 8 or 16 bits. It is distinguished by O(1) complexity, total absence of information loss, and perfect interpretability.<\/p>\n<p class=\"p1\"><b>3. <a href=\"https:\/\/www.pauljorion.com\/blog_en\/2025\/10\/18\/pribor-che-contextual-hyper-embedding-uint8\/\" target=\"_blank\" rel=\"noopener\">CHE (Contextual Hyper-Embedding uint8)<\/a><\/b><\/p>\n<p class=\"p1\">CHE is an extremely economical contextual encoding approach, which represents each token by a uint8 integer. Unlike the floating-point attention of Transformers, it avoids matrices and softmax, reducing energy consumption by a factor of up to 5000.<\/p>\n<p class=\"p1\"><b>4. Comparative Table<\/b><\/p>\n<table class=\"t1\" cellspacing=\"0\" cellpadding=\"0\">\n<tbody>\n<tr>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">Feature<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">S2R (Google)<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">Combinatorial Magic<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">CHE<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td class=\"td2\" valign=\"top\">\n<p class=\"p1\">Data type<\/p>\n<\/td>\n<td class=\"td2\" valign=\"top\">\n<p class=\"p1\">float16 \/ float32<\/p>\n<\/td>\n<td class=\"td2\" valign=\"top\">\n<p class=\"p1\">symbolic indices + meta uint8<\/p>\n<\/td>\n<td class=\"td2\" valign=\"top\">\n<p class=\"p1\">uint8<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">Dimension<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">512\u20134096D<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">4D \/ 5D<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">1 byte\/token<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">Complexity<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">O(n\u00b2)<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">O(1)<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">O(n) linear<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">Information loss<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">with loss<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">none<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">bounded \/ quantized<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td class=\"td3\" valign=\"top\">\n<p class=\"p1\">Energy efficiency<\/p>\n<\/td>\n<td class=\"td3\" valign=\"top\">\n<p class=\"p1\">low<\/p>\n<\/td>\n<td class=\"td3\" valign=\"top\">\n<p class=\"p1\">extreme<\/p>\n<\/td>\n<td class=\"td3\" valign=\"top\">\n<p class=\"p1\">extreme (\u00d7500\u20135000)<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">Interpretability<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">low<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">total<\/p>\n<\/td>\n<td class=\"td1\" valign=\"top\">\n<p class=\"p1\">medium<\/p>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p class=\"p1\"><b>5. Synergies and Integration<\/b><\/p>\n<p class=\"p1\">The three approaches can be integrated into a hybrid architecture: S2R provides the global semantic geometry, CHE ensures contextual efficiency through uint8 quantisation, and Combinatorial Magic formalises symbolic propositions without loss. This combination gives rise to a family of S2R\u2013CHE\u2013CM models combining semantic generalisation, energy frugality and complete interpretability.<\/p>\n<p class=\"p1\">* Ehsan Variani and Michael Riley, Research Scientists, Google Research, <a href=\"https:\/\/research.google\/blog\/speech-to-retrieval-s2r-a-new-approach-to-voice-search\/\"><span class=\"s1\">&#8220;Speech-to-Retrieval (S2R): A new approach to voice search&#8221;<\/span><\/a>, October 7, 2025<\/p>\n","protected":false},"excerpt":{"rendered":"<p class=\"p1\"><b>Comparison between Google S2R, Pribor\u2019s Combinatorial Magic and Pribor\u2019s CHE (Contextual Hyper-Embedding)<\/b><\/p>\n<p class=\"p1\">This document presents the characteristics, divergences and synergies between three approaches: Google S2R, Pribor\u2019s Combinatorial Magic and Pribor\u2019s CHE (Contextual Hyper-Embedding).<\/p>\n<p class=\"p1\"><b>1. Google\u2019s S2R *<\/b><\/p>\n<p class=\"p1\">\u201cS2R\u201d means Speech-to-Retrieval. It is a recent voice search architecture that Google is deploying, [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_crdt_document":"","footnotes":""},"categories":[3,12],"tags":[562,559,564,563],"class_list":["post-2290","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-human-complex-systems","tag-che-contextual-hyper-embedding","tag-combinatorial-magic","tag-pribor-ai","tag-s2r"],"_links":{"self":[{"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/posts\/2290","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/comments?post=2290"}],"version-history":[{"count":2,"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/posts\/2290\/revisions"}],"predecessor-version":[{"id":2292,"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/posts\/2290\/revisions\/2292"}],"wp:attachment":[{"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/media?parent=2290"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/categories?post=2290"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.pauljorion.com\/blog_en\/wp-json\/wp\/v2\/tags?post=2290"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}