{"id":1352,"date":"2025-04-13T22:48:58","date_gmt":"2025-04-13T22:48:58","guid":{"rendered":"https:\/\/techtrendfeed.com\/?p=1352"},"modified":"2025-04-13T22:48:58","modified_gmt":"2025-04-13T22:48:58","slug":"sesame-speech-mannequin-how-this-viral-ai-mannequin-generates-human-like-speech","status":"publish","type":"post","link":"https:\/\/techtrendfeed.com\/?p=1352","title":{"rendered":"Sesame\u200a Speech Mannequin: \u200aHow This Viral AI Mannequin Generates Human-Like Speech"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p class=\"wp-block-paragraph\"> revealed a demo of their newest Speech-to-Speech mannequin. A conversational AI agent who&#8217;s <em>actually<\/em> good at talking, they supply related solutions, they communicate with expressions, and truthfully, they&#8217;re simply very enjoyable and interactive to play with.<\/p>\n<p class=\"wp-block-paragraph\"><em>Word {that a} technical paper just isn&#8217;t out but, however they do have a <\/em><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/www.sesame.com\/research\/crossing_the_uncanny_valley_of_voice\" target=\"_blank\" rel=\"noreferrer noopener\"><em>quick weblog put up<\/em><\/a><em> that gives numerous details about the methods they used and former algorithms they constructed upon.<\/em>\u00a0<\/p>\n<p class=\"wp-block-paragraph\">Fortunately, they offered sufficient info for me to jot down this text and make a <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/youtu.be\/ThG9EBbMhP8\" target=\"_blank\" rel=\"noreferrer noopener\">YouTube video<\/a> out of it. Learn on!<\/p>\n<h3 class=\"wp-block-heading\">Coaching a Conversational Speech\u00a0Mannequin<\/h3>\n<p class=\"wp-block-paragraph\">Sesame is a <strong>Conversational Speech Mannequin<\/strong>, or a CSM. It inputs each textual content and audio, and generates speech as audio. Whereas they haven\u2019t revealed their coaching knowledge sources within the articles, we are able to nonetheless attempt to take a stable guess. The <i>weblog put up<\/i> closely cites one other CSM, <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/moshi.chat\/\" target=\"_blank\" rel=\"noreferrer noopener\">2024\u2019s Moshi<\/a>, and thankfully, the creators of Moshi did reveal their knowledge sources of their <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2410.00037\" target=\"_blank\" rel=\"noreferrer noopener\">paper<\/a>. Moshi makes use of <em>7 million hours<\/em> of unsupervised speech knowledge, <em>170 hours<\/em> of pure and scripted conversations (for multi-stream coaching), and <em>2000 extra hours<\/em> of phone conversations (The Fischer Dataset).<\/p>\n<hr class=\"wp-block-separator has-alpha-channel-opacity is-style-dotted\"\/>\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/contributor.insightmediagroup.io\/wp-content\/uploads\/2025\/04\/Abstract-1024x1004.png\" alt=\"\" class=\"wp-image-601440\"\/><figcaption class=\"wp-element-caption\">Sesame builds upon the <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2410.00037\" target=\"_blank\" rel=\"noreferrer noopener\">Moshi Paper<\/a> (2024)<\/figcaption><\/figure>\n<h2 class=\"wp-block-heading\">However what does it actually take to generate\u00a0audio?<\/h2>\n<p class=\"wp-block-paragraph\"><strong>In uncooked type, audio is only a lengthy sequence of amplitude values<\/strong>\u200a\u2014\u200aa waveform. For instance, in case you\u2019re sampling audio at 24 kHz, you might be capturing 24,000 float values each second.<\/p>\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/contributor.insightmediagroup.io\/wp-content\/uploads\/2025\/04\/7-1024x384.png\" alt=\"\" class=\"wp-image-601448\"\/><figcaption class=\"wp-element-caption\">There are 24000 values right here to symbolize 1 second of speech! (Picture generated by creator)<\/figcaption><\/figure>\n<p class=\"wp-block-paragraph\"><strong>After all, it&#8217;s fairly resource-intensive to course of 24000 float values for only one second of information<\/strong>, particularly as a result of transformer computations scale quadratically with sequence size. It will be nice if we might compress this sign and scale back the variety of samples required to course of the audio.<\/p>\n<p class=\"wp-block-paragraph\">We&#8217;ll take a deep dive into the <strong>Mimi encoder<\/strong> and particularly <strong>Residual Vector Quantizers (RVQ)<\/strong>, that are the spine of Audio\/Speech modeling in <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/towardsdatascience.com\/tag\/deep-learning\/\" title=\"Deep Learning\">Deep Studying<\/a> in the present day. We&#8217;ll finish the article by studying about how Sesame generates audio utilizing its particular dual-transformer structure.<\/p>\n<h3 class=\"wp-block-heading\">Preprocessing audio<\/h3>\n<p class=\"wp-block-paragraph\">Compression and have extraction are the place convolution helps us. Sesame makes use of the Mimi speech encoder to course of audio. <strong>Mimi was launched within the aforementioned <\/strong><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2410.00037\" target=\"_blank\" rel=\"noreferrer noopener\"><strong>Moshi paper<\/strong><\/a><strong> as properly<\/strong>. Mimi is a self-supervised audio encoder-decoder mannequin that converts audio waveforms into discrete \u201clatent\u201d tokens first, after which reconstructs the unique sign. Sesame solely makes use of the encoder part of Mimi to tokenize the enter audio tokens. Let\u2019s learn the way.<\/p>\n<p class=\"wp-block-paragraph\">Mimi inputs the uncooked speech waveform at 24Khz, passes them by way of a number of strided convolution layers to downsample the sign, with a stride issue of 4, 5, 6, 8, and a pair of. Which means the primary CNN block downsamples the audio by 4x, then 5x, then 6x, and so forth. In the long run, it downsamples by an element of 1920, lowering it to only 12.5 frames per second.<\/p>\n<p class=\"wp-block-paragraph\">The convolution blocks additionally undertaking the unique float values to an embedding dimension of 512. Every embedding aggregates the native options of the unique 1D waveform. 1 second of audio is now represented as round 12 vectors of measurement 512. This manner, Mimi reduces the sequence size from 24000 to only 12 and converts them into dense steady vectors.<\/p>\n<figure class=\"wp-block-image alignwide size-large\"><img decoding=\"async\" src=\"https:\/\/contributor.insightmediagroup.io\/wp-content\/uploads\/2025\/04\/1-1024x176.png\" alt=\"\" class=\"wp-image-601441\"\/><figcaption class=\"wp-element-caption\">Earlier than making use of any quantization, the Mimi Encoder downsamples the enter 24KHz audio by 1920 occasions, and embeds it into 512 dimensions. In different phrases, you get 12.5 frames per second with every body as a 512-dimensional vector. <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/youtu.be\/ThG9EBbMhP8\" target=\"_blank\" rel=\"noreferrer noopener\">(Picture from creator\u2019s video)<\/a><\/figcaption><\/figure>\n<h3 class=\"wp-block-heading\">What&#8217;s Audio Quantization?<\/h3>\n<p class=\"wp-block-paragraph\">Given the continual embeddings obtained after the convolution layer, we need to tokenize the enter speech. <strong>If we are able to symbolize speech as a sequence of tokens, we are able to apply normal language studying transformers to coach generative fashions.<\/strong><\/p>\n<p class=\"wp-block-paragraph\">Mimi makes use of a <strong>Residual Vector Quantizer or RVQ tokenizer<\/strong> to realize this. We&#8217;ll speak concerning the residual half quickly, however first, let\u2019s have a look at what a easy vanilla Vector quantizer does.<\/p>\n<h4 class=\"wp-block-heading\">Vector Quantization<\/h4>\n<p class=\"wp-block-paragraph\">The concept behind Vector Quantization is easy: you prepare a codebook\u200a,\u200awhich is a set of, say, 1000 random vector codes all of measurement 512 (identical as your embedding dimension).<\/p>\n<figure class=\"wp-block-image alignwide size-large\"><img decoding=\"async\" src=\"https:\/\/contributor.insightmediagroup.io\/wp-content\/uploads\/2025\/04\/2-1-1024x646.png\" alt=\"\" class=\"wp-image-601442\"\/><figcaption class=\"wp-element-caption\">A Vanilla Vector Quantizer. A codebook of embeddings is educated. Given an enter embedding, we map\/quantize it to the closest codebook entry. <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/youtu.be\/ThG9EBbMhP8\" target=\"_blank\" rel=\"noreferrer noopener\">(Screenshot from creator\u2019s\u00a0video)<\/a><\/figcaption><\/figure>\n<p class=\"wp-block-paragraph\">Then, given the enter vector, we are going to map it to the closest vector in our codebook\u200a\u2014\u200aprincipally snapping some extent to its nearest cluster heart. This implies we&#8217;ve got successfully created a hard and fast vocabulary of tokens to symbolize every audio body, as a result of regardless of the enter body embedding could also be, we are going to symbolize it with the closest cluster centroid. If you wish to study extra about Vector Quantization, take a look at my video on this subject the place I am going a lot deeper with this.<\/p>\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\">\n<div class=\"jeg_video_container jeg_video_content\"><iframe loading=\"lazy\" title=\"If LLMs are text models, how do they generate images?\" width=\"500\" height=\"281\" src=\"https:\/\/www.youtube.com\/embed\/EzDsrEvdgNQ?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/div>\n<figcaption class=\"wp-element-caption\">Extra about Vector Quantization! (Video by creator)<\/figcaption><\/figure>\n<h4 class=\"wp-block-heading\">Residual Vector Quantization<\/h4>\n<p class=\"wp-block-paragraph\">The issue with easy vector quantization is that the lack of info could also be too excessive as a result of we&#8217;re mapping every vector to its cluster\u2019s centroid. This <em>\u201csnap\u201d<\/em> is never excellent, so there&#8217;s all the time an error between the unique embedding and the closest codebook.<\/p>\n<p class=\"wp-block-paragraph\">The massive concept of <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/towardsdatascience.com\/tag\/residual-vector-quantization\/\" title=\"Residual Vector Quantization\">Residual Vector Quantization<\/a> is that it doesn\u2019t cease at having only one codebook. As a substitute, it tries to make use of a number of codebooks to symbolize the enter vector.<\/p>\n<ol class=\"wp-block-list\">\n<li class=\"wp-block-list-item\"><strong>First<\/strong>, you quantize the unique vector utilizing the primary codebook.<\/li>\n<li class=\"wp-block-list-item\"><strong>Then<\/strong>, you subtract that centroid out of your authentic vector. What you\u2019re left with is the <strong>residual<\/strong>\u200a\u2014\u200athe error that wasn\u2019t captured within the first quantization.<\/li>\n<li class=\"wp-block-list-item\">Now take this residual, and <strong>quantize it once more<\/strong>, utilizing a <strong>second codebook full of name new code vectors\u200a<\/strong>\u2014\u200aonce more by snapping it to the closest centroid.<\/li>\n<li class=\"wp-block-list-item\">Subtract <em>that<\/em> too, and also you get a smaller residual. Quantize once more with a 3rd codebook\u2026 and you&#8217;ll preserve doing this for as many codebooks as you need.<\/li>\n<\/ol>\n<figure class=\"wp-block-image alignwide size-large\"><img decoding=\"async\" src=\"https:\/\/contributor.insightmediagroup.io\/wp-content\/uploads\/2025\/04\/3-1-1024x485.png\" alt=\"\" class=\"wp-image-601443\"\/><figcaption class=\"wp-element-caption\">Residual Vector Quantizers (RVQ) hierarchically encode the enter embeddings through the use of a brand new codebook and VQ layer to symbolize the earlier codebook\u2019s error. (Illustration by the creator)<\/figcaption><\/figure>\n<p class=\"wp-block-paragraph\">Every step hierarchically captures somewhat extra element that was missed within the earlier spherical. Should you repeat this for, let\u2019s say, N codebooks, you get a set of N discrete tokens from every stage of quantization to symbolize one audio body.<\/p>\n<p class=\"wp-block-paragraph\">The best factor about RVQs is that they&#8217;re designed to have a excessive inductive bias in direction of capturing probably the most important content material within the very first quantizer. Within the subsequent quantizers, they study an increasing number of fine-grained options. <\/p>\n<p class=\"wp-block-paragraph\">Should you\u2019re accustomed to PCA, you&#8217;ll be able to consider the primary codebook as containing the first principal elements, capturing probably the most important info. The next codebooks symbolize higher-order elements, containing info that provides extra particulars.<\/p>\n<figure class=\"wp-block-image alignwide size-large\"><img decoding=\"async\" src=\"https:\/\/contributor.insightmediagroup.io\/wp-content\/uploads\/2025\/04\/4-1-1024x424.png\" alt=\"\" class=\"wp-image-601444\"\/><figcaption class=\"wp-element-caption\">Residual Vector Quantizers (RVQ) makes use of a number of codebooks to encode the enter vector\u200a\u2014\u200aone entry from every codebook. <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/youtu.be\/ThG9EBbMhP8\" rel=\"noreferrer noopener\" target=\"_blank\">(Screenshot from creator\u2019s\u00a0video)<\/a><\/figcaption><\/figure>\n<h4 class=\"wp-block-heading\">Acoustic vs Semantic Codebooks<\/h4>\n<p class=\"wp-block-paragraph\">Since Mimi is educated on the duty of audio reconstruction, the encoder compresses the sign to the discretized latent area, and the decoder reconstructs it again from the latent area. When optimizing for this process, the RVQ codebooks study to seize the important acoustic content material of the enter audio contained in the compressed latent area.\u00a0<\/p>\n<p class=\"wp-block-paragraph\">Mimi additionally individually trains a single codebook (vanilla VQ) that solely focuses on embedding the semantic content material of the audio. This is the reason <strong>Mimi is known as a split-RVQ tokenizer<\/strong> \u2013 it divides the quantization course of into two unbiased parallel paths: one for semantic info and one other for acoustic info.<\/p>\n<figure class=\"wp-block-image alignwide size-large\"><img decoding=\"async\" src=\"https:\/\/contributor.insightmediagroup.io\/wp-content\/uploads\/2025\/04\/5-1-1024x456.png\" alt=\"\" class=\"wp-image-601445\"\/><figcaption class=\"wp-element-caption\">The Mimi Structure (Supply: Moshi paper) License: Free<\/figcaption><\/figure>\n<p class=\"wp-block-paragraph\">To coach semantic representations, Mimi used information distillation with an present speech mannequin referred to as WavLM as a semantic trainer. Mainly, Mimi introduces an extra loss operate that decreases the cosine distance between the semantic RVQ code and the WavLM-generated embedding.<\/p>\n<hr class=\"wp-block-separator has-alpha-channel-opacity is-style-dotted\"\/>\n<h2 class=\"wp-block-heading\">Audio Decoder<\/h2>\n<p class=\"wp-block-paragraph\">Given a dialog containing textual content and audio, we first convert them right into a sequence of token embeddings utilizing the textual content and audio tokenizers. This token sequence is then enter right into a transformer mannequin as a time collection. Within the weblog put up, this mannequin is known as the Autoregressive Spine Transformer. Its process is to course of this time collection and output the \u201czeroth\u201d codebook token.<\/p>\n<p class=\"wp-block-paragraph\">A lighterweight transformer referred to as the audio decoder then reconstructs the subsequent codebook tokens conditioned on this zeroth code generated by the spine transformer. Word that the zeroth code already incorporates numerous details about the historical past of the dialog because the spine transformer has visibility of all the previous sequence. <strong>The light-weight audio decoder solely operates on the zeroth token and generates the opposite N-1<\/strong> codes. These codes are generated through the use of N-1 distinct linear layers that output the chance of selecting every code from their corresponding codebooks.\u00a0<\/p>\n<p class=\"wp-block-paragraph\">You possibly can think about this course of as predicting a textual content token from the vocabulary in a text-only LLM. Simply {that a} text-based LLM has a single vocabulary, however the RVQ-tokenizer has a number of vocabularies within the type of the N codebooks, so you could prepare a separate linear layer to mannequin the codes for every.<\/p>\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/contributor.insightmediagroup.io\/wp-content\/uploads\/2025\/04\/8-1024x576.jpg\" alt=\"\" class=\"wp-image-601537\"\/><figcaption class=\"wp-element-caption\">The Sesame Structure (Illustration by the creator)<\/figcaption><\/figure>\n<p class=\"wp-block-paragraph\">Lastly, after the codewords are all generated, we mixture them to type the mixed steady audio embedding. The ultimate job is to transform this audio again to a waveform. For this, we apply transposed convolutional layers to upscale the embedding again from 12.5 Hz again to KHz waveform audio. Mainly, reversing the transforms we had utilized initially throughout audio preprocessing.<\/p>\n<h3 class=\"wp-block-heading\">In Abstract<\/h3>\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\">\n<div class=\"jeg_video_container jeg_video_content\"><iframe loading=\"lazy\" title=\"Sesame AI and RVQs - the network architecture behind VIRAL speech models\" width=\"500\" height=\"281\" src=\"https:\/\/www.youtube.com\/embed\/ThG9EBbMhP8?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/div>\n<figcaption class=\"wp-element-caption\">Take a look at the accompanying video on this text! (Video by creator)<\/figcaption><\/figure>\n<p class=\"wp-block-paragraph\">So, right here is the general abstract of the Sesame mannequin in some bullet factors.<\/p>\n<ol class=\"wp-block-list\">\n<li class=\"wp-block-list-item\">\u00a0Sesame is constructed on a multimodal Dialog Speech Mannequin or a CSM.<\/li>\n<li class=\"wp-block-list-item\">Textual content and audio are tokenized collectively to type a sequence of tokens and enter into the spine transformer that autoregressively processes the sequence.<\/li>\n<li class=\"wp-block-list-item\">Whereas the textual content is processed like another text-based LLM, the audio is processed immediately from its waveform illustration. They use the Mimi encoder to transform the waveform into latent codes utilizing a cut up RVQ tokenizer.<\/li>\n<li class=\"wp-block-list-item\">The multimodal spine transformers devour a sequence of tokens and predict the subsequent zeroth codeword.<\/li>\n<li class=\"wp-block-list-item\">\u00a0One other light-weight transformer referred to as the Audio Decoder predicts the subsequent codewords from the zeroth codeword.<\/li>\n<li class=\"wp-block-list-item\">The ultimate audio body illustration is generated from combining all of the generated codewords and upsampled again to the waveform illustration.<\/li>\n<\/ol>\n<p class=\"wp-block-paragraph\">Thanks for studying!<\/p>\n<h3 class=\"wp-block-heading\">References and Should-read papers<\/h3>\n<p class=\"wp-block-paragraph\"><strong><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/www.youtube.com\/@avb_fj\">Take a look at my ML YouTube Channel<\/a><\/strong><\/p>\n<p class=\"wp-block-paragraph\"><strong><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/www.sesame.com\/research\/crossing_the_uncanny_valley_of_voice\">Sesame Blogpost and Demo<\/a><\/strong><\/p>\n<p class=\"wp-block-paragraph\"><strong>Related papers:\u00a0<br \/><\/strong>Moshi: <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2410.00037\" rel=\"noreferrer noopener\" target=\"_blank\">https:\/\/arxiv.org\/abs\/2410.00037<\/a>\u00a0<br \/>SoundStream: <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2107.03312\" rel=\"noreferrer noopener\" target=\"_blank\">https:\/\/arxiv.org\/abs\/2107.03312<\/a>\u00a0<br \/>HuBert: <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2106.07447\" rel=\"noreferrer noopener\" target=\"_blank\">https:\/\/arxiv.org\/abs\/2106.07447<\/a>\u00a0<br \/>Speech Tokenizer: <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2308.16692\" rel=\"noreferrer noopener\" target=\"_blank\">https:\/\/arxiv.org\/abs\/2308.16692<\/a><\/p>\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n<\/div>\n\n","protected":false},"excerpt":{"rendered":"<p>revealed a demo of their newest Speech-to-Speech mannequin. A conversational AI agent who&#8217;s actually good at talking, they supply related solutions, they communicate with expressions, and truthfully, they&#8217;re simply very enjoyable and interactive to play with. Word {that a} technical paper just isn&#8217;t out but, however they do have a quick weblog put up that [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":1354,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[55],"tags":[510,1235,358,1232,1233,1234],"class_list":["post-1352","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-machine-learning","tag-generates","tag-humanlike","tag-model","tag-sesame","tag-speech","tag-viral"],"_links":{"self":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/1352","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=1352"}],"version-history":[{"count":1,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/1352\/revisions"}],"predecessor-version":[{"id":1353,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/1352\/revisions\/1353"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/media\/1354"}],"wp:attachment":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=1352"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=1352"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=1352"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<!-- This website is optimized by Airlift. Learn more: https://airlift.net. Template:. Learn more: https://airlift.net. Template: 69d9690a190636c2e0989534. Config Timestamp: 2026-04-10 21:18:02 UTC, Cached Timestamp: 2026-05-15 01:38:04 UTC -->