Tuesday, April 7, 2026
Home Blog

Migraines might be handled by ramping up the mind’s cleansing system

0


Novel approaches are edging us nearer to relieving the agonising ache of migraines for all affected

Sergey Khakimullin/Getty Pictures

One-third of individuals with migraines don’t reply to present remedies, however harnessing the mind’s cleansing system may open up a brand new therapy possibility. A drug that ordinarily treats hypertension helped this method extra successfully take away a chemical substance from the brains of mice that may be a potent driver of migraines. Because of this, the mice confirmed fewer indicators of facial ache, which impacts about 60 per cent of individuals with migraines throughout an episode.

Round 1 in 7 individuals worldwide have migraines. Ache, strain or throbbing within the cheeks, jaw, brow or behind the eyes are frequent signs, and will be exacerbated by even mild contact. “Merely brushing their hair will be painful for [people with migraines],” says Adriana Della Pietra on the College of Iowa, who offered the analysis on the Oxford Glymphatic and Mind Clearance Symposium within the UK on 1 April.

Commonplace migraine remedies embody painkillers similar to triptans, that are thought to cut back irritation and decrease ranges of a neurotransmitter referred to as calcitonin gene-related peptide (CGRP) – a driver of migraines and the goal of our most potent therapies towards the situation. “However numerous individuals don’t reply to the medicine they’re provided, and very often, persons are dwelling by way of hell that may final for days. Very often, they will’t do on a regular basis duties,” says Valentina Mosienko on the College of Bristol, UK, who wasn’t concerned within the analysis.

In an earlier experiment, the researchers discovered that prazosin – a drug that’s accredited to deal with hypertension – relieves facial ache attributable to traumatic mind accidents in mice. Such accidents impair the mind’s waste disposal system, often known as the glymphatic system, whereas prazosin boosts the circulate of waste fluid from mind cells by way of this method. Nevertheless, some migraine-affected mice that the researchers used as a management additionally appeared to profit.

Investigating this additional, the workforce added prazosin to the consuming water of a gaggle of mice for six weeks, whereas a separate group got common water. The researchers then induced migraines in the entire rodents by injecting them with CGRP.

Half an hour later, the workforce prodded the mice’s foreheads with plastic filaments of accelerating thickness. Though none of those is ordinarily painful, the filaments are more and more noticeable as they grow to be thicker. The researchers discovered that the mice within the prazosin group might be touched by considerably thicker filaments with out flinching than the management mice. The previous group acted equally to mice that hadn’t been injected with CGRP, says Della Pietra.

Additional evaluation confirmed that prazosin reversed impairment of the glymphatic system attributable to CGRP, which most likely enhanced its clearance and that of different molecules that transmit ache alerts, says Della Pietra.

The researchers hope to check whether or not the identical results happen in individuals. “If it really works in people, that will be improbable,” says Mosienko. “The drug is already in use, so we all know it’s secure to be used.”

Subjects:

Serving to information facilities ship increased efficiency with much less {hardware} | MIT Information

0

To enhance information middle effectivity, a number of storage units are sometimes pooled collectively over a community so many functions can share them. However even with pooling, important system capability stays underutilized as a result of efficiency variability throughout the units.

MIT researchers have now developed a system that enhances the efficiency of storage units by dealing with three main sources of variability concurrently. Their strategy delivers important pace enhancements over conventional strategies that deal with just one supply of variability at a time.

The system makes use of a two-tier structure, with a central controller that makes big-picture selections about which duties every storage system performs, and native controllers for every machine that quickly reroute information if that system is struggling.

The tactic, which might adapt in real-time to shifting workloads, doesn’t require specialised {hardware}. When the researchers examined this technique on practical duties like AI mannequin coaching and picture compression, it almost doubled the efficiency delivered by conventional approaches. By intelligently balancing the workloads of a number of storage units, the system can improve total information middle effectivity.

“There’s a tendency to need to throw extra sources at an issue to unravel it, however that’s not sustainable in some ways. We would like to have the ability to maximize the longevity of those very costly and carbon-intensive sources,” says Gohar Chaudhry, {an electrical} engineering and pc science (EECS) graduate scholar and lead creator of a paper on this system. “With our adaptive software program answer, you may nonetheless squeeze numerous efficiency out of your current units earlier than it’s good to throw them away and purchase new ones.”

Chaudhry is joined on the paper by Ankit Bhardwaj, an assistant professor at Tufts College; Zhenyuan Ruan PhD ’24; and senior creator Adam Belay, an affiliate professor of EECS and a member of the MIT Laptop Science and Synthetic Intelligence Laboratory. The analysis will probably be introduced on the USENIX Symposium on Networked Methods Design and Implementation.

Leveraging untapped efficiency

Stable-state drives (SSDs) are high-performance digital storage units that enable functions to learn and write information. As an example, an SSD can retailer huge datasets and quickly ship information to a processor for machine-learning mannequin coaching.   

Pooling a number of SSDs collectively so many functions can share them improves effectivity, since not each software wants to make use of your entire capability of an SSD at a given time. However not all SSDs carry out equally, and the slowest system can restrict the general efficiency of the pool.

These inefficiencies come up from variability in SSD {hardware} and the duties they carry out.

To make the most of this untapped SSD efficiency, the researchers developed Sandook, a software-based system that tackles three main types of performance-hampering variability concurrently. “Sandook” is an Urdu phrase meaning “field,” to indicate “storage.”

One kind of variability is attributable to variations within the age, quantity of damage, and capability of SSDs that will have been bought at completely different instances from a number of distributors.

The second kind of variability is as a result of mismatch between learn and write operations occurring on the identical SSD. To put in writing new information to the system, the SSD should erase some current information. This course of can decelerate information reads, or retrievals, occurring on the similar time.

The third supply of variability is rubbish assortment, a means of gathering and eradicating outdated information to unlock area. This course of, which slows SSD operations, is triggered at random intervals {that a} information middle operator can’t management.

“I can’t assume all SSDs will behave identically via my complete deployment cycle. Even when I give all of them the identical workload, a few of them will probably be stragglers, which hurts the web throughput I can obtain,” Chaudhry explains.

Plan globally, react domestically

To deal with all three sources of variability, Sandook makes use of a two-tier construction. A worldwide schedular optimizes the distribution of duties for the general pool, whereas quicker schedulers on every SSD react to pressing occasions and shift operations away from congested units.

The system overcomes delays from read-write interference by rotating which SSDs an software can use for reads and writes. This reduces the possibility reads and writes occur concurrently on the identical machine.

Sandook additionally profiles the everyday efficiency of every SSD. It makes use of this info to detect when rubbish assortment is probably going slowing operations down. As soon as detected, Sandook reduces the workload on that SSD by diverting some duties till rubbish assortment is completed.

“If that SSD is doing rubbish assortment and might’t deal with the identical workload anymore, I need to give it a smaller workload and slowly ramp issues again up. We need to discover the candy spot the place it’s nonetheless performing some work, and faucet into that efficiency,” Chaudhry says.

The SSD profiles additionally enable Sandook’s international controller to assign workloads in a weighted style that considers the traits and capability of every system.

As a result of the worldwide controller sees the general image and the native controllers react on the fly, Sandook can concurrently handle types of variability that occur over completely different time scales. As an example, delays from rubbish assortment happen instantly, whereas latency attributable to put on and tear builds up over many months.

The researchers examined Sandook on a pool of 10 SSDs and evaluated the system on 4 duties: working a database, coaching a machine-learning mannequin, compressing photos, and storing person information. Sandook boosted the throughput of every software between 12 and 94 % when in comparison with static strategies, and improved the general utilization of SSD capability by 23 %.

The system enabled SSDs to realize 95 % of their theoretical most efficiency, with out the necessity for specialised {hardware} or application-specific updates.

“Our dynamic answer can unlock extra efficiency for all of the SSDs and actually push them to the restrict. Each little bit of capability it can save you actually counts at this scale,” Chaudhry says.

Sooner or later, the researchers need to incorporate new protocols obtainable on the newest SSDs that give operators extra management over information placement. In addition they need to leverage the predictability in AI workloads to extend the effectivity of SSD operations.

“Flash storage is a robust expertise that underpins trendy datacenter functions, however sharing this useful resource throughout workloads with extensively various efficiency calls for stays an excellent problem. This work strikes the needle meaningfully ahead with a chic and sensible answer prepared for deployment, bringing flash storage nearer to its full potential in manufacturing clouds,” says Josh Fried, a software program engineer at Google and incoming assistant professor on the College of Pennsylvania, who was not concerned with this work.

This analysis was funded, partially, by the Nationwide Science Basis, the U.S. Protection Superior Analysis Initiatives Company, and the Semiconductor Analysis Company.

It is not your tech stack, it is your construction — repair it

0


I’ve had this dialog many instances: A expertise chief pulls me apart and says some model of: “We preserve faltering. Handoffs disintegrate. No one owns the result. How can we repair it?”

My reply? It is in all probability not your course of or your tech stack. It is your organizational mannequin.

Challenge-based buildings have been a dependable administration method in tech-first organizations, and for good purpose. It is what scaled advanced programs, delivered large infrastructure and introduced extremely coordinated engineering efforts to life. It is sensible: outline the deliverable, set the timeline, employees a staff, ship it, exit. Repeat. However the tempo of expertise has outrun it.

What challenge fashions really price you

At the moment’s merchandise should not static deliverables — they’re programs that evolve constantly. They require fixed iteration, ongoing funding and tight alignment with buyer conduct.

A mannequin constructed round outlined endpoints creates friction in that type of atmosphere.

Associated:InformationWeek Podcast: Safeguarding IT ecosystems from exterior entry

Work will get handed off throughout groups. The staff that builds one thing is commonly not the one chargeable for operating or bettering it. Finish-to-end possession will get diluted. Context fades at every transition. Accountability begins to interrupt down.

I’ve seen organizations with distinctive expertise gradual to a crawl as a result of the mannequin surrounding it created friction at each handoff.

Planview’s 2023 Challenge to Product State of the Trade Report discovered that even with Agile groups, solely 8% of deliberate work finally delivered its supposed worth for the 326 IT leaders surveyed at 253 distinctive corporations. The precise quantity is much less vital than what it displays. Delivery one thing and making it work should not the identical factor.

As a result of on this mannequin, when one thing breaks or a buyer expertise degrades, the response is commonly to spin up one other challenge. A brand new workstream and new constitution. In the meantime, months cross and rivals transfer forward.

The result’s a niche between supply and consequence. And the construction is guilty.

The shift most organizations keep away from

Shifting to a product-led group does not require a brand new methodology or new instruments — only a completely different philosophy.

In a challenge mannequin, groups are accountable for delivering an outlined scope inside a set timeframe. As soon as that work is full, possession strikes on.

In a product mannequin, a cross-functional staff owns a product or buyer consequence. They do not hand it off when the challenge closes. They construct it, function it, and enhance it. They’re accountable for not simply whether or not one thing shipped, but in addition for the way it performs, the way it evolves and what it delivers to the enterprise and the shopper.

Associated:People are the North Star for AI-native workplaces — Gartner

Microsoft, Adobe and Spotify are the examples everybody cites as a result of the outcomes are arduous to argue with. When Satya Nadella reshaped Microsoft round merchandise and platforms moderately than challenge deliverables, it modified how briskly the corporate might transfer and the way cleanly it might coordinate throughout a really giant group.

That shift sounds easy. In apply, it is arduous however price it.

What a product mannequin really modifications

When possession is steady, the way in which groups function modifications with it.

Selections are made with an extended view. Tradeoffs are clearer as a result of the identical staff lives with the results. Work is prioritized based mostly on influence, not simply timelines.

The suggestions loop between constructing and studying tightens. Points are addressed sooner as a result of context stays with the staff. Enhancements construct on one another as an alternative of beginning over.

Velocity appears to be like completely different on this mannequin: It isn’t about shifting sooner from challenge to challenge. It’s about lowering the friction between steps so progress carries ahead.

The way to make the change with out breaking all the pieces

This transition doesn’t require a full reset, nevertheless it does require readability. Earlier than you embark on a migration, deal with three areas. Getting these proper is commonly the distinction between a transition that stalls and one which positive factors momentum.

Associated:Why CIOs cannot let customers wait on IT

  • Be trustworthy about the place possession actually lives and the place it breaks down. The place does the buck cease on a roadmap resolution? Who will get the two a.m. name when one thing breaks? Undecided? That is your baseline for change.

  • Second, resist the urge to alter all the pieces directly. Begin with one or two areas the place a persistent product staff would have a transparent, measurable influence. Early progress ought to construct momentum, not create chaos.

  • Third, have a look at how groups are evaluated. If success is outlined by supply milestones, groups will optimize for supply. If success is outlined by outcomes, conduct modifications. Ask: Did buyer satisfaction go up? Did our retention charges enhance? What’s the bottom-line influence? Did merchandise or programs get extra dependable? When you may ask these sorts of questions as an alternative of “Did it ship?” the tradition will observe.

The payoff

This type of transition requires wholesale rethinking round profession paths, reporting strains, planning and budgeting. Nevertheless, the payoff is there.

Organizations that make this transfer expertise sooner iteration, clearer accountability and stronger alignment between groups and enterprise outcomes.

Although the work usually does not really feel completed in the identical method as a result of there isn’t a clear endpoint, the progress turns into steady. As well as, groups can see the correlation between their work and the end result. 

The challenge mannequin was proper for a unique period. It rewarded precision, sequencing and management. However at present calls for pace, adaptability and possession. These reside in product groups, not challenge charters. 

Why it issues now

AI is accelerating how merchandise are constructed and the way rapidly they’ll evolve. Buyer expectations are rising on the similar time. On this atmosphere, organizations that may study and adapt rapidly get the benefit. Success at present nonetheless requires the correct instruments and expertise, nevertheless it’s now essential to have groups structured with a product focus that emphasizes continuity and possession.

If execution feels slower than it ought to, it’s price trying past your stack. The constraint is commonly not technical. It’s structural.



Posit AI Weblog: Illustration studying with MMD-VAE


Not too long ago, we confirmed the best way to generate pictures utilizing generative adversarial networks (GANs). GANs could yield wonderful outcomes, however the contract there mainly is: what you see is what you get.
Generally this can be all we wish. In different instances, we could also be extra concerned with really modelling a website. We don’t simply wish to generate realistic-looking samples – we wish our samples to be situated at particular coordinates in area house.

For instance, think about our area to be the house of facial expressions. Then our latent house is perhaps conceived as two-dimensional: In accordance with underlying emotional states, expressions differ on a positive-negative scale. On the similar time, they differ in depth. Now if we skilled a VAE on a set of facial expressions adequately masking the ranges, and it did actually “uncover” our hypothesized dimensions, we may then use it to generate previously-nonexisting incarnations of factors (faces, that’s) in latent house.

Variational autoencoders are just like probabilistic graphical fashions in that they assume a latent house that’s accountable for the observations, however unobservable. They’re just like plain autoencoders in that they compress, after which decompress once more, the enter area. In distinction to plain autoencoders although, the essential level right here is to plan a loss perform that enables to acquire informative representations in latent house.

In a nutshell

In commonplace VAEs (Kingma and Welling 2013), the target is to maximise the proof decrease sure (ELBO):

[ELBO = E[log p(x|z)] – KL(q(z)||p(z))]

In plain phrases and expressed when it comes to how we use it in observe, the primary element is the reconstruction loss we additionally see in plain (non-variational) autoencoders. The second is the Kullback-Leibler divergence between a previous imposed on the latent house (usually, an ordinary regular distribution) and the illustration of latent house as realized from the info.

A significant criticism concerning the normal VAE loss is that it leads to uninformative latent house. Options embody (beta)-VAE(Burgess et al. 2018), Information-VAE (Zhao, Music, and Ermon 2017), and extra. The MMD-VAE(Zhao, Music, and Ermon 2017) applied beneath is a subtype of Information-VAE that as an alternative of constructing every illustration in latent house as related as attainable to the prior, coerces the respective distributions to be as shut as attainable. Right here MMD stands for most imply discrepancy, a similarity measure for distributions primarily based on matching their respective moments. We clarify this in additional element beneath.

Our goal at this time

On this publish, we’re first going to implement an ordinary VAE that strives to maximise the ELBO. Then, we evaluate its efficiency to that of an Information-VAE utilizing the MMD loss.

Our focus will likely be on inspecting the latent areas and see if, and the way, they differ as a consequence of the optimization standards used.

The area we’re going to mannequin will likely be glamorous (trend!), however for the sake of manageability, confined to measurement 28 x 28: We’ll compress and reconstruct pictures from the Style MNIST dataset that has been developed as a drop-in to MNIST.

An ordinary variational autoencoder

Seeing we haven’t used TensorFlow keen execution for some weeks, we’ll do the mannequin in an keen approach.
If you happen to’re new to keen execution, don’t fear: As each new approach, it wants some getting accustomed to, however you’ll rapidly discover that many duties are made simpler when you use it. A easy but full, template-like instance is obtainable as a part of the Keras documentation.

Setup and knowledge preparation

As regular, we begin by ensuring we’re utilizing the TensorFlow implementation of Keras and enabling keen execution. Moreover tensorflow and keras, we additionally load tfdatasets to be used in knowledge streaming.

By the way in which: No must copy-paste any of the beneath code snippets. The 2 approaches can be found amongst our Keras examples, specifically, as eager_cvae.R and mmd_cvae.R.

The info comes conveniently with keras, all we have to do is the same old normalization and reshaping.

trend <- dataset_fashion_mnist()

c(train_images, train_labels) %<-% trend$practice
c(test_images, test_labels) %<-% trend$check

train_x <- train_images %>%
  `/`(255) %>%
  k_reshape(c(60000, 28, 28, 1))

test_x <- test_images %>% `/`(255) %>%
  k_reshape(c(10000, 28, 28, 1))

What do we’d like the check set for, given we’re going to practice an unsupervised (a greater time period being: semi-supervised) mannequin? We’ll use it to see how (beforehand unknown) knowledge factors cluster collectively in latent house.

Now put together for streaming the info to keras:

buffer_size <- 60000
batch_size <- 100
batches_per_epoch <- buffer_size / batch_size

train_dataset <- tensor_slices_dataset(train_x) %>%
  dataset_shuffle(buffer_size) %>%
  dataset_batch(batch_size)

test_dataset <- tensor_slices_dataset(test_x) %>%
  dataset_batch(10000)

Subsequent up is defining the mannequin.

Encoder-decoder mannequin

The mannequin actually is 2 fashions: the encoder and the decoder. As we’ll see shortly, in the usual model of the VAE there’s a third element in between, performing the so-called reparameterization trick.

The encoder is a customized mannequin, comprised of two convolutional layers and a dense layer. It returns the output of the dense layer cut up into two elements, one storing the imply of the latent variables, the opposite their variance.

latent_dim <- 2

encoder_model <- perform(identify = NULL) {
  
  keras_model_custom(identify = identify, perform(self) {
    self$conv1 <-
      layer_conv_2d(
        filters = 32,
        kernel_size = 3,
        strides = 2,
        activation = "relu"
      )
    self$conv2 <-
      layer_conv_2d(
        filters = 64,
        kernel_size = 3,
        strides = 2,
        activation = "relu"
      )
    self$flatten <- layer_flatten()
    self$dense <- layer_dense(items = 2 * latent_dim)
    
    perform (x, masks = NULL) {
      x %>%
        self$conv1() %>%
        self$conv2() %>%
        self$flatten() %>%
        self$dense() %>%
        tf$cut up(num_or_size_splits = 2L, axis = 1L) 
    }
  })
}

We select the latent house to be of dimension 2 – simply because that makes visualization straightforward.
With extra advanced knowledge, you’ll in all probability profit from selecting the next dimensionality right here.

So the encoder compresses actual knowledge into estimates of imply and variance of the latent house.
We then “not directly” pattern from this distribution (the so-called reparameterization trick):

reparameterize <- perform(imply, logvar) {
  eps <- k_random_normal(form = imply$form, dtype = tf$float64)
  eps * k_exp(logvar * 0.5) + imply
}

The sampled values will function enter to the decoder, who will try and map them again to the unique house.
The decoder is mainly a sequence of transposed convolutions, upsampling till we attain a decision of 28×28.

decoder_model <- perform(identify = NULL) {
  
  keras_model_custom(identify = identify, perform(self) {
    
    self$dense <- layer_dense(items = 7 * 7 * 32, activation = "relu")
    self$reshape <- layer_reshape(target_shape = c(7, 7, 32))
    self$deconv1 <-
      layer_conv_2d_transpose(
        filters = 64,
        kernel_size = 3,
        strides = 2,
        padding = "similar",
        activation = "relu"
      )
    self$deconv2 <-
      layer_conv_2d_transpose(
        filters = 32,
        kernel_size = 3,
        strides = 2,
        padding = "similar",
        activation = "relu"
      )
    self$deconv3 <-
      layer_conv_2d_transpose(
        filters = 1,
        kernel_size = 3,
        strides = 1,
        padding = "similar"
      )
    
    perform (x, masks = NULL) {
      x %>%
        self$dense() %>%
        self$reshape() %>%
        self$deconv1() %>%
        self$deconv2() %>%
        self$deconv3()
    }
  })
}

Word how the ultimate deconvolution doesn’t have the sigmoid activation you might need anticipated. It is because we will likely be utilizing tf$nn$sigmoid_cross_entropy_with_logits when calculating the loss.

Talking of losses, let’s examine them now.

Loss calculations

One solution to implement the VAE loss is combining reconstruction loss (cross entropy, within the current case) and Kullback-Leibler divergence. In Keras, the latter is obtainable immediately as loss_kullback_leibler_divergence.

Right here, we observe a current Google Colaboratory pocket book in batch-estimating the whole ELBO as an alternative (as an alternative of simply estimating reconstruction loss and computing the KL-divergence analytically):

[ELBO batch estimate = log p(x_{batch}|z_{sampled})+log p(z)−log q(z_{sampled}|x_{batch})]

Calculation of the conventional loglikelihood is packaged right into a perform so we are able to reuse it in the course of the coaching loop.

normal_loglik <- perform(pattern, imply, logvar, reduce_axis = 2) {
  loglik <- k_constant(0.5, dtype = tf$float64) *
    (k_log(2 * k_constant(pi, dtype = tf$float64)) +
    logvar +
    k_exp(-logvar) * (pattern - imply) ^ 2)
  - k_sum(loglik, axis = reduce_axis)
}

Peeking forward some, throughout coaching we are going to compute the above as follows.

First,

crossentropy_loss <- tf$nn$sigmoid_cross_entropy_with_logits(
  logits = preds,
  labels = x
)
logpx_z <- - k_sum(crossentropy_loss)

yields (log p(x|z)), the loglikelihood of the reconstructed samples given values sampled from latent house (a.ok.a. reconstruction loss).

Then,

logpz <- normal_loglik(
  z,
  k_constant(0, dtype = tf$float64),
  k_constant(0, dtype = tf$float64)
)

provides (log p(z)), the prior loglikelihood of (z). The prior is assumed to be commonplace regular, as is most frequently the case with VAEs.

Lastly,

logqz_x <- normal_loglik(z, imply, logvar)

vields (log q(z|x)), the loglikelihood of the samples (z) given imply and variance computed from the noticed samples (x).

From these three parts, we are going to compute the ultimate loss as

loss <- -k_mean(logpx_z + logpz - logqz_x)

After this peaking forward, let’s rapidly end the setup so we prepare for coaching.

Remaining setup

Moreover the loss, we’d like an optimizer that may try to decrease it.

optimizer <- tf$practice$AdamOptimizer(1e-4)

We instantiate our fashions …

encoder <- encoder_model()
decoder <- decoder_model()

and arrange checkpointing, so we are able to later restore skilled weights.

checkpoint_dir <- "./checkpoints_cvae"
checkpoint_prefix <- file.path(checkpoint_dir, "ckpt")
checkpoint <- tf$practice$Checkpoint(
  optimizer = optimizer,
  encoder = encoder,
  decoder = decoder
)

From the coaching loop, we are going to, in sure intervals, additionally name three capabilities not reproduced right here (however accessible within the code instance): generate_random_clothes, used to generate garments from random samples from the latent house; show_latent_space, that shows the whole check set in latent (2-dimensional, thus simply visualizable) house; and show_grid, that generates garments in keeping with enter values systematically spaced out in a grid.

Let’s begin coaching! Really, earlier than we try this, let’s take a look at what these capabilities show earlier than any coaching: As a substitute of garments, we see random pixels. Latent house has no construction. And several types of garments don’t cluster collectively in latent house.

Coaching loop

We’re coaching for 50 epochs right here. For every epoch, we loop over the coaching set in batches. For every batch, we observe the same old keen execution stream: Contained in the context of a GradientTape, apply the mannequin and calculate the present loss; then outdoors this context calculate the gradients and let the optimizer carry out backprop.

What’s particular right here is that we now have two fashions that each want their gradients calculated and weights adjusted. This may be taken care of by a single gradient tape, supplied we create it persistent.

After every epoch, we save present weights and each ten epochs, we additionally save plots for later inspection.

num_epochs <- 50

for (epoch in seq_len(num_epochs)) {
  iter <- make_iterator_one_shot(train_dataset)
  
  total_loss <- 0
  logpx_z_total <- 0
  logpz_total <- 0
  logqz_x_total <- 0
  
  until_out_of_range({
    x <-  iterator_get_next(iter)
    
    with(tf$GradientTape(persistent = TRUE) %as% tape, {
      
      c(imply, logvar) %<-% encoder(x)
      z <- reparameterize(imply, logvar)
      preds <- decoder(z)
      
      crossentropy_loss <-
        tf$nn$sigmoid_cross_entropy_with_logits(logits = preds, labels = x)
      logpx_z <-
        - k_sum(crossentropy_loss)
      logpz <-
        normal_loglik(z,
                      k_constant(0, dtype = tf$float64),
                      k_constant(0, dtype = tf$float64)
        )
      logqz_x <- normal_loglik(z, imply, logvar)
      loss <- -k_mean(logpx_z + logpz - logqz_x)
      
    })

    total_loss <- total_loss + loss
    logpx_z_total <- tf$reduce_mean(logpx_z) + logpx_z_total
    logpz_total <- tf$reduce_mean(logpz) + logpz_total
    logqz_x_total <- tf$reduce_mean(logqz_x) + logqz_x_total
    
    encoder_gradients <- tape$gradient(loss, encoder$variables)
    decoder_gradients <- tape$gradient(loss, decoder$variables)
    
    optimizer$apply_gradients(
      purrr::transpose(listing(encoder_gradients, encoder$variables)),
      global_step = tf$practice$get_or_create_global_step()
    )
    optimizer$apply_gradients(
      purrr::transpose(listing(decoder_gradients, decoder$variables)),
      global_step = tf$practice$get_or_create_global_step()
    )
    
  })
  
  checkpoint$save(file_prefix = checkpoint_prefix)
  
  cat(
    glue(
      "Losses (epoch): {epoch}:",
      "  {(as.numeric(logpx_z_total)/batches_per_epoch) %>% spherical(2)} logpx_z_total,",
      "  {(as.numeric(logpz_total)/batches_per_epoch) %>% spherical(2)} logpz_total,",
      "  {(as.numeric(logqz_x_total)/batches_per_epoch) %>% spherical(2)} logqz_x_total,",
      "  {(as.numeric(total_loss)/batches_per_epoch) %>% spherical(2)} whole"
    ),
    "n"
  )
  
  if (epoch %% 10 == 0) {
    generate_random_clothes(epoch)
    show_latent_space(epoch)
    show_grid(epoch)
  }
}

Outcomes

How effectively did that work? Let’s see the sorts of garments generated after 50 epochs.

Additionally, how disentangled (or not) are the totally different courses in latent house?

And now watch totally different garments morph into each other.

How good are these representations? That is laborious to say when there may be nothing to match with.

So let’s dive into MMD-VAE and see the way it does on the identical dataset.

MMD-VAE

MMD-VAE guarantees to generate extra informative latent options, so we might hope to see totally different conduct particularly within the clustering and morphing plots.

Knowledge setup is identical, and there are solely very slight variations within the mannequin. Please try the whole code for this instance, mmd_vae.R, as right here we’ll simply spotlight the variations.

Variations within the mannequin(s)

There are three variations as regards mannequin structure.

One, the encoder doesn’t must return the variance, so there isn’t a want for tf$cut up. The encoder’s name technique now simply is

perform (x, masks = NULL) {
  x %>%
    self$conv1() %>%
    self$conv2() %>%
    self$flatten() %>%
    self$dense() 
}

Between the encoder and the decoder, we don’t want the sampling step anymore, so there isn’t a reparameterization.
And since we received’t use tf$nn$sigmoid_cross_entropy_with_logits to compute the loss, we let the decoder apply the sigmoid within the final deconvolution layer:

self$deconv3 <- layer_conv_2d_transpose(
  filters = 1,
  kernel_size = 3,
  strides = 1,
  padding = "similar",
  activation = "sigmoid"
)

Loss calculations

Now, as anticipated, the massive novelty is within the loss perform.

The loss, most imply discrepancy (MMD), relies on the concept that two distributions are an identical if and provided that all moments are an identical.
Concretely, MMD is estimated utilizing a kernel, such because the Gaussian kernel

[k(z,z’)=frac{e^z-z’}{2sigma^2}]

to evaluate similarity between distributions.

The concept then is that if two distributions are an identical, the common similarity between samples from every distribution must be an identical to the common similarity between combined samples from each distributions:

[MMD(p(z)||q(z))=E_{p(z),p(z’)}[k(z,z’)]+E_{q(z),q(z’)}[k(z,z’)]−2E_{p(z),q(z’)}[k(z,z’)]]
The next code is a direct port of the creator’s authentic TensorFlow code:

compute_kernel <- perform(x, y) {
  x_size <- k_shape(x)[1]
  y_size <- k_shape(y)[1]
  dim <- k_shape(x)[2]
  tiled_x <- k_tile(
    k_reshape(x, k_stack(listing(x_size, 1, dim))),
    k_stack(listing(1, y_size, 1))
  )
  tiled_y <- k_tile(
    k_reshape(y, k_stack(listing(1, y_size, dim))),
    k_stack(listing(x_size, 1, 1))
  )
  k_exp(-k_mean(k_square(tiled_x - tiled_y), axis = 3) /
          k_cast(dim, tf$float64))
}

compute_mmd <- perform(x, y, sigma_sqr = 1) {
  x_kernel <- compute_kernel(x, x)
  y_kernel <- compute_kernel(y, y)
  xy_kernel <- compute_kernel(x, y)
  k_mean(x_kernel) + k_mean(y_kernel) - 2 * k_mean(xy_kernel)
}

Coaching loop

The coaching loop differs from the usual VAE instance solely within the loss calculations.
Listed below are the respective strains:

 with(tf$GradientTape(persistent = TRUE) %as% tape, {
      
      imply <- encoder(x)
      preds <- decoder(imply)
      
      true_samples <- k_random_normal(
        form = c(batch_size, latent_dim),
        dtype = tf$float64
      )
      loss_mmd <- compute_mmd(true_samples, imply)
      loss_nll <- k_mean(k_square(x - preds))
      loss <- loss_nll + loss_mmd
      
    })

So we merely compute MMD loss in addition to reconstruction loss, and add them up. No sampling is concerned on this model.
After all, we’re curious to see how effectively that labored!

Outcomes

Once more, let’s take a look at some generated garments first. It looks as if edges are a lot sharper right here.

The clusters too look extra properly unfold out within the two dimensions. And, they’re centered at (0,0), as we might have hoped for.

Lastly, let’s see garments morph into each other. Right here, the graceful, steady evolutions are spectacular!
Additionally, almost all house is stuffed with significant objects, which hasn’t been the case above.

MNIST

For curiosity’s sake, we generated the identical sorts of plots after coaching on authentic MNIST.
Right here, there are hardly any variations seen in generated random digits after 50 epochs of coaching.

Left: random digits as generated after training with ELBO loss. Right: MMD loss.

Additionally the variations in clustering are usually not that massive.

Left: latent space as observed after training with ELBO loss. Right: MMD loss.

However right here too, the morphing seems way more natural with MMD-VAE.

Left: Morphing as observed after training with ELBO loss. Right: MMD loss.

Conclusion

To us, this demonstrates impressively what massive a distinction the fee perform could make when working with VAEs.
One other element open to experimentation stands out as the prior used for the latent house – see this speak for an summary of different priors and the “Variational Combination of Posteriors” paper (Tomczak and Welling 2017) for a preferred current method.

For each price capabilities and priors, we anticipate efficient variations to turn out to be approach larger nonetheless after we go away the managed setting of (Style) MNIST and work with real-world datasets.

Burgess, C. P., I. Higgins, A. Pal, L. Matthey, N. Watters, G. Desjardins, and A. Lerchner. 2018. “Understanding Disentangling in Beta-VAE.” ArXiv e-Prints, April. https://arxiv.org/abs/1804.03599.
Doersch, C. 2016. “Tutorial on Variational Autoencoders.” ArXiv e-Prints, June. https://arxiv.org/abs/1606.05908.

Kingma, Diederik P., and Max Welling. 2013. “Auto-Encoding Variational Bayes.” CoRR abs/1312.6114.

Tomczak, Jakub M., and Max Welling. 2017. “VAE with a VampPrior.” CoRR abs/1705.07120.

Zhao, Shengjia, Jiaming Music, and Stefano Ermon. 2017. “InfoVAE: Info Maximizing Variational Autoencoders.” CoRR abs/1706.02262. http://arxiv.org/abs/1706.02262.

I’ve been ready years for Google Pictures to get this one primary characteristic — and it is lastly right here

0


What you have to know

  • Google Pictures lastly provides video playback pace management.
  • Now you can watch movies from 0.5x to 2x straight contained in the app.
  • It is rolling out now on Android, however availability is gradual and iOS continues to be ready.

Google Pictures has been the primary gallery app for many individuals for years, however it at all times missed one easy characteristic: you couldn’t change video playback pace. In the event you needed to rapidly get by way of an extended recital or decelerate a sports activities spotlight, you had been caught at regular pace. That’s lastly altering.

Android Central’s Take

It is a customary characteristic in most media apps, so it’s shocking it took this lengthy to reach in Google Pictures.

Google is including a video playback pace management to Google Pictures on Android. Now you may pace up or decelerate movies proper within the app.

Scientists say 7 days of meditation can rewire your mind

0


Researchers on the College of California San Diego report {that a} weeklong program combining meditation and different mind-body strategies can rapidly produce measurable modifications in each mind exercise and blood biology. The research discovered that these practices activated pure pathways concerned in mind flexibility, metabolism, immune operate, and ache aid. Printed in Communications Biology, the findings provide new proof that psychological practices can affect bodily well being in vital methods.

Meditation and comparable approaches have been used for hundreds of years to assist well-being, however scientists have struggled to elucidate precisely how they have an effect on the physique. This new analysis, half of a big initiative funded by the InnerScience Analysis Fund, is the primary to systematically measure the mixed organic results of a number of mind-body strategies delivered over a short while.

“We have identified for years that practices like meditation can affect well being, however what’s placing is that combining a number of mind-body practices right into a single retreat produced modifications throughout so many organic methods that we may measure immediately within the mind and blood,” stated senior research creator Hemal H. Patel, Ph.D., professor of anesthesiology at UC San Diego College of Medication and analysis profession scientist on the Veterans Affairs San Diego Healthcare System. “This is not about simply stress aid or leisure; that is about essentially altering how the mind engages with actuality and quantifying these modifications biologically.”

Contained in the 7-Day Meditation Program

The research adopted 20 wholesome adults who took half in a 7-day residential retreat led by neuroscience educator and creator Joe Dispenza, D.C. Members attended lectures and accomplished about 33 hours of guided meditation together with group-based therapeutic actions.

These periods used an “open-label placebo” method, that means members had been conscious that some practices had been offered as placebos. Even so, such interventions can nonetheless produce actual results by way of expectation, shared expertise, and social connection.

Earlier than and after the retreat, researchers used useful magnetic resonance imaging (fMRI) to observe mind exercise. Blood samples had been additionally analyzed to trace modifications in metabolism, immune operate, and different organic markers.

Mind, Immune, and Metabolic Adjustments Noticed

After the retreat, a number of notable modifications had been detected:

  • Mind community modifications: Exercise decreased in areas linked to inside psychological chatter, suggesting extra environment friendly mind operate.
  • Enhanced neuroplasticity: Blood plasma collected after the retreat inspired lab-grown neurons to increase and type new connections.
  • Metabolic shifts: Cells uncovered to post-retreat plasma confirmed elevated glycolytic (sugar-burning) metabolism, indicating improved metabolic flexibility.
  • Pure ache aid: Ranges of endogenous opioids, the physique’s pure painkillers, rose following the retreat.
  • Immune activation: Each inflammatory and anti inflammatory alerts elevated, pointing to a balanced and adaptive immune response.
  • Gene and molecular signaling modifications: Small RNA and gene exercise shifted in methods linked to brain-related organic pathways.

Mystical Experiences Linked to Mind Connectivity

Members additionally accomplished the Mystical Expertise Questionnaire (MEQ-30), which measures emotions akin to unity, transcendence, and altered consciousness throughout meditation. Scores elevated from a median of two.37 earlier than the retreat to three.02 afterward.

Those that reported stronger mystical experiences additionally confirmed extra pronounced organic modifications, together with better coordination between totally different mind areas. This means that deeper subjective experiences could also be tied to measurable modifications in mind operate.

Meditation and Psychedelic-Like Mind States

The researchers discovered that the mind exercise patterns noticed after the retreat carefully resembled these beforehand linked to psychedelic substances.

“We’re seeing the identical mystical experiences and neural connectivity patterns that sometimes require psilocybin, now achieved by way of meditation apply alone,” added Patel. “Seeing each central nervous system modifications in mind scans and systemic modifications in blood chemistry underscores that these mind-body practices are performing on a whole-body scale.”

The findings assist clarify how non-drug approaches like meditation might assist total well being. By boosting neuroplasticity and influencing immune exercise, these practices may enhance emotional regulation, stress resilience, and psychological well-being. The rise in pure pain-relief chemical substances additionally factors to potential functions for managing persistent ache.

What Comes Subsequent for Thoughts-Physique Analysis

Though the research targeted on wholesome people, the researchers word that extra work is required to find out how these results translate to medical populations. Future research will discover whether or not comparable packages may assist individuals with persistent ache, temper issues, or immune-related circumstances.

The staff additionally plans to look at how totally different components of the retreat, together with meditation, reconceptualization, and open-label placebo therapeutic, contribute individually and collectively. One other key query is how lengthy these organic modifications final and whether or not repeated apply can strengthen or preserve them.

“This research reveals that our minds and our bodies are deeply interconnected — what we consider, how we focus our consideration, and the practices we take part in can depart measurable fingerprints on our biology,” stated first creator Alex Jinich-Diamant, a doctoral pupil within the Departments of Cognitive Science and Anesthesiology at UC San Diego. “It is an thrilling step towards understanding how aware expertise and bodily well being are intertwined, and the way we’d harness that connection to advertise well-being in new methods.”

Extra coauthors of the research embrace Sierra Simpson, Juan P. Zuniga-Hertz, Ramamurthy Chitteti, Jan M. Schilling, Jacqueline A. Bonds, Laura Case, Andrei V. Chernov, Natalia Esther Amkie Stahl, Michael Licamele, Narin Fazlalipour and, Swetha Devulapalli, at UC San Diego; Joe Dispenza and Michelle A. Poirier at Metamorphosis LLC; Jacqueline Maree and Tobias Moeller-Bertram at VitaMed Analysis; and Leonardo Christov-Moore and Nicco Reggente on the Institute for Superior Consciousness Research.

This work was supported by the InnerScience Analysis Fund and a Veterans Administration Analysis Profession Scientist Award (BX005229).

Disclosure: One co-author (Joe Dispenza) is employed by Encephalon, Inc., the corporate providing the retreat; all different authors declare no competing pursuits.

Constructing sophisticated expressions the simple means

0


Have you ever each wished to make an “simple” calculation–say, after becoming a mannequin–and gotten misplaced since you simply weren’t positive the place to seek out the levels of freedom of the residual or the usual error of the coefficient? Have you ever ever been within the midst of setting up an “simple” calculation and was all of the sudden uncertain simply what e(df_r) actually was? I’ve an answer.

It’s known as Stata’s expression builder. You may get to it from the show dialog (Knowledge->Different Utilities->Hand Calculator)

Within the dialog, click on the Create button to convey up the builder. Actually, it doesn’t seem like a lot:

I wish to present you find out how to use this expression builder; in the event you’ll follow me, it’ll be value your time.

Let’s begin over once more and assume you’re within the midst of an evaluation, say,

. sysuse auto, clear
. regress worth mpg size

Subsequent invoke the expression builder by knocking down the menu Knowledge->Different Utilities->Hand Calculator. Click on Create. It seems like this:

Now click on on the tree node icon (+) in entrance of “Estimation outcomes” after which scroll right down to see what’s beneath. You’ll see

Click on on Scalars:

The center field now incorporates the scalars saved in e(). N occurs to be highlighted, however you would click on on any of the scalars. If you happen to look beneath the 2 packing containers, you see the worth of the e() scalar chosen in addition to its worth and a brief description. e(N) is 74 and is the “variety of observations”.

It really works the identical means for all the opposite classes within the field on the left: Operators, Capabilities, Variables, Coefficients, Estimation outcomes, Returned outcomes, System parameters, Matrices, Macros, Scalars, Notes, and Traits. You merely click on on the tree node icon (+), and the class expands to point out what is on the market.

You’ve now mastered the expression builder!

Let’s attempt it out.

Say you wish to confirm that the p-value of the coefficient on mpg is accurately calculated by regress–which reviews 0.052–or extra possible, you wish to confirm that you understand how it was calculated. You assume the method is

or, as an expression in Stata,

2*ttail(e(df_r), abs(_b[mpg]/_se[mpg]))

However I’m leaping forward. It’s possible you’ll not keep in mind that _b[mpg] is the coefficient on variable mpg, or that _se[mpg] is its corresponding customary error, or that abs() is Stata’s absolute worth perform, or that e(df_r) is the residual levels of freedom from the regression, or that ttail() is Stata’s Pupil’s t distribution perform. We are able to construct the above expression utilizing the builder as a result of all of the parts could be accessed by way of the builder. The ttail() and abs() capabilities are within the Capabilities class, the e(df_r) scalar is within the Estimation outcomes class, and _b[mpg] and _se[mpg] are within the Coefficients class.

What’s good concerning the builder is that not solely are the merchandise names listed but in addition a definition, syntax, and worth are displayed once you click on on an merchandise. Having all this data in a single place makes constructing a posh expression a lot simpler.

One other instance of when the expression builder turns out to be useful is when computing intraclass correlations after xtmixed. Think about a easy two-level mannequin from Instance 1 in [XT] xtmixed, which fashions weight trajectories of 48 pigs from 9 successive weeks:

. use http://www.stata-press.com/information/r12/pig
. xtmixed weight week || id:, variance

The intraclass correlation is a nonlinear perform of variance parts. On this instance, the (residual) intraclass correlation is the ratio of the between-pig variance, var(_cons), to the full variance, between-pig variance plus residual (within-pig) variance, or var(_cons) + var(residual).

The xtmixed command doesn’t retailer the estimates of variance parts instantly. As a substitute, it shops them as log customary deviations in e(b) such that _b[lns1_1_1:_cons] is the estimated log of between-pig customary deviation, and _b[lnsig_e:_cons] is the estimated log of residual (within-pig) customary deviation. So to compute the intraclass correlation, we should first rework log customary deviations to variances:

exp(2*_b[lns1_1_1:_cons])
exp(2*_b[lnsig_e:_cons])

The ultimate expression for the intraclass correlation is then

exp(2*_b[lns1_1_1:_cons]) / (exp(2*_b[lns1_1_1:_cons])+exp(2*_b[lnsig_e:_cons]))

The issue is that few individuals keep in mind that _b[lns1_1_1:_cons] is the estimated log of between-pig customary deviation. The few who do definitely don’t wish to kind it. So use the expression builder as we do beneath:

On this case, we’re utilizing the expression builder accessed from Stata’s nlcom dialog, which reviews estimated nonlinear mixtures together with their customary errors. As soon as we press OK right here and within the nlcom dialog, we’ll see

. nlcom (exp(2*_b[lns1_1_1:_cons])/(exp(2*_b[lns1_1_1:_cons])+exp(2*_b[lnsig_e:_cons])))

  _nl_1:  exp(2*_b[lns1_1_1:_cons])/(exp(2*_b[lns1_1_1:_cons])+exp(2*_b[lnsig_e:_cons]))

------------------------------------------------------------------------------
      weight |      Coef.   Std. Err.      z    P>|z|     [95% Conf. Interval]
-------------+----------------------------------------------------------------
       _nl_1 |   .7717142   .0393959    19.59   0.000     .6944996    .8489288
------------------------------------------------------------------------------

The above might simply be prolonged to computing various kinds of intraclass correlations arising in higher-level random-effects fashions. Using the expression builder for that turns into much more useful.



Rethinking Enterprise Search with Cortex Search

0


In response to Stack Overflow and Atlassian, builders lose between 6 and 10 hours each week looking for info or clarifying unclear documentation. For a 50-developer staff, that provides as much as $675,000–$1.1 million in wasted productiveness yearly. This isn’t only a tooling subject. It’s a retrieval drawback.
Enterprises have loads of knowledge however lack quick, dependable methods to search out the precise info. Conventional search fails as techniques develop advanced, slowing onboarding, selections, and assist. On this article, we discover how fashionable enterprise search solves these gaps.

Why Conventional Enterprise Search Falls Quick

Most enterprise search techniques had been constructed for a special period. They assume comparatively static content material, predictable bugs and queries, and guide tuning to remain related. In fashionable knowledge setting none of these assumptions maintain significance. 

Groups work throughout quickly altering datasets. Queries are ambiguous and conversational. Context issues as a lot as Key phrases. But many search instruments nonetheless depend on brittle guidelines and precise matches, forcing customers guess the precise phrasing reasonably than expressing actual intent.  

The result’s acquainted. Individuals search repeatedly, refine queries manually or abandon search altogether. In AI-powered purposes, the issue turns into extra severe. Poor retrieval doesn’t simply gradual customers down. It typically feeds incomplete or irrelevant context into language fashions, growing the danger of low-quality or deceptive outputs.  

The Change to Hybrid Retrieval

The subsequent technology of enterprise search is constructed on hybrid retrieval. As an alternative of selecting between key phrase search and semantic search, fashionable techniques mix each of them.  

Key phrase search excels at precision. Vector search captures which means and intent. Collectively, they allow search experiences which might be quick, versatile and resilient throughout a variety of queries. 

Cortex Search is designed orienting this hybrid method from the beginning. It offers low latency, high-quality fuzzy search instantly over Snowflake knowledge, with out requiring groups to handle embeddings and tune relevance parameters or preserve customized infrastructure. The retrieval layer adapts to the info, not the opposite method round. 

Fairly than treating search as an add on characteristic, Coretx Search makes it a foundational functionality that scales with enterprise knowledge complexity. 

Cortex Search helps two major use instances which might be more and more central to fashionable knowledge methods. 

First is Retrieval Augmented Technology. Cortex Search acts because the retrieval engine that provides massive language fashions with correct, up-to-date enterprise context. This grounding layer is what permits AI chat purposes to ship responses which might be particular, related and aligned with proprietary knowledge reasonably than generic patterns. 

Second is Enterprise Search. Cortex Search can energy high-quality search experiences embedded instantly into purposes, instruments and workflows. Customers ask questions in pure language and obtain outcomes ranked by each semantic relevance and key phrase precision.

Below the hood, cortex search indexes textual content knowledge, applies hybrid retrieval and makes use of semantic reranking to floor probably the most related outcomes. Refreshes are automated and incremental, so search outcomes keep aligned with the present state of the info with out guide intervention. 

This issues as a result of retrieval high quality instantly shapes consumer belief. When search works persistently, folks depend on it. When it doesn’t, they cease utilizing it and fall again to slower, dearer paths. 

How Cortex Search Works in Apply

At a excessive stage, Cortex Search abstracts away the toughest elements of constructing a contemporary retrieval system. 

What we’ll Construct: A buyer assist AI assistant that solutions consumer questions by retrieving grounded context from historic assist tickets and transcripts: then passing that context to a Snowflake Cortex LLM to generate correct, particular solutions.

Stipulations

Requirement Particulars
Snowflake Account Free trial at trial.snowflake.com — Enterprise tier or above
Snowflake Position SYSADMIN or a task with CREATE DATABASE, CREATE WAREHOUSE, CREATE CORTEX SEARCH SERVICE privileges
Python 3.9+
Packages snowflake-snowpark-python, snowflake-core

Establishing Snowflake account

  1. Head over to trial.snowflake.com and Join the Enterprise account
  1. Now you will note one thing like this:
Snowflake dashboard

Step 1 — Set Up Snowflake Atmosphere

Run the next in a Snowflake Worksheet to create the database, schema 

First create a brand new sql file.

Setting Snowflake Environment
CREATE DATABASE IF NOT EXISTS SUPPORT_DB;

CREATE WAREHOUSE IF NOT EXISTS COMPUTE_WH
WAREHOUSE_SIZE = 'X-SMALL'
AUTO_SUSPEND = 60
AUTO_RESUME = TRUE;

USE DATABASE SUPPORT_DB;
USE WAREHOUSE COMPUTE_WH;

Step 2 — Create and Populate the Supply Desk

This desk simulates historic assist tickets. In manufacturing, this might be a stay desk synced out of your CRM, ticketing system, or knowledge pipeline.

CREATE TABLE IF NOT EXISTS SUPPORT_DB.PUBLIC.support_tickets (
    ticket_id VARCHAR(20),
    issue_category VARCHAR(100),
    user_query TEXT,
    decision TEXT,
    created_at TIMESTAMP_NTZ DEFAULT CURRENT_TIMESTAMP()
);

INSERT INTO SUPPORT_DB.PUBLIC.support_tickets (ticket_id, issue_category, user_query, decision) VALUES
('TKT-001', 'Connectivity',
'My web retains dropping each couple of minutes. The router lights look regular.',
'Agent checked line diagnostics. Discovered intermittent sign degradation on the coax line. Dispatched technician to exchange splitter. Problem resolved after {hardware} swap.'),

('TKT-002', 'Connectivity',
'Web may be very gradual throughout evenings however high quality within the morning.',
'Community congestion detected in buyer section throughout peak hours (6–10 PM). Upgraded buyer to a much less congested node. Speeds normalized inside 24 hours.'),

('TKT-003', 'Billing',
'I used to be charged twice for a similar month. Want a refund.',
'Duplicate billing confirmed because of fee gateway retry error. Refund of $49.99 issued. Buyer notified through e-mail. Root trigger patched in billing system.'),

('TKT-004', 'Machine Setup',
'My new router just isn't exhibiting up within the Wi-Fi listing on my laptop computer.',
'Router was broadcasting on 5GHz solely. Buyer laptop computer had outdated Wi-Fi driver that didn't assist 5GHz. Guided buyer to replace driver. Each 2.4GHz and 5GHz bands now seen.'),

('TKT-005', 'Connectivity',
'Frequent packet loss throughout video calls. Wired connection additionally affected.',
'Packet loss traced to defective ethernet port on modem. Changed modem beneath guarantee. Buyer confirmed secure connection post-replacement.'),

('TKT-006', 'Account',
'Can't log into the client portal. Password reset emails aren't arriving.',
'E-mail supply blocked by SPF report misconfiguration on buyer area. Suggested buyer to offer assist area. Reset e-mail delivered efficiently.'),

('TKT-007', 'Connectivity',
'Web unstable solely when microwave is working within the kitchen.',
'2.4GHz Wi-Fi interference brought on by microwave proximity to router. Beneficial switching router channel from 6 to 11 and enabling 5GHz band. Problem eradicated.'),

('TKT-008', 'Pace',
'Marketed pace is 500Mbps however I solely get round 120Mbps on speedtest.',
'Pace check confirmed 480Mbps at node. Buyer router restricted to 100Mbps because of Quick Ethernet port. Beneficial router improve. Put up-upgrade pace confirmed at 470Mbps.');

Step 3 — Create the Cortex Search Service

This single SQL command handles embedding technology, indexing, and hybrid retrieval setup routinely. The ON clause specifies which column to index for full-text and semantic search. ATTRIBUTES defines filterable metadata columns.

CREATE OR REPLACE CORTEX SEARCH SERVICE SUPPORT_DB.PUBLIC.support_search_svc
ON decision
ATTRIBUTES issue_category, ticket_id
WAREHOUSE = COMPUTE_WH
TARGET_LAG = '1 minute'
AS (
    SELECT
        ticket_id,
        issue_category,
        user_query,
        decision
    FROM SUPPORT_DB.PUBLIC.support_tickets
);

What occurs right here: Snowflake routinely generates vector embeddings for the decision column, builds each a key phrase index and a vector index, and exposes a unified hybrid retrieval endpoint. No embedding mannequin administration, no separate vector database.

You may confirm the service is energetic:

SHOW CORTEX SEARCH SERVICES IN SCHEMA SUPPORT_DB.PUBLIC;

Output:

Snowflake table

Step 4 — Question the Search Service from Python

Hook up with Snowflake and use the snowflake-core SDK to question the service:

First Set up required packages:

pip set up snowflake-snowpark-python snowflake-core

Now to search out your account particulars go to your account and click on on “Join a device to Snowflake”

Snowflake account creation
from snowflake.snowpark import Session
from snowflake.core import Root

# --- Connection config ---
connection_params = {
    "account": "YOUR_ACCOUNT_IDENTIFIER",  # e.g. abc12345.us-east-1
    "consumer": "YOUR_USERNAME",
    "password": "YOUR_PASSWORD",
    "function": "SYSADMIN",
    "warehouse": "COMPUTE_WH",
    "database": "SUPPORT_DB",
    "schema": "PUBLIC",
}

# --- Create Snowpark session ---
session = Session.builder.configs(connection_params).create()
root = Root(session)

# --- Reference the Cortex Search service ---
search_svc = (
    root.databases["SUPPORT_DB"]
    .schemas["PUBLIC"]
    .cortex_search_services["SUPPORT_SEARCH_SVC"]
)

def retrieve_context(question: str, category_filter: str = None, top_k: int = 3):
    """Run hybrid search towards the Cortex Search service."""
    filter_expr = {"@eq": {"issue_category": category_filter}} if category_filter else None

    response = search_svc.search(
        question=question,
        columns=["ticket_id", "issue_category", "user_query", "resolution"],
        filter=filter_expr,
        restrict=top_k,
    )
    return response.outcomes

# --- Take a look at retrieval ---
user_question = "Why is my web unstable?"
outcomes = retrieve_context(user_question, top_k=3)

print(f"n🔍 Question: {user_question}n")
print("=" * 60)

for i, r in enumerate(outcomes, 1):
    print(f"n[Result {i}]")
    print(f"  Ticket ID : {r['ticket_id']}")
    print(f"  Class  : {r['issue_category']}")
    print(f"  Consumer Question: {r['user_query']}")
    print(f"  Decision: {r['resolution'][:200]}...")

Output:

Output fetched via RAG

Step 5 — Construct the Full RAG Pipeline

Now move the retrieved context into Snowflake Cortex LLM (mistral-large or llama3.1-70b) to generate a grounded reply:

import json

def build_rag_prompt(user_question: str, retrieved_results: listing) -> str:
    """Format retrieved context into an LLM-ready immediate."""
    context_blocks = []
    for r in retrieved_results:
        context_blocks.append(
            f"- Ticket {r['ticket_id']} ({r['issue_category']}): "
            f"Buyer reported '{r['user_query']}'. "
            f"Decision: {r['resolution']}"
        )
    context_str = "n".be part of(context_blocks)

    return f"""You're a useful buyer assist assistant. Use ONLY the context under
to reply the client's query. Be particular and concise.

CONTEXT FROM HISTORICAL TICKETS:
{context_str}

CUSTOMER QUESTION: {user_question}
ANSWER:"""

def ask_rag_assistant(user_question: str, mannequin: str = "mistral-large2"):
    """Full RAG pipeline: retrieve → increase → generate."""
    print(f"n📡 Retrieving context for: '{user_question}'")
    outcomes = retrieve_context(user_question, top_k=3)
    print(f"   ✅ Retrieved {len(outcomes)} related tickets")

    immediate = build_rag_prompt(user_question, outcomes)

    safe_prompt = immediate.exchange("'", "'")
    sql = f"""
        SELECT SNOWFLAKE.CORTEX.COMPLETE(
            '{mannequin}',
            '{safe_prompt}'
        ) AS reply
    """

    end result = session.sql(sql).gather()
    reply = end result[0]["ANSWER"]
    return reply, outcomes

# --- Run the assistant ---
questions = [
    "Why is my internet unstable?",
    "I'm being charged incorrectly, what should I do?",
    "My router is not visible on my devices",
]

for q in questions:
    reply, ctx = ask_rag_assistant(q)
    print(f"n{'='*60}")
    print(f"❓ Buyer: {q}")
    print(f"n🤖 AI Assistant:n{reply.strip()}")
    print(f"n📎 Grounded in tickets: {[r['ticket_id'] for r in ctx]}")

Output:

Output

Key takeaway: The AI by no means generates generic solutions. Each response is traceable to particular historic tickets, dramatically decreasing hallucination threat and making outputs auditable.

Instance: Constructing Enterprise Search into Functions

What we’ll construct: 

A pure language assist ticket search interface — embedded instantly into an software — that lets brokers and clients search historic tickets utilizing plain English. No new infrastructure is required: this instance reuses the very same support_tickets desk and support_search_svc Cortex Search service created within the RAG part above.

This exhibits how the identical Cortex Search service can energy two completely completely different surfaces: an AI assistant on one hand, and a browsable search UI on the opposite.

Step 1 — Verify the Current Service is Lively

Confirm the service created within the earlier part remains to be working:

USE DATABASE SUPPORT_DB;

USE SCHEMA PUBLIC;

SHOW CORTEX SEARCH SERVICES IN SCHEMA RAG_SCHEMA;

Output:

Table

Step 2 — Construct the Enterprise Search Shopper

This module connects to the identical Snowpark session and support_search_svc service, and exposes a search perform with class filtering and ranked end result show — the form of interface you’d embed right into a assist portal, an inner data device, or an agent dashboard.

# enterprise_search.py

from snowflake.snowpark import Session
from snowflake.core import Root

# --- Connection config ---
connection_params = {
    "account": "YOUR_ACCOUNT_IDENTIFIER",  # e.g. abc12345.us-east-1
    "consumer": "YOUR_USERNAME",
    "password": "YOUR_PASSWORD",
    "function": "SYSADMIN",
    "warehouse": "COMPUTE_WH",
    "database": "SUPPORT_DB",
    "schema": "PUBLIC",
}

session = Session.builder.configs(connection_params).create()
root = Root(session)

# --- Identical service because the RAG instance — no new service wanted ---
search_svc = (
    root.databases["SUPPORT_DB"]
    .schemas["RAG_SCHEMA"]
    .cortex_search_services["SUPPORT_SEARCH_SVC"]
)

def search_tickets(question: str, class: str = None, top_k: int = 5) -> listing:
    """Pure language ticket search with non-compulsory class filter."""
    filter_expr = {"@eq": {"issue_category": class}} if class else None

    response = search_svc.search(
        question=question,
        columns=["ticket_id", "issue_category", "user_query", "resolution"],
        filter=filter_expr,
        restrict=top_k,
    )
    return response.outcomes

def display_tickets(question: str, outcomes: listing, filter_label: str = None):
    """Render search outcomes as a formatted ticket listing."""
    label = f" [{filter_label}]" if filter_label else ""

    print(f"n🔎 Search{label}: "{question}"")
    print(f"   {len(outcomes)} ticket(s) foundn")
    print("-" * 72)

    for i, r in enumerate(outcomes, 1):
        print(f"  #{i}  {r['ticket_id']}  |  Class: {r['issue_category']}")
        print(f"       Buyer: {r['user_query']}")
        print(f"       Decision: {r['resolution'][:160]}...n")

Step 3 — Run Pure Language Ticket Searches

# --- Search 1: Semantic question — no precise match wanted ---
outcomes = search_tickets("machine not connecting to the community")
display_tickets("machine not connecting to the community", outcomes)

Output:

Output
# --- Search 2: Class-filtered search (Billing solely) ---
outcomes = search_tickets(
    question="incorrect fee or refund request",
    class="Billing"
)

display_tickets(
    "incorrect fee or refund request",
    outcomes,
    filter_label="Billing"
)

2nd Output:

Output in Billing
# --- Search 3: Account & entry points ---
outcomes = search_tickets("cannot log in or entry my account", class="Account")
display_tickets("cannot log in or entry my account", outcomes, filter_label="Account")

Output:

Output in Accounting

Step 4 — Expose as a Flask Search API (Optionally available)

Wrap the search perform in a REST endpoint to embed it into any assist portal, inner device, or chatbot backend:

# app.py

from flask import Flask, request, jsonify
from enterprise_search import search_tickets

app = Flask(__name__)

@app.route("/tickets/search", strategies=["GET"])

def ticket_search():
    question    = request.args.get("q", "")
    class = request.args.get("class")        # non-compulsory filter
    top_k    = int(request.args.get("restrict", 5))

    if not question:
        return jsonify({"error": "Question parameter 'q' is required"}), 400

    outcomes = search_tickets(question, class=class, top_k=top_k)

    return jsonify({
        "question":    question,
        "class": class,
        "rely":    len(outcomes),
        "outcomes":  outcomes,
    })

if __name__ == "__main__":
    app.run(port=5001, debug=True)

Take a look at with curl:

# Free-text pure language search
curl "http://localhost:5001/tickets/search?q=web+retains+dropping&restrict=3"

Output:

JSON Output
# Filtered by class

curl "http://localhost:5001/tickets/search?q=charged+incorrectly&class=Billing"

Output:

JSON output

Key takeaway: The identical Cortex Search service that grounds the RAG assistant additionally powers a completely useful enterprise search UI — no duplication of infrastructure, no second index to keep up. One service definition delivers each experiences, and each keep routinely in sync as tickets are added or up to date.

The Enterprise Impression of Higher Retrieval

Poor knowledge search strategies quietly erode enterprise efficiency. Time is misplaced to repeated queries and rework. However, assist groups get entangled in resolving questions that ought to have been self-served within the first place. New hires and clients take longer to succeed in productiveness. AI initiatives stall when outputs can’t be trusted. 

In contrast, robust retrieval adjustments how organizations function. 

Groups transfer quicker as a result of solutions are simpler to search out. AI purposes carry out higher as a result of they’re grounded in related, present knowledge. Characteristic adoption improves as a result of customers can uncover and perceive capabilities with out friction. Help prices decline as search absorbs routine questions. 

Cortex Search turns retrieval from a background utility right into a strategic lever. It helps enterprises unlock the worth already current of their knowledge by making it accessible, searchable and usable at scale.  

Steadily Requested Questions

Q1. Why does conventional enterprise search fail in fashionable techniques?

A. It depends on key phrase matching and static indexes, which fail to seize intent and sustain with dynamic, distributed knowledge environments.

Q2. What makes hybrid retrieval simpler than conventional search?

A. It combines key phrase precision with semantic understanding, enabling quicker, extra related outcomes even for ambiguous or conversational queries.

Q3. How does Cortex Search enhance AI and enterprise purposes?

A. It offers correct, real-time retrieval that grounds AI responses and powers search experiences with out advanced infrastructure or guide tuning.

Dentsu’s world functionality heart, Dentsu World Companies (DGS), is shaping the longer term as an innovation engine. DGS has 5,600+ specialists focusing on digital platforms, efficiency advertising and marketing, product engineering, knowledge science, automation and AI, with media transformation on the core. DGS delivers AI-first, scalable options by dentsu’s community seamlessly integrating folks, know-how, and craft. They mix human creativity and superior know-how, constructing a various, future-focused group that adapts shortly to shopper wants whereas guaranteeing reliability, collaboration and excellence in each engagement.

DGS brings collectively world-class expertise, breakthrough know-how and daring concepts to ship impression at scale—for dentsu’s purchasers, its folks and the world. It’s a future-focused, industry-leading office the place expertise meets alternative. At DGS, staff can speed up their profession, collaborate with world groups and contribute to work that shapes the longer term. Discover out extra: Dentsu World Companies

Login to proceed studying and luxuriate in expert-curated content material.

Meta AI Releases EUPE: A Compact Imaginative and prescient Encoder Household Beneath 100M Parameters That Rivals Specialist Fashions Throughout Picture Understanding, Dense Prediction, and VLM Duties


Working highly effective AI in your smartphone isn’t only a {hardware} downside — it’s a mannequin structure downside. Most state-of-the-art imaginative and prescient encoders are huge, and while you trim them down to suit on an edge system, they lose the capabilities that made them helpful within the first place. Worse, specialised fashions are likely to excel at one sort of activity — picture classification, say, or scene segmentation — however collapse while you ask them to do one thing outdoors their lane.

Meta’s AI analysis groups are actually proposing a special path. They launched the Environment friendly Common Notion Encoder (EUPE): a compact imaginative and prescient encoder that handles various imaginative and prescient duties concurrently while not having to be massive.

The Core Downside: Specialists vs. Generalists

To know why EUPE issues, it helps to know how imaginative and prescient encoders work and why specialization is an issue.

A imaginative and prescient encoder is the a part of a pc imaginative and prescient mannequin that converts uncooked picture pixels right into a compact illustration — a set of characteristic vectors — that downstream duties (like classification, segmentation, or answering questions on a picture) can use. Consider it because the ‘eyes’ of an AI pipeline.

Fashionable basis imaginative and prescient encoders are educated with particular targets, which supplies them an edge particularly domains. For instance:

  • CLIP and SigLIP 2 are educated on text-image pairs. They’re robust at picture understanding and vision-language modeling, however their efficiency on dense prediction duties (which require spatially exact, pixel-level options) typically falls under expectations.
  • DINOv2 and its successor DINOv3 are self-supervised fashions that be taught distinctive structural and geometric descriptors, making them robust at dense prediction duties like semantic segmentation and depth estimation. However they lack passable vision-language capabilities.
  • SAM (Phase Something Mannequin) achieves spectacular zero-shot segmentation via coaching on large segmentation datasets, however once more falls quick on vision-language duties.

For an edge system — a smartphone or AR headset — that should deal with all of those activity varieties concurrently, the standard answer is to deploy a number of encoders directly. That shortly turns into compute-prohibitive. The choice is accepting {that a} single encoder will underperform in a number of domains.

Earlier Makes an attempt: Why Agglomerative Strategies Fell Quick on Environment friendly Backbones

Researchers have tried to mix the strengths of a number of specialist encoders via a household of strategies referred to as agglomerative multi-teacher distillation. The fundamental concept: practice a single pupil encoder to concurrently mimic a number of instructor fashions, every of which is a site professional.

AM-RADIO and its follow-up RADIOv2.5 are maybe essentially the most well-known examples of this strategy. They confirmed that agglomerative distillation can work effectively for giant encoders — fashions with greater than 300 million parameters. However the EUPE analysis demonstrates a transparent limitation: while you apply the identical recipe to environment friendly backbones, the outcomes degrade considerably. RADIOv2.5-B, the ViT-B-scale variant, has important gaps in comparison with area specialists on dense prediction and VLM duties.

One other agglomerative methodology, DUNE, merges 2D imaginative and prescient and 3D notion academics via heterogeneous co-distillation, however equally struggles on the environment friendly spine scale.

The foundation trigger, the analysis workforce argue, is capability. Environment friendly encoders merely don’t have sufficient representational capability to straight take up various characteristic representations from a number of specialist academics and unify them right into a common illustration. Making an attempt to take action in a single step produces a mannequin that’s mediocre throughout the board.

https://arxiv.org/pdf/2603.22387

EUPE’s Reply: Scale Up First, Then Scale Down

The important thing perception behind EUPE is a precept named ‘first scaling up after which cutting down.

As a substitute of distilling straight from a number of domain-expert academics right into a small pupil, EUPE introduces an intermediate mannequin: a big proxy instructor with sufficient capability to unify the information from all of the area specialists. This proxy instructor then transfers its unified, common information to the environment friendly pupil via distillation.

The complete pipeline has three phases:

Stage 1 — Multi-Instructor Distillation into the Proxy Mannequin. A number of massive basis encoders function academics concurrently, processing label-free photos at their native resolutions. Every instructor outputs a category token and a set of patch tokens. The proxy mannequin — a 1.9B parameter mannequin educated with 4 register tokens — is educated to imitate all academics directly. The chosen academics are:

  • PEcore-G (1.9B parameters), chosen because the area professional for zero-shot picture classification and retrieval
  • PElang-G (1.7B parameters), which the analysis workforce discovered is essential for vision-language modeling, significantly OCR efficiency
  • DINOv3-H+ (840M parameters), chosen because the area professional for dense prediction

To stabilize coaching, instructor outputs are normalized by subtracting the per-coordinate imply and dividing by the usual deviation, computed as soon as over 500 iterations earlier than coaching begins and saved mounted thereafter. That is intentionally less complicated than the complicated PHI-S normalization utilized in RADIOv2.5, and avoids the cross-GPU reminiscence overhead of computing normalization statistics on-the-fly.

Stage 2 — Fastened-Decision Distillation into the Environment friendly Scholar. With the proxy mannequin now serving as a single common instructor, the goal environment friendly encoder is educated at a set decision of 256×256. This mounted decision makes coaching computationally environment friendly, permitting an extended studying schedule: 390,000 iterations with a batch measurement of 8,192, cosine studying charge schedule, a base studying charge of 2e-5, and weight decay of 1e-4. Customary information augmentation applies: random resized cropping, horizontal flipping, shade jittering, Gaussian blur, and random solarization. For the distillation loss, the category token loss makes use of cosine similarity, whereas the patch token loss combines cosine similarity (weight α=0.9) and clean L1 loss (weight β=0.1). Adapter head modules — 2-layer MLPs — are appended to the coed to match every instructor’s characteristic dimension. If pupil and instructor patch token spatial dimensions differ, 2D bicubic interpolation is utilized to align them.

Stage 3 — Multi-Decision Finetuning. Ranging from the Stage 2 checkpoint, the coed undergoes a shorter finetuning part utilizing a picture pyramid of three scales: 256, 384, and 512. The scholar and the proxy instructor independently and randomly choose one scale per iteration — so they might course of the identical picture at completely different resolutions. This forces the coed to be taught representations that generalize throughout spatial granularities, accommodating downstream duties that function at numerous resolutions. This stage runs for 100,000 iterations at a batch measurement of 4,096 and base studying charge of 1e-5. It’s deliberately shorter as a result of multi-resolution coaching is computationally expensive — one iteration in Stage 3 takes roughly twice so long as in Stage 2.

Coaching Information. All three phases use the identical DINOv3 dataset, LVD-1689M, which offers balanced protection of visible ideas from the net alongside high-quality public datasets together with ImageNet-1k. The sampling likelihood from ImageNet-1k is 10%, with the remaining 90% from LVD-1689M. In an ablation examine, coaching on LVD-1689M constantly outperformed coaching on MetaCLIP (2.5B photos) on almost all benchmarks — regardless of MetaCLIP being roughly 800M photos bigger — indicating increased information high quality in LVD.

https://arxiv.org/pdf/2603.22387

An Vital Detrimental Outcome: Not All Lecturers Mix Effectively

One of many extra virtually helpful findings issues instructor choice. Intuitively, including extra robust academics ought to assist. However the analysis workforce discovered that together with SigLIP2-G alongside PEcore-G and DINOv3-H+ considerably degrades OCR efficiency. On the proxy mannequin stage, TextVQA drops from 56.2 to 53.2; on the ViT-B pupil stage, it drops from 48.6 to 44.8. The analysis groups’ speculation: having two CLIP-style fashions (PEcore-G and SigLIP2-G) within the instructor set concurrently causes characteristic incompatibility. PElang-G, a language-focused mannequin derived from PEcore-G via alignment with language fashions, proved a much better complement — bettering OCR and common VLM efficiency with out sacrificing picture understanding or dense prediction.

What the Numbers Say

The ablation research validate the three-stage design. Distilling straight from a number of academics to an environment friendly pupil (“Stage 2 solely”) yields poor VLM efficiency, particularly on OCR-type duties, and poor dense prediction. Including Stage 1 (the proxy mannequin) considerably improves VLM duties — TextVQA rises from 46.8 to 48.3, and Realworld from 53.5 to 55.1 — however nonetheless lags on dense duties. Stage 1+3 (skipping Stage 2) offers the strongest dense prediction outcomes (SPair: 53.3, NYUv2: 0.388) however leaves VLM gaps and is dear to run for a full schedule. The complete three-stage pipeline achieves the perfect total steadiness.

On the principle ViT-B benchmark, EUPE-ViT-B constantly stands out:

  • Picture understanding: EUPE achieves 84.1 on IN1k-KNN, outperforming PEcore-B (79.7), SigLIP2-B (83.2), and DINOv3-ViT-B (83.0). On IN1k-ZS (zero-shot), it scores 79.7, outperforming PEcore-B (78.4) and SigLIP2-B (78.2).
  • Dense prediction: EUPE achieves 52.4 mIoU on ADE20k, outperforming the dense prediction professional DINOv3-ViT-B (51.8). On SPair-71k semantic correspondence, it scores 51.3, matching DINOv3-ViT-B.
  • Imaginative and prescient-language modeling: EUPE outperforms each PEcore-B and SigLIP2-B on RealworldQA (55.5 vs. 52.9 and 52.5) and GQA (67.3 vs. 65.6 and 65.2), whereas staying aggressive on TextVQA, SQA, and POPE.
  • Vs. agglomerative strategies: EUPE outperforms RADIOv2.5-B and DUNE-B on all VLM duties and most dense prediction duties by important margins.

What the Options Really Look Like

The analysis additionally consists of qualitative characteristic visualization utilizing PCA projection of patch tokens into RGB area — a method that reveals the spatial and semantic construction an encoder has realized. The outcomes are telling:

  • PEcore-B and SigLIP2-B patch tokens include semantic info however are usually not spatially constant, resulting in noisy representations.
  • DINOv3-ViT-B has extremely sharp, semantically coherent options, however lacks fine-grained discrimination (meals and plates find yourself with comparable representations within the final row instance).
  • RADIOv2.5-B options are overly delicate, breaking semantic coherence — for instance, black canine fur merges visually with the background.
  • EUPE-ViT-B combines semantic coherence, positive granularity, complicated spatial construction, and textual content consciousness concurrently — capturing the perfect qualities throughout all area specialists directly.

A Full Household of Edge-Prepared Fashions

EUPE is an entire household spanning two structure varieties:

  • ViT household: ViT-T (6M parameters), ViT-S (21M), ViT-B (86M)
  • ConvNeXt household: ConvNeXt-Tiny (29M), ConvNeXt-Small (50M), ConvNeXt-Base (89M)

All fashions are underneath 100M parameters. Inference latency is measured on iPhone 15 Professional CPU through ExecuTorch-exported fashions. At 256×256 decision: ViT-T runs in 6.8ms, ViT-S in 17.1ms, and ViT-B in 55.2ms. The ConvNeXt variants have decrease FLOPs than ViTs of comparable measurement, however don’t essentially obtain decrease latency on CPU — as a result of convolutional operations are sometimes much less environment friendly on CPU structure in comparison with the extremely optimized matrix multiplication (GEMM) operations utilized in ViTs.

For the ConvNeXt household, EUPE constantly outperforms the DINOv3-ConvNeXt household of the identical sizes throughout Tiny, Small, and Base variants on dense prediction, whereas additionally unlocking higher VLM functionality — significantly for OCR and vision-centric duties — that DINOv3-ConvNeXt totally lacks.

Key Takeaways

  • One encoder to rule all of them. EUPE is a single compact imaginative and prescient encoder (underneath 100M parameters) that matches or outperforms specialised domain-expert fashions throughout picture understanding, dense prediction, and vision-language modeling — duties that beforehand required separate, devoted encoders.
  • Scale up earlier than you scale down. The core innovation is a three-stage “proxy instructor” distillation pipeline: first combination information from a number of massive professional fashions right into a 1.9B parameter proxy, then distill from that single unified instructor into an environment friendly pupil — fairly than straight distilling from a number of academics directly.
  • Instructor choice is a design determination, not a given. Including extra academics doesn’t all the time assist. Together with SigLIP2-G alongside PEcore-G degraded OCR efficiency considerably. PElang-G turned out to be the correct VLM complement — a discovering with direct sensible implications for anybody constructing multi-teacher distillation pipelines.
  • Constructed for actual edge deployment. The complete EUPE household spans six fashions throughout ViT and ConvNeXt architectures. The smallest, ViT-T, runs in 6.8ms on iPhone 15 Professional CPU. All fashions are exported through ExecuTorch and out there on Hugging Face — prepared for on-device integration, not simply benchmarking.
  • Information high quality beats information amount. In ablation experiments, coaching on LVD-1689M outperformed coaching on MetaCLIP throughout almost all benchmarks — regardless of MetaCLIP containing roughly 800 million extra photos. A helpful reminder that greater datasets don’t routinely imply higher fashions.

Try the Paper, Mannequin Weight and Repo.  Additionally, be at liberty to observe us on Twitter and don’t neglect to affix our 120k+ ML SubReddit and Subscribe to our E-newsletter. Wait! are you on telegram? now you may be a part of us on telegram as effectively.

Must companion with us for selling your GitHub Repo OR Hugging Face Web page OR Product Launch OR Webinar and many others.? Join with us


This is when the Galaxy S25 might lastly get steady One UI 8.5

0


Joe Maring / Android Authority

TL;DR

  • A leak suggests Samsung will launch the steady One UI 8.5 replace for the Galaxy S25 sequence on April 30, beginning in South Korea.
  • Worldwide customers can seemingly count on the rollout to start on Might 4, 2026, based mostly on earlier launch patterns.
  • The replace follows a prolonged interval of eight beta variations and weeks of exclusivity for the newer Galaxy S26 sequence.

Galaxy S25 homeowners, your look forward to One UI 8.5 may lastly be coming to an finish soon-ish. After releasing eight One UI 8.5 betas, Samsung is now anticipated to lastly launch One UI 8.5 steady for the Galaxy S25 sequence on the finish of this month, in keeping with a brand new leak.

As per Tarun Vats on X, Samsung is predicted to launch One UI 8.5 steady for the Galaxy S25 sequence on April 30, beginning with its house area of South Korea. Presuming the corporate follows final yr’s sample, worldwide rollout might begin from Might 4, 2026, onwards.

Galaxy S25 series Expected One UI 8.5 stable release date

Don’t need to miss the very best from Android Authority?

google preferred source badge light@2xgoogle preferred source badge dark@2x

One UI 8.5 has been round in its steady type with the launch of the Galaxy S26 sequence. It’s been many weeks since then, and we frankly anticipated Samsung to maintain software program exclusivity for under a few weeks earlier than gracing its now-older flagship with the most recent Android replace. However late final month, we obtained yet one more One UI 8.5 beta for the Galaxy S25 sequence, bringing the whole to eight.

Whereas we recognize that Samsung is doing its finest to repair bugs, it’s nonetheless a bit disappointing to see the corporate’s ex-flagship nonetheless working an outdated Android model, provided that Android 16 has been out since June 2025 and we’ve nearly reached steady Android 17 too. Even with this leak, we’re nonetheless a couple of month away from a One UI 8.5 steady launch (based mostly on Android 16), and it’s frankly embarrassing given how flagships from different firms have lengthy been up to date to Android 16.

We hope Samsung allocates extra sources, fixes bugs, and rolls out a steady One UI 8.5 replace for the Galaxy S25 sequence quickly.

Thanks for being a part of our group. Learn our Remark Coverage earlier than posting.