Friday, March 13, 2026
Home Blog Page 111

Air for Tomorrow: Mapping the Digital Air-High quality Panorama, from Repositories and Information Sorts to Starter Code

0


highway in Lao PDR. The college is 200 meters away. Site visitors roars, smoke from burning rubbish drifts throughout the trail, and kids stroll straight by way of it. What are they respiration at present? With out native knowledge, nobody actually is aware of. 

Throughout East Asia and the Pacific, 325 million kids [1] breathe poisonous air daily, generally at ranges 10 instances above secure limits. The harm is usually silent: affected lungs, bronchial asthma, however it might result in missed faculty days in acute circumstances. The futures are at stake. In the long term, the well being methods are strained, and economies should bear the prices.

In lots of circumstances, air high quality knowledge just isn’t even out there.

No displays. No proof. No safety. 

On this second a part of the weblog sequence [2], we examine the information repositories the place helpful air-quality knowledge is out there, tips on how to import them, and tips on how to get them up and operating in your pocket book. We might additionally demystify knowledge codecs equivalent to GeoJSON, Parquet/GeoParquet, NetCDF/HDF5, COG, GRIB, and Zarr so you’ll be able to decide the correct software for the job. We are constructing it up in order that within the subsequent half, we will go step-by-step by way of how we developed an open-source air high quality mannequin. 

In the previous couple of years, there was a big push to generate and use air-quality knowledge. These knowledge come from totally different sources, and their high quality varies accordingly. Just a few repositories will help quantify them: regulatory stations for floor reality, neighborhood sensors to know hyperlocal variations, satellites for regional context, and mannequin reanalyses for estimates (Determine 2). The excellent news: most of that is open. The higher information: the code to get began is comparatively brief. 

Determine 2: Hearth hotspot as on 20.04.2024 and the interpolated density map created utilizing a number of knowledge sources. Supply: @UNICEF. All rights reserved.

Repository quick-starts (with minimal Python) 

On this part, we transfer from ideas to observe. Under, we stroll by way of a set of generally used open-source repositories and present the smallest doable code it’s essential begin pulling knowledge from every of them. All examples assume Python ≥3.10 with pip set up as wanted. 

For every numbered repository, one can find: 

  • a brief description of what the information supply is and the way it’s maintained, 
  • typical use-cases (when this supply is an efficient match), 
  • tips on how to entry it (API keys, sign-up notes, or direct URLs), and 
  • a minimal Python code snippet to extract knowledge. 

Think about this as a sensible information the place you’ll be able to skim the descriptions, decide the supply that matches your downside, after which adapt the code to plug immediately into your individual evaluation or mannequin pipeline. 

Tip: Maintain secrets and techniques out of code. Use setting variables for tokens (e.g., export AIRNOW_API_KEY=…). 

1) OpenAQ (international floor measurements; open API) 

OpenAQ [3] is an open-source knowledge platform that hosts international knowledge for air high quality knowledge, equivalent to PM2.5, PM10, and O3. They supply air high quality knowledge by partnering with varied governmental companions, neighborhood companions, and air high quality sensor corporations equivalent to Air Gradient, IQAir, amongst others.

Nice for: fast cross-country pulls, harmonised items/metadata, reproducible pipelines. 

Join an OpenAQ API key at https://discover.openaq.org. After signing up, discover your API key in your settings. Use this key to authenticate requests. 

!pip set up openaq pandas
import pandas as pd
from pandas import json_normalize
from openaq import OpenAQ
import datetime
from datetime import timedelta
import geopandas as gpd
import requests
import time
import json

# comply with the quickstart to get the api key https://docs.openaq.org/using-the-api/quick-start
api_key = '' #enter you API Key earlier than executing
consumer = OpenAQ(api_key=api_key) #use the API key generated earlier

# get the places of each sensors within the chosen international locations codes: https://docs.openaq.org/assets/international locations
places = consumer.places.record(
countries_id=[68,111],
restrict = 1000
)

data_locations = places.dict()
df_sensors_country = json_normalize(data_locations ['results'])
df_sensors_exploded = df_sensors_country.explode('sensors')
df_sensors_exploded['sensor_id']=df_sensors_exploded['sensors'].apply(lambda x: x['id'])
df_sensors_exploded['sensor_type']=df_sensors_exploded['sensors'].apply(lambda x: x['name'])
df_sensors_pm25 = df_sensors_exploded[df_sensors_exploded['sensor_type'] == "pm25 µg/m³"]
df_sensors_pm25

# undergo every location and extract the  hourly measurements
df_concat_aq_data=pd.DataFrame()
to_date = datetime.datetime.now()
from_date = to_date - timedelta(days=2) # get the previous 2 days knowledge
sensor_list = df_sensors_pm25.sensor_id

for sensor_id in sensor_list[0:5]:
    print("-----")
    response = consumer.measurements.record(
        sensors_id= sensor_id,
        datetime_from = from_date,
        datetime_to = to_date,
        restrict = 500 )
    print(response)

    data_measurements = response.dict()
    df_hourly_data = json_normalize(data_measurements ['results'])
    df_hourly_data["sensor_id"] = sensor_id
    if len(df_hourly_data) > 0:
        df_concat_aq_data=pd.concat([df_concat_aq_data,df_hourly_data])
        df_concat_aq_data = df_concat_aq_data[["sensor_id","period.datetime_from.utc","period.datetime_to.utc","parameter.name","value"]]

    df_concat_aq_data

2) EPA AQS Information Mart (U.S. regulatory archive; token wanted) 

The EPA AQS Information Mart [4] is a U.S. regulatory knowledge archive that hosts quality-controlled air-quality measurements from 1000’s of monitoring stations throughout the nation. It offers long-term data for standards pollution equivalent to PM₂․₅, PM₁₀, O₃, NO₂, SO₂, and CO, together with detailed website metadata and QA flags, and is freely accessible through an API when you register and acquire an entry token. It offers meteorological knowledge as properly. 

Nice for: authoritative QA/QC-d U.S. knowledge. 

Join an AQS Information Mart account on the US EPA web site at: https://aqs.epa.gov/aqsweb/paperwork/data_api.html
Create a .env file in your setting and add your credentials, together with AQS e-mail and AQS key.

# pip set up requests pandas 

import os, requests, pandas as pd 
AQS_EMAIL = os.getenv("AQS_EMAIL") 
AQS_KEY   = os.getenv("AQS_KEY") 

url = "https://aqs.epa.gov/knowledge/api/sampleData/byState" 
params = {"e-mail": AQS_EMAIL, "key": AQS_KEY, "param": "88101", "b date":"20250101", "edate": "20250107", "state": "06"} 
r = requests.get(url, params=params, timeout=60) 

df = pd.json_normalize(r.json()["Data"]) 
print(df[["state_name","county_name","date_local","sample_measurement","units_of_measure"]].head()) 

3) AirNow (U.S. real-time indices; API key) 

AirNow [5] is a U.S. authorities platform that gives close to real-time air-quality index (AQI) info primarily based on regulatory monitoring knowledge. It publishes present and forecast AQI values for pollution equivalent to PM₂․₅ and O₃, together with class breakpoints (“Good”, “Reasonable”, and so on.) which are straightforward to speak to the general public. Information may be accessed programmatically through the AirNow API when you register and acquire an API key. 

Nice for: wildfire and public-facing AQI visuals. 

Register for an AirNow API account through the AirNow API portal: https://docs.airnowapi.org/ 

From the Log In web page, choose “Request an AirNow API Account” and full the registration type together with your e-mail and fundamental particulars. After you activate your account, you will discover your API key in your AirNow API dashboard; use this key to authenticate all calls to the AirNow net providers. 

import os, requests, pandas as pd 

API_KEY = os.getenv("AIRNOW_API_KEY") 
url = "https://www.airnowapi.org/aq/statement/latLong/present/" 
params = {"format":"software/json", "latitude": 37.7749, "longitude": -122.4194, "distance":25, "API_KEY": API_KEY} 
df = pd.DataFrame(requests.get(url, params=params, timeout=30).json()) 

print(df[["ParameterName", "AQI" ,"Category.Name ","DateObserved", "HourObserved"]]) 

4) Copernicus Environment Monitoring Service (CAMS; Environment Information Retailer)

The Copernicus Environment Monitoring Service [6], carried out by ECMWF for the EU’s Copernicus programme, offers international reanalyses and near-real-time forecasts of atmospheric composition. By the Environment Information Retailer (ADS), you’ll be able to entry gridded fields for aerosols, reactive gases (O₃, NO₂, and so on.), greenhouse gases and associated meteorological variables, with multi-year data appropriate for each analysis and operational purposes. All CAMS merchandise within the ADS are open and freed from cost, topic to accepting the Copernicus licence. 

Nice for: international background fields (aerosols & hint gases), forecasts and reanalyses. 

Find out how to register and get API entry 

  1. Go to the Environment Information Retailer: https://advertisements.environment.copernicus.eu
  1. Click on Login / Register within the top-right nook and create a (free) Copernicus/ECMWF account. 
  1. After confirming your e-mail, log in and go to your profile web page to seek out your ADS API key (UID + key). 
  1. Comply with the ADS “Find out how to use the API” directions to create a configuration file (usually ~/.cdsapirc) with: 
  1. url: https://advertisements.environment.copernicus.eu/api 
  1. key: : 
  1. On the internet web page of every CAMS dataset you wish to use, go to the Obtain knowledge tab and settle for the licence on the backside as soon as; solely then will API requests for that dataset succeed. 

As soon as that is arrange, you need to use the usual cdsapi Python consumer to programmatically obtain CAMS datasets from the ADS. 

# pip set up cdsapi xarray cfgrib 

import cdsapi 
c = cdsapi.Shopper() 

# Instance: CAMS international reanalysis (EAC4) complete column ozone (toy instance) 
c.retrieve( 
    "cams-global-reanalysis-eac4", 
    {"variable":"total_column_ozone","date":"2025-08-01/2025-08-02","time":["00:00","12:00"], 
     "format":"grib"}, "cams_ozone.grib") 

5) NASA Earthdata (LAADS DAAC / GES DISC; token/login) 

NASA Earthdata [7] offers unified sign-on entry to a variety of Earth science knowledge, together with satellite tv for pc aerosol and hint fuel merchandise which are essential for air-quality purposes. Two key centres for atmospheric composition are: 

  • LAADS DAAC (Stage-1 and Environment Archive and Distribution System DAAC), which hosts MODIS, VIIRS and different instrument merchandise (e.g., AOD, cloud, fireplace, radiance). 
  • GES DISC (Goddard Earth Sciences Information and Info Companies Middle), which serves mannequin and satellite tv for pc merchandise equivalent to MERRA-2 reanalysis, OMI, TROPOMI, and associated atmospheric datasets. 

Most of those datasets are free to make use of however require a NASA Earthdata Login; downloads are authenticated both through HTTP fundamental auth (username/password saved in .netrc) or through a private entry token (PAT) in request headers. 

Nice for: MODIS/VIIRS AOD, MAIAC, TROPOMI trace-gas merchandise.  

Find out how to register and get API/obtain entry: 

  1. Create a NASA Earthdata Login account at: 
    https://urs.earthdata.nasa.gov 
  1. Verify your e-mail and log in to your Earthdata profile. 
  1. Beneath your profile, generate a private entry token (PAT). Save this token securely; you need to use it in scripts through an Authorization: Bearer  header or in instruments that help Earthdata tokens. 
  1. For traditional wget/curl-based downloads, you’ll be able to alternatively create a ~/.netrc file to retailer your Earthdata username and password, for instance: 
machine urs.earthdata.nasa.gov 
login  
password 

Then set file permissions to user-only (chmod 600 ~/.netrc) so command-line instruments can authenticate robotically. 

  1. For LAADS DAAC merchandise, go to https://ladsweb.modaps.eosdis.nasa.gov, log in together with your Earthdata credentials, and use the Search & Obtain interface to construct obtain URLs; you’ll be able to copy the auto-generated wget/curl instructions into your scripts. 
  1. For GES DISC datasets, begin from https://disc.gsfc.nasa.gov, select a dataset (e.g., MERRA-2), and use the “Information Entry” or “Subset/Get Information” instruments. The positioning can generate script templates (Python, wget, and so on.) that already embody the right endpoints for authenticated entry. 

As soon as your Earthdata Login and token are arrange, LAADS DAAC and GES DISC behave like customary HTTPS APIs: you’ll be able to name them from Python (e.g., with requests, xarray + pydap/OPeNDAP, or s3fs for cloud buckets) utilizing your credentials or token for authenticated, scriptable downloads. 

#Downloads through HTTPS with Earthdata login. 

# pip set up requests 
import requests 
url = "https://ladsweb.modaps.eosdis.nasa.gov/archive/allData/6/MCD19A2/2025/214/MCD19A2.A2025214.h21v09.006.2025xxxxxx.hdf" 

# Requires a legitimate token cookie; advocate utilizing .netrc or requests.Session() with auth 
# See NASA docs for token-based obtain; right here we solely illustrate the sample: 
# s = requests.Session(); s.auth = (USERNAME, PASSWORD); r = s.get(url) 

6) STAC catalogues (search satellites programmatically) 

SpatioTemporal Asset Catalog (STAC) [8] is an open specification for describing geospatial property, equivalent to satellite tv for pc scenes, tiles, and derived merchandise, in a constant, machine-readable approach. As an alternative of manually shopping obtain portals, you question a STAC API with filters like time, bounding field, cloud cowl, platform (e.g., Sentinel-2, Landsat-8, Sentinel-5P), or processing degree, and get again JSON gadgets with direct hyperlinks to COGs, NetCDF, Zarr, or different property.  

Nice for: uncover and stream property (COGs/NetCDF) with out bespoke APIs and works properly with Sentinel-5P, Landsat, Sentinel-2, extra. 

Find out how to register and get API entry: 
STAC itself is simply an ordinary; entry is dependent upon the particular STAC API you utilize: 

  • Many public STAC catalogues (e.g., demo or analysis endpoints) are absolutely open and require no registration—you’ll be able to hit their /search endpoint immediately with HTTP POST/GET. 
  • Some cloud platforms that expose STAC (for instance, industrial or giant cloud suppliers) require you to create a free account and acquire credentials earlier than you’ll be able to learn the underlying property (e.g., blobs in S3/Blob storage), though the STAC metadata is open. 

A generic sample you’ll be able to describe is: 

  1. Decide a STAC API endpoint for the satellite tv for pc knowledge you care about (typically documented as one thing alongside the strains of https:///stac or …/stac/search). 
  1. If the supplier requires sign-up, create an account of their portal and acquire the API key or storage credentials they advocate (this may be a token, SAS URL, or cloud entry position). 
  1. Use a STAC consumer library in Python (for instance, pystac-client) to go looking {the catalogue}: 
# pip set up pystac-client 
from pystac_client import Shopper 

api = Shopper.open("https://instance.com/stac") 
search = api.search( 
    collections=["sentinel-2-l2a"], 
    bbox=[102.4, 17.8, 103.0, 18.2],   # minx, miny, maxx, maxy 
    datetime="2024-01-01/2024-01-31", 
    question={"eo:cloud_cover": {"lt": 20}}, 
    )
gadgets = record(search.get_items()) 
first_item = gadgets[0] 
property = first_item.property  # e.g., COGs, QA bands, metadata 
  1. For every returned STAC merchandise, comply with the asset href hyperlinks (typically HTTPS URLs or cloud URIs like s3://…) and skim them with the applicable library (rasterio/xarray/zarr and so on.). If credentials are wanted, configure them through setting variables or your cloud SDK as per the supplier’s directions. 

As soon as arrange, STAC catalogues provide you with a uniform, programmatic strategy to search and retrieve satellite tv for pc knowledge throughout totally different suppliers, with out rewriting your search logic each time you turn from one archive to a different. 

# pip set up pystac-client planetary-computer rasterio 
from pystac_client import Shopper 
from shapely.geometry import field, mapping 
import geopandas as gpd 

catalog = Shopper.open("https://earth-search.aws.element84.com/v1") 
aoi = mapping(field(-0.3, 5.5, 0.3, 5.9))  # bbox round Accra
search = catalog.search(collections=["sentinel-2-l2a"], intersects=aoi, restrict=5) 
gadgets = record(search.get_items()) 
for it in gadgets: 
    print(it.id, record(it.property.keys())[:5])   # e.g., "B04", "B08", "SCL", "visible" 

It’s preferrable to make use of STAC the place doable as they supply clear metadata, cloud-optimised property, and straightforward filtering by time/house. 

7) Google Earth Engine (GEE; quick prototyping at scale) 

Google Earth Engine [9] is a cloud-based geospatial evaluation platform that hosts a big catalogue of satellite tv for pc, local weather, and land-surface datasets (e.g., MODIS, Landsat, Sentinel, reanalyses) and allows you to course of them at scale with out managing your individual infrastructure. You write brief scripts in JavaScript or Python, and GEE handles the heavy lifting like knowledge entry, tiling, reprojection, and parallel computation thus making it superb for quick prototyping, exploratory analyses, and instructing. 

Nevertheless, GEE itself just isn’t open supply: it’s a proprietary, closed platform the place the underlying codebase just isn’t publicly out there. This has implications for open, reproducible workflows mentioned within the first Air for Tomorrow weblog [add link]: 
 
Nice for: testing fusion/downscaling over a metropolis/area utilizing petabyte-scale datasets. 
 
Find out how to register and get entry 

  1. Go to the Earth Engine sign-up web page: https://earthengine.google.com
  1. Sign up with a Google account and full the non-commercial sign-up type, describing your meant use (analysis, schooling, or private, non-commercial tasks). 
  1. As soon as your account is accepted, you’ll be able to: 
  • use the browser-based Code Editor to jot down JavaScript Earth Engine scripts; and 
  • allow the Earth Engine API in Google Cloud and set up the earthengine-api Python bundle (pip set up earthengine-api) to run workflows from Python notebooks. 
  1. When sharing your work, take into account exporting key intermediate outcomes (e.g., GeoTIFF/COG, NetCDF/Zarr) and documenting your processing steps in open-source code in order that others can re-create the evaluation with out relying solely on GEE. 

When used this manner, Earth Engine turns into a robust “fast laboratory” for testing concepts, which you’ll be able to then harden into absolutely open, moveable pipelines for manufacturing and long-term stewardship. 

# pip set up earthengine-api 
import ee 

ee.Initialize()  # first run: ee.Authenticate() in a console 
s5p = ee.ImageCollection('COPERNICUS/S5P/OFFL/L3_NO2').choose('NO2_column_number_density') 
       .filterDate('2025-08-01', '2025-08-07').imply() 

print(s5p.getInfo()['bands'][0]['id']) 

# Exporting and visualization occur inside GEE; you'll be able to pattern to a grid then .getDownloadURL() 

8) HIMAWARI

Himawari-8 and Himawari-9 are geostationary meteorological satellites operated by the Japan Meteorological Company (JMA). Their Superior Himawari Imager (AHI) offers multi-band seen, near-infrared and infrared imagery over East Asia and the western–central Pacific, with full-disk scans each 10 minutes and even quicker refresh over goal areas. This high-cadence view is extraordinarily helpful for monitoring smoke plumes, mud, volcanic eruptions, convective storms and the diurnal evolution of clouds—precisely the sorts of processes that modulate near-surface air high quality. 
 
Nice for: monitoring diurnal haze/smoke plumes and fireplace occasions, producing high-frequency AOD to fill polar-orbit gaps, and fast situational consciousness for cities throughout SE/E Asia (through JAXA P-Tree L3 merchandise). 

Find out how to entry and register 

Choice A – Open archive through NOAA on AWS (no sign-up required) 

  1. Browse the dataset description on the AWS Registry of Open Information: https://registry.opendata.aws/noaa-himawari/ 
  1. Himawari-8 and Himawari-9 imagery are hosted in public S3 buckets (s3://noaa-himawari8/ and s3://noaa-himawari9/). As a result of the buckets are world-readable, you’ll be able to record or obtain information anonymously, for instance: 

aws s3 ls --no-sign-request s3://noaa-himawari9/ 

or entry particular person objects through HTTPS (e.g., https://noaa-himawari9.s3.amazonaws.com/…). 

  1. For Python workflows, you need to use libraries like s3fs, fsspec, xarray, or rasterio to stream knowledge immediately from these buckets with out prior registration, holding in thoughts the attribution steering from JMA/NOAA if you publish outcomes. 

Choice B – JAXA Himawari Monitor / P-Tree (analysis & schooling account) 

  1. Go to the JAXA Himawari Monitor / P-Tree portal: 
    https://www.eorc.jaxa.jp/ptree/ 
  1. Click on Person Registration / Account request and skim the “Precautions” and “Phrases of Use”. Information entry is restricted to non-profit functions equivalent to analysis and schooling; industrial customers are directed to the Japan Meteorological Enterprise Assist Middle. 
  1. Submit your e-mail tackle within the account request type. You’ll obtain a short lived acceptance e-mail, then a hyperlink to finish your consumer info. After handbook evaluate, JAXA allows your entry and notifies you as soon as you’ll be able to obtain Himawari Customary Information and geophysical parameter merchandise. 
  1. As soon as accepted, you’ll be able to log in to obtain near-real-time and archived Himawari knowledge through the P-Tree FTP/HTTP providers, following JAXA’s steering on non-redistribution and quotation. 

In observe, a standard sample is to make use of the NOAA/AWS buckets for open, scriptable entry to uncooked imagery, and the JAXA P-Tree merchandise if you want value-added parameters (e.g., cloud or aerosol properties) and are working inside non-profit analysis or instructional tasks. 

# open the downloaded file
!pip set up xarray netCDF4
!pip set up rasterio polars_h3
!pip set up geopandas pykrige
!pip set up polars==1.25.2
!pip set up dask[complete] rioxarray h3==3.7.7
!pip set up h3ronpy==0.21.1
!pip set up geowrangler
# Himawari utilizing – JAXA Himawari Monitor / P-Tree
# create your account right here and use the username and password despatched by e-mail - https://www.eorc.jaxa.jp/ptree/registration_top.html

consumer = '' # enter the username 
password = '' # enter the password 
from ftplib import FTP
from pathlib import Path
import rasterio
from rasterio.rework import from_origin
import xarray as xr
import os
import matplotlib.pyplot as plt


def get_himawari_ftp_past_2_days(consumer, password):

    # FTP connection particulars
    ftp = FTP('ftp.ptree.jaxa.jp')
    ftp.login(consumer=consumer, passwd=password)

    # test the listing content material : /pub/himawari/L2/ARP/031/
    # particulars of AOD directoty right here: https://www.eorc.jaxa.jp/ptree/paperwork/README_HimawariGeo_en.txt

    overall_path= "/pub/himawari/L3/ARP/031/"
    directories = overall_path.strip("/").break up("/")

    for listing in directories:
      ftp.cwd(listing)

    # Checklist information within the goal listing
    date_month_files = ftp.nlst()

    # order information desc
    date_month_files.kind(reverse=False)
    print("Information in goal listing:", date_month_files)

    # get a listing of all of the month / days throughout the "/pub/himawari/L3/ARP/031/" path throughout the previous 2 months
    limited_months_list = date_month_files[-2:]

    i=0
    # for every month within the limited_months_list, record all the times inside in
    for month in limited_months_list:
      ftp.cwd(month)
      date_day_files = ftp.nlst()
      date_day_files.kind(reverse=False)


      # mix every component of the date_day_file record with the month : month +"/" + date_day_file
      list_combined_days_month_inter = [month + "/" + date_day_file for date_day_file in date_day_files]
      if i ==0:
        list_combined_days_month= list_combined_days_month_inter
        i=i+1
      else:
        list_combined_days_month= list_combined_days_month + list_combined_days_month_inter
      ftp.cwd("..")

    # take away all parts containing every day or month-to-month from list_combined_days_month
    list_combined_days_month = [item for item in list_combined_days_month if 'daily' not in item and 'monthly' not in item]

    # get the record of days we wish to obtain : in our case final 2 days - for NRT
    limited_list_combined_days_month=list_combined_days_month[-2:]


    for month_day_date in limited_list_combined_days_month:
      #navigate to the related listing
      ftp.cwd(month_day_date)
      print(f"listing: {month_day_date}")

      # get the record of the hourly information inside every listing
      date_hour_files = ftp.nlst()
      !mkdir -p ./raw_data/{month_day_date}

      #for every hourly file within the record
      for date_hour_file in date_hour_files:
        target_file_path=f"./raw_data/{month_day_date}/{date_hour_file}"
        # Obtain the goal file - provided that it doesn't exist already

        if not os.path.exists(target_file_path):
            with open(target_file_path, "wb") as local_file:
              ftp.retrbinary(f"RETR {date_hour_file}", local_file.write)
              print(f"Downloaded {date_hour_file} efficiently!")
        else:
            print(f"File already exists: {date_hour_file}")



      print("--------------")
      # return 2 steps within the ftp tree
      ftp.cwd("..")
      ftp.cwd("..")
def transform_to_tif():
    # get record of information in raw_data folder
    month_file_list = os.listdir("./raw_data")
    month_file_list

    #order month_file_list
    month_file_list.kind(reverse=False)

    nb_errors=0
    # get record of every day folder for the previous 2 months solely

    for month_file in month_file_list[-2:]:
        print(f"-----------------------------------------")
        print(f"Month thought of: {month_file}")
        date_file_list=os.listdir(f"./raw_data/{month_file}")
        date_file_list.kind(reverse=False)

        # get record of information for every day folder

        for date_file in date_file_list[-2:]:
            print(f"---------------------------")
            print(f"Day thought of: {date_file}")
            hour_file_list=os.listdir(f"./raw_data/{month_file}/{date_file}")
            hour_file_list.kind(reverse=False)

            #course of every hourly file right into a tif file and rework it into an h3 processed dataframe
            for hour_file in hour_file_list:
                file_path = f"./raw_data/{month_file}/{date_file}/{hour_file}"
                hour_file_tif=hour_file.exchange(".nc",".tif")
                output_tif = f"./tif/{month_file}/{date_file}/{hour_file_tif}"
                if os.path.exists(output_tif):
                   print(f"File already exists: {output_tif}")
                else:

                   attempt:
                      dataset = xr.open_dataset(file_path, engine='netcdf4')
                   besides:
                      #go to subsequent hour_file
                      print(f"error opening {hour_file} file - skipping ")
                      nb_errors=nb_errors+1
                      proceed

                   # Entry a selected variable
                   variable_name = record(dataset.data_vars.keys())[1] # Merged AOT product
                   knowledge = dataset[variable_name]

                   # Plot knowledge (if it is 2D and appropriate)
                   plt.determine()
                   knowledge.plot()
                   plt.title(f'{date_file}')
                   plt.present()

                   # Extract metadata (exchange with precise coordinates out of your knowledge if out there)
                   lon = dataset['longitude'] if 'longitude' in dataset.coords else None
                   lat = dataset['latitude'] if 'latitude' in dataset.coords else None

                   # Deal with lacking lat/lon (instance assumes evenly spaced grid)
                   if lon is None or lat is None:
                        lon_start, lon_step = -180, 0.05 # Instance values
                        lat_start, lat_step = 90, -0.05 # Instance values
                        lon = xr.DataArray(lon_start + lon_step * vary(knowledge.form[-1]), dims=['x'])
                        lat = xr.DataArray(lat_start + lat_step * vary(knowledge.form[-2]), dims=['y'])

                   # Outline the affine rework for georeferencing
                   rework = from_origin(lon.min().merchandise(), lat.max().merchandise(), abs(lon[1] - lon[0]).merchandise(), abs(lat[0] - lat[1]).merchandise())

                   # Save to GeoTIFF
                   !mkdir -p ./tif/{month_file}/{date_file}

                   with rasterio.open(
                   output_tif,
                   'w',
                   driver='GTiff',
                   peak=knowledge.form[-2],
                   width=knowledge.form[-1],
                   rely=1, # Variety of bands
                   dtype=knowledge.dtype.identify,
                   crs='EPSG:4326', # Coordinate Reference System (e.g., WGS84)
                   rework=rework
                   ) as dst:

                        dst.write(knowledge.values, 1) # Write the information to band 1
                   print(f"Saved {output_tif} efficiently!")
                   print(f"{nb_errors} error(s) ")
get_himawari_ftp_past_2_days(consumer, password)
transform_to_tif()

9) NASA — FIRMS [Special Highlight] 

NASA’s Hearth Info for Useful resource Administration System (FIRMS) [10] offers near-real-time info on energetic fires and thermal anomalies detected by devices equivalent to MODIS and VIIRS. It affords international protection with low latency (on the order of minutes to hours), supplying attributes equivalent to fireplace radiative energy, confidence, and acquisition time. FIRMS is broadly used for wildfire monitoring, agricultural burning, forest administration, and as a proxy enter for air-quality and smoke dispersion modelling. 
 
Nice for: pinpointing fireplace hotspots that drive AQ spikes, monitoring plume sources and fire-line development, monitoring crop-residue/forest burns, and triggering fast response. Easy accessibility through CSV/GeoJSON/Shapefile, map tiles/API, with 24–72 h rolling feeds and full archives for seasonal evaluation. 

Find out how to register and get API entry 

  1. Create a free NASA Earthdata Login account at: 
    https://urs.earthdata.nasa.gov 
  1. Verify your e-mail and register together with your new credentials. 
  1. Go to the FIRMS website you propose to make use of, for instance: 
  1. Click on Login (high proper) and authenticate together with your Earthdata username and password. As soon as logged in, you’ll be able to: 
  • customise map views and obtain choices from the net interface, and 
  • generate or use FIRMS Internet Companies/API URLs that honour your authenticated session. 
  1. For scripted entry, you’ll be able to name the FIRMS obtain or net service endpoints (e.g., GeoJSON, CSV) utilizing customary HTTP instruments (e.g., curl, requests in Python). If an endpoint requires authentication, provide your Earthdata credentials through a .netrc file or session cookies, as you’d for different Earthdata providers. 

In observe, FIRMS is a handy strategy to pull current fireplace places into an air-quality workflow: you’ll be able to fetch every day or hourly fireplace detections for a area, convert them to a GeoDataFrame, after which intersect with wind fields, inhabitants grids, or sensor networks to know potential smoke impacts. 

#FIRMS  
!pip set up geopandas rtree shapely 
import pandas as pd 
import geopandas as gpd 
from shapely.geometry import Level 
import numpy as np 
import matplotlib.pyplot as plt 
import rtree 

# get boundaries of Thailand 
boundaries_country = gpd.read_file(f'https://github.com/wmgeolab/geoBoundaries/uncooked/fcccfab7523d4d5e55dfc7f63c166df918119fd1/releaseData/gbOpen/THA/ADM0/geoBoundaries-THA-ADM0.geojson') 
boundaries_country.plot() 

# Actual time knowledge supply: https://corporations.modaps.eosdis.nasa.gov/active_fire/ 
# Previous 7 days hyperlinks: 
modis_7d_url= "https://corporations.modaps.eosdis.nasa.gov/knowledge/active_fire/modis-c6.1/csv/MODIS_C6_1_SouthEast_Asia_7d.csv" 
suomi_7d_url= "https://corporations.modaps.eosdis.nasa.gov/knowledge/active_fire/suomi-npp-viirs-c2/csv/SUOMI_VIIRS_C2_SouthEast_Asia_7d.csv" 
j1_7d_url= "https://corporations.modaps.eosdis.nasa.gov/knowledge/active_fire/noaa-20-viirs-c2/csv/J1_VIIRS_C2_SouthEast_Asia_7d.csv" 
j2_7d_url="https://corporations.modaps.eosdis.nasa.gov/knowledge/active_fire/noaa-21-viirs-c2/csv/J2_VIIRS_C2_SouthEast_Asia_7d.csv" 
urls = [modis_7d_url, suomi_7d_url, j1_7d_url, j2_7d_url] 

# Create an empty GeoDataFrame to retailer the mixed knowledge 
gdf = gpd.GeoDataFrame() 

for url in urls: 
    df = pd.read_csv(url) 

    # Create a geometry column from latitude and longitude 
    geometry = [Point(xy) for xy in zip(df['longitude'], df['latitude'])] 
    gdf_temp = gpd.GeoDataFrame(df, crs="EPSG:4326", geometry=geometry)
     
    # Concatenate the short-term GeoDataFrame to the primary GeoDataFrame 
    gdf = pd.concat([gdf, gdf_temp], ignore_index=True) 

# Filter to maintain solely fires throughout the nation boundaries 
gdf = gpd.sjoin(gdf, boundaries_country, how="inside", predicate="inside") 

# Show fires on map  
frp = gdf["frp"].astype(float) 
fig, ax = plt.subplots(figsize=(9,9)) 
boundaries_country.plot(ax=ax, facecolor="none", edgecolor="0.3", linewidth=0.8) 
gdf.plot(ax=ax, markersize=frp, shade="crimson", alpha=0.55) 
ax.set_title("Fires inside nation boundaries (bubble dimension = Hearth Radiative Energy )") 
ax.set_axis_off() 
plt.present() 

Information sorts you will meet (and tips on how to learn them proper) 

Air-quality work hardly ever lives in a single, tidy CSV. So, it helps to know what the file sorts you’ll meet. You’ll transfer between multidimensional mannequin outputs (NetCDF/GRIB/Zarr), satellite tv for pc rasters (COG/GeoTIFF), level measurements (CSV /Parquet /GeoParquet), and web-friendly codecs (JSON/GeoJSON), typically in the identical pocket book. 

This part is a fast subject information to these codecs and tips on how to open them with out getting caught. 

There isn’t any have to memorise any of this, so be at liberty to skim the record as soon as, then come again if you hit an unfamiliar file extension within the wild. 

  1. NetCDF4 / HDF5 (self-describing scientific arrays): Extensively used for reanalyses, satellite tv for pc merchandise, and fashions. Wealthy metadata, multi-dimensional (time, degree, lat, lon) Typical extensions: .nc, .nc4, .h5, .hdf5 

Learn: 

# pip set up xarray netCDF4 

import xarray as xr 
ds = xr.open_dataset("modis_aod_2025.nc") 
ds = ds.sel(time=slice("2025-08-01","2025-08-07")) 
print(ds) 
  1. Cloud-Optimised GeoTIFF (COG): Raster format tuned for HTTP vary requests (stream simply what you want). Widespread for satellite tv for pc imagery and gridded merchandise. Typical extensions: .tif, .tiff 

 Learn: 

# pip set up rasterio 

import rasterio 
with rasterio.open("https://example-bucket/no2_mean_2025.tif") as src: 
    window = rasterio.home windows.from_bounds(*(-0.3,5.5,0.3,5.9), src.rework) 
    arr = src.learn(1, window=window)
  1. JSON (nested) & GeoJSON (options + geometry): Nice for APIs and light-weight geospatial. GeoJSON makes use of WGS84 (EPSG:4326) by default. Typical extensions: json, .jsonl, .ndjson, .geojsonl, .ndgeojson 

Learn: 

# pip set up geopandas 

import geopandas as gpd 
gdf = gpd.read_file("factors.geojson")  # columns + geometry 
gdf = gdf.set_crs(4326)                # guarantee WGS84 
  1. GRIB2 (meteorology, mannequin outputs): Compact, tiled; typically utilized by CAMS/ECMWF/NWP. Typical extensions: .grib2, .grb2, .grib, .grb. In observe, knowledge suppliers typically add compression suffixes too, e.g. .grib2.gz or .grb2.bz2. 

Learn: 

# pip set up xarray cfgrib 

import xarray as xr 
ds = xr.open_dataset("cams_ozone.grib", engine="cfgrib") 
  1. Parquet & GeoParquet (columnar, compressed): Greatest for large tables: quick column choice, predicate pushdown, partitioning (e.g., by date/metropolis). GeoParquet provides an ordinary for geometries. Typical extensions: .parquet, .parquet.gz 

 Learn/Write: 

# pip set up pandas pyarrow geopandas geoparquet 

import pandas as pd, geopandas as gpd 
df = pd.read_parquet("openaq_accra_2025.parquet")   # columns solely 

# Convert a GeoDataFrame -> GeoParquet 
gdf = gpd.read_file("factors.geojson") 
gdf.to_parquet("factors.geoparquet")  # preserves geometry & CRS 
  1. CSV/TSV (textual content tables): Easy, common. Weak at giant scale (gradual I/O, no schema), no geometry. Typical extensions: .csv, .tsv (additionally generally .tab, much less widespread)

 Learn: 

# pip set up pandas 

import pandas as pd
df = pd.read_csv("measurements.csv", parse_dates=["datetime"], dtype={"site_id":"string"}) 
  1. Zarr (chunked, cloud-native): Supreme for evaluation within the cloud with parallel reads (works nice with Dask). Typical extension: .zarr (typically a listing / retailer ending in .zarr; sometimes packaged as .zarr.zip) 

 Learn: 

# pip set up xarray zarr s3fs 

import xarray as xr
ds = xr.open_zarr("s3://bucket/cams_eac4_2025.zarr", consolidated=True) 

Observe: Shapefile (legacy vector): Works, however brittle (many information, 10-char subject restrict). . It is a legacy codecs and it’s higher to make use of the alternate options like GeoPackage or GeoParquet  

It is very important select the correct geospatial (or scientific) file format as it’s not only a storage determination nevertheless it immediately impacts how rapidly you’ll be able to learn knowledge, software compatibility, how simply you’ll be able to share it, and the way properly it scales from a desktop workflow to cloud-native processing. The next desk (Desk 1) offers a sensible “format-to-task” cheat sheet: for every widespread want (from fast API dumps to cloud-scale arrays and net mapping), it lists essentially the most appropriate format, the extensions you’ll usually encounter, and the core motive that format is an efficient match. It may be used as a default place to begin when designing pipelines, publishing datasets, or choosing what to obtain from an exterior repository. 

Want Greatest Wager Typical Extension Why 
Human-readable logs or fast API dumps  CSV/JSON  .csv, .json (additionally .jsonl, .ndjson)  Ubiquitous, straightforward to examine 
Large tables (tens of millions of rows)  Parquet/ GeoParquet  .parquet  Quick scans, column pruning, and partitioning 
Giant rasters  over HTTP  COG  .tif, .tiff  Vary requests; no full obtain 
Multi-dimensional scientific knowledge  NetCDF4/HDF5  .nc, .nc4, .h5, .hdf5  Self-describing, items/attrs 
Meteorological mannequin outputs  GRIB2  .grib2, .grb2, .grib, .grb  Compact, broadly supported by wx instruments 
Cloud-scale arrays  Zarr  .zarr  Chunked + parallel; cloud-native 
Exchangeable vector file  GeoPackage  .gpkg  Single file; strong 
Internet mapping geometries  GeoJSON  .geojsonl, 
.ndgeojson 
Easy; native to net stacks 
Desk 1: Choosing the right format for the job 

Tip: An attention-grabbing discuss on STAC and knowledge sorts (particularly GeoParquet): https://github.com/GSA/gtcop-wiki/wiki/June-2025:-GeoParquet,-Iceberg-and-CloudpercentE2percent80percent90Native-Spatial-Information-Infrastructures

A number of open STAC catalogues are actually out there, together with public endpoints for optical, radar, and atmospheric merchandise (for instance, Landsat and Sentinel imagery through suppliers equivalent to Factor 84’s Earth Search or Microsoft’s Planetary Pc). STAC makes it a lot simpler to script “discover and obtain all scenes for this polygon and time vary” and to combine totally different datasets into the identical workflow. 

Conclusion — from “the place” the information lives to “how” you utilize it 

Determine 3: Creating publicity maps from hotspots  © UNICEF/UNI724381/Kongchan Phi. All rights reserved. 

Air for Tomorrow: We began with the query “What are these children respiration at present?” This publish offers a sensible path and instruments that can assist you reply this query. You now know the place open-air high quality knowledge resides, together with regulatory networks, neighborhood sensors, satellite tv for pc measurements, and reanalysis. You additionally perceive what these information are (GeoJSON, Parquet/GeoParquet, NetCDF/HDF5, COG, GRIB, Zarr) and tips on how to retrieve them with compact, reproducible snippets. The aim is past simply downloading them; it’s to make defensible, quick, and shareable analyses that maintain up tomorrow

You may assemble a reputable native image in hours, not weeks. From fireplace hotspots (Determine 2) to school-route publicity (Determine 1), you’ll be able to create publicity maps (Determine 3).

Up subsequent: We might showcase an precise Air High quality Mannequin developed by us on the UNICEF Nation Workplace of Lao PDR with the UNICEF EAPRO’s Frontier Information Workforce. We might undergo an open, end-to-end mannequin pipeline. When there are ground-level air high quality knowledge streams out there, we’d cowl how function engineering, bias correction, normalisation, and a mannequin may be developed with an actionable floor {that a} regional can use tomorrow morning. 

Contributors: Prithviraj Pramanik, AQAI; Hugo Ruiz Verastegui, Anthony Mockler, Judith Hanan, Frontier Information Lab; Risdianto Irawan, UNICEF EAPRO; Soheib Abdalla, Andrew Dunbrack, UNICEF Lao PDR Nation Workplace; Halim Jun, Daniel Alvarez, Shane O’Connor, UNICEF Workplace of Innovation; 

Use Instances, Structure & Shopping for Suggestions


Introduction – What Makes Nvidia GH200 the Star of 2026?

Fast Abstract: What’s the Nvidia GH200 and why does it matter in 2026? – The Nvidia GH200 is a hybrid superchip that merges a 72‑core Arm CPU (Grace) with a Hopper/H200 GPU utilizing NVLink‑C2C. This integration creates as much as 624 GB of unified reminiscence accessible to each CPU and GPU, enabling reminiscence‑sure AI workloads like lengthy‑context LLMs, retrieval‑augmented era (RAG) and exascale simulations. In 2026, as fashions develop bigger and extra advanced, the GH200’s reminiscence‑centric design delivers efficiency and value effectivity not achievable with conventional GPU playing cards. Clarifai presents enterprise‑grade GH200 internet hosting with good autoscaling and cross‑cloud orchestration, making this know-how accessible for builders and companies.

Synthetic intelligence is evolving at breakneck pace. Mannequin sizes are growing from hundreds of thousands to trillions of parameters, and generative functions equivalent to retrieval‑augmented chatbots and video synthesis require large key–worth caches and embeddings. Conventional GPUs just like the A100 or H100 present excessive compute throughput however can turn into bottlenecked by reminiscence capability and information motion. Enter the Nvidia GH200, usually nicknamed the Grace Hopper superchip. As an alternative of connecting a CPU and GPU by way of a gradual PCIe bus, the GH200 fuses them on the identical bundle and hyperlinks them by NVLink‑C2C—a excessive‑bandwidth, low‑latency interconnect that delivers 900 GB/s of bidirectional bandwidth. This structure permits the GPU to entry the CPU’s reminiscence straight, leading to a unified reminiscence pool of as much as 624 GB (when combining the 96 GB or 144 GB HBM on the GPU with 480 GB LPDDR5X on the CPU).

This information presents an in depth have a look at the GH200: its structure, efficiency, best use instances, deployment fashions, comparability to different GPUs (H100, H200, B200), and sensible steerage on when and the way to decide on it. Alongside the way in which we are going to spotlight Clarifai’s compute options that leverage GH200 and supply greatest practices for deploying reminiscence‑intensive AI workloads.

Fast Digest: How This Information Is Structured

  • Understanding the GH200 Structure – We look at how the hybrid CPU–GPU design and unified reminiscence system work, and why HBM3e issues.
  • Benchmarks & Price Effectivity – See how GH200 performs in inference and coaching in contrast with H100/H200, and the impact on value per token.
  • Use Instances & Workload Match – Be taught which AI and HPC workloads profit from the superchip, together with RAG, LLMs, graph neural networks and exascale simulations.
  • Deployment Fashions & Ecosystem – Discover on‑premises DGX techniques, hyperscale cloud situations, specialist GPU clouds, and Clarifai’s orchestration options.
  • Choice Framework – Perceive when to decide on GH200 vs H100/H200 vs B200/Rubin based mostly on reminiscence, bandwidth, software program and funds.
  • Challenges & Future Traits – Contemplate limitations (ARM software program, energy, latency) and sit up for HBM3e, Blackwell, Rubin and new supercomputers.

Let’s dive in.


GH200 Structure and Reminiscence Improvements

Fast Abstract: How does the GH200’s structure differ from conventional GPUs? – In contrast to standalone GPU playing cards, the GH200 integrates a 72‑core Grace CPU and a Hopper/H200 GPU on a single module. The 2 chips talk by way of NVLink‑C2C delivering 900 GB/s bandwidth. The GPU consists of 96 GB HBM3 or 144 GB HBM3e, whereas the CPU gives 480 GB LPDDR5X. NVLink‑C2C permits the GPU to straight entry CPU reminiscence, making a unified reminiscence pool of as much as 624 GB. This eliminates expensive information transfers and is vital to the GH200’s reminiscence‑centric design.

Hybrid CPU–GPU Fusion

At its core, the GH200 combines a Grace CPU and a Hopper GPU. The CPU options 72 Arm Neoverse V2 cores (or 72 Grace cores), delivering excessive reminiscence bandwidth and power effectivity. The GPU relies on the Hopper structure (used within the H100) however could also be upgraded to the H200 in newer revisions, including quicker HBM3e reminiscence. NVLink‑C2C is the key sauce: a cache‑coherent interface enabling each chips to share reminiscence coherently at 900 GB/s – roughly 7× quicker than PCIe Gen5. This design makes the GH200 successfully an enormous APU or system‑on‑chip tailor-made for AI.

Unified Reminiscence Pool

Conventional GPU servers depend on discrete reminiscence swimming pools: CPU DRAM and GPU HBM. Information have to be copied throughout the PCIe bus, incurring latency and overhead. The GH200’s unified reminiscence eliminates this barrier. The Grace CPU brings 480 GB of LPDDR5X reminiscence with bandwidth of 546 GB/s, whereas the Hopper GPU consists of 96 GB HBM3 delivering 4 000 GB/s bandwidth. The upcoming HBM3e variant will increase reminiscence capability to 141–144 GB and boosts bandwidth by over 25 %. Mixed with NVLink‑C2C, this gives a shared reminiscence pool of as much as 624 GB, enabling the GPU to cache large datasets and key–worth caches for LLMs with out repeatedly fetching from CPU reminiscence. NVLink can also be scalable: NVL2 pairs two superchips to create a node with 288 GB HBM and 10 TB/s bandwidth, and the NVLink change system can join 256 superchips to behave as one large GPU with 1 exaflop efficiency and 144 TB unified reminiscence.

HBM3e and Rubin Platform

The GH200 began with HBM3 however is already evolving. The HBM3e revision provides 144 GB of HBM for the GPU, elevating efficient reminiscence capability by round 50 % and growing bandwidth from 4 000 GB/s to about 4.9 TB/s. This improve helps massive fashions retailer extra key–worth pairs and embeddings solely in on‑chip reminiscence. Trying forward, Nvidia’s Rubin platform (introduced 2025) will introduce a brand new CPU with 88 Olympus cores, 1.8 TB/s NVLink‑C2C bandwidth and 1.5 TB LPDDR5X reminiscence, doubling reminiscence capability over Grace. Rubin may even help NVLink 6 and NVL72 rack techniques that scale back inference token value by 10× and coaching GPU rely by in contrast with Blackwell—an indication that reminiscence‑centric design will proceed to evolve.

Skilled Insights

  • Unified reminiscence is a paradigm shift – By exposing GPU reminiscence as a CPU NUMA node, NVLink‑C2C eliminates the necessity for specific information copying and permits CPU code to entry HBM straight. This simplifies programming and accelerates reminiscence‑sure duties.
  • HBM3e vs HBM3 – The 50 % enhance in capability and 25 % enhance in bandwidth of HBM3e considerably extends the dimensions of fashions that may be served on a single chip, pushing the GH200 into territory beforehand reserved for multi‑GPU clusters.
  • Scalability by way of NVLink change – Connecting a whole lot of superchips by way of NVLink change ends in a single logical GPU with terabytes of shared reminiscence—essential for exascale techniques like Helios and JUPITER.
  • Grace vs Rubin – Whereas Grace presents 72 cores and 480 GB reminiscence, Rubin will ship 88 cores and as much as 1.5 TB reminiscence with NVLink 6, hinting that future AI workloads might require much more reminiscence and bandwidth.

Efficiency Benchmarks & Price Effectivity

Fast Abstract: How does GH200 carry out relative to H100/H200, and what does this imply for value? – Benchmarks reveal that the GH200 delivers 1.4×–1.8× greater MLPerf inference efficiency per accelerator than the H100. In sensible checks on Llama 3 fashions, GH200 achieved 7.6× greater throughput and lowered value per token by 8× in contrast with H100. Clarifai reviews a 17 % efficiency acquire over H100 of their MLPerf outcomes. These features stem from unified reminiscence and NVLink‑C2C, which scale back latency and allow bigger batches.

MLPerf and Vendor Benchmarks

In Nvidia’s MLPerf Inference v4.1 outcomes, the GH200 delivered as much as 1.4× extra efficiency per accelerator than the H100 on generative AI duties. When configured in NVL2, two superchips achieved 3.5× extra reminiscence and 3× extra bandwidth than a single H100, translating into higher scaling for big fashions. Clarifai’s inner benchmarking confirmed a 17 % throughput enchancment over H100 for MLPerf duties.

Actual‑World Inference (LLM and RAG)

In a extensively shared weblog put up, Lambda AI in contrast GH200 to H100 for single‑node Llama 3.1 70B inference. GH200 delivered 7.6× greater throughput and 8× decrease value per token than H100, due to the flexibility to dump key–worth caches to CPU reminiscence. Baseten ran related experiments with Llama 3.3 70B and located that GH200 outperformed H100 by 32 % as a result of the reminiscence pool allowed bigger batch sizes. Nvidia’s technical weblog on RAG functions confirmed that GH200 gives 2.7×–5.7× speedups in contrast with A100 throughout embedding era, index construct, vector search and LLM inference.

Price‑Per‑Hour & Cloud Pricing

Price is a vital issue. An evaluation of GPU rental markets discovered that GH200 situations value $4–$6 per hour on hyperscalers, barely greater than H100 however with improved efficiency, whereas specialist GPU clouds typically supply GH200 at aggressive charges. Decentralised marketplaces might permit cheaper entry however usually restrict options. Clarifai’s compute platform makes use of good autoscaling and GPU fractioning to optimise useful resource utilisation, lowering value per token additional.

Reminiscence‑Certain vs Compute‑Certain Workloads

Whereas GH200 shines for reminiscence‑sure duties, it doesn’t at all times beat H100 for compute‑sure kernels. Some compute‑intensive kernels saturate the GPU’s compute items and aren’t restricted by reminiscence bandwidth, so the efficiency benefit shrinks. Fluence’s information notes that GH200 shouldn’t be the fitting selection for easy single‑GPU coaching or compute‑solely duties. In such instances, H100 or H200 may ship related or higher efficiency at decrease value.

Skilled Insights

  • Price per token issues – Inference value isn’t nearly GPU worth; it’s about throughput. GH200’s means to make use of bigger batches and retailer key–worth caches on CPU reminiscence drastically cuts value per token.
  • Batch measurement is the important thing – Bigger unified reminiscence permits greater batches and reduces the overhead of reloading contexts, resulting in large throughput features.
  • Stability compute and reminiscence – For compute‑heavy duties like CNN coaching or matrix multiplications, H100 or H200 might suffice. GH200 is focused at reminiscence‑sure workloads, so select accordingly.

Use Instances and Workload Match

Fast Abstract: Which workloads profit most from GH200? – GH200 excels in massive language mannequin inference and coaching, retrieval‑augmented era (RAG), multimodal AI, vector search, graph neural networks, advanced simulations, video era, and scientific HPC. Its unified reminiscence permits storing massive key–worth caches and embeddings in RAM, enabling quicker response occasions and bigger context home windows. Exascale supercomputers like JUPITER make use of tens of 1000’s of GH200 chips to simulate local weather and physics at unprecedented scale.

Massive Language Fashions and Chatbots

Fashionable LLMs equivalent to Llama 3, Llama 2, GPT‑J and different 70 B+ parameter fashions require storing gigabytes of weights and key–worth caches. GH200’s unified reminiscence helps as much as 624 GB of accessible reminiscence, which means that lengthy context home windows (128 okay tokens or extra) will be served with out swapping to disk. Nvidia’s weblog on multiturn interactions exhibits that offloading KV caches to CPU reminiscence reduces time‑to‑first token by as much as 14× and improves throughput in contrast with x86‑H100 servers. This makes GH200 best for chatbots requiring actual‑time responses and deep context.

Retrieval‑Augmented Era (RAG)

RAG pipelines combine massive language fashions with vector databases to fetch related data. This requires producing embeddings, constructing vector indices and performing similarity search. Nvidia’s RAG benchmark exhibits GH200 achieves 2.7× quicker embedding era, 2.9× quicker index construct, 3.3× quicker vector search, and 5.7× quicker LLM inference in comparison with A100. The power to maintain vector databases in unified reminiscence reduces information motion and improves latency. Clarifai’s RAG APIs can run on GH200 to deploy chatbots with area‑particular information and summarisation capabilities.

Multimodal AI and Video Era

The GH200’s reminiscence capability additionally advantages multimodal fashions (textual content + picture + video). Fashions like VideoPoet or diffusion‑based mostly video synthesizers require storing frames and cross‑modal embeddings. GH200’s reminiscence can maintain longer sequences and unify CPU and GPU reminiscence, accelerating coaching and inference. That is particularly precious for firms engaged on video era or massive‑scale picture captioning.

Graph Neural Networks and Suggestion Techniques

Massive recommender techniques and graph neural networks deal with billions of nodes and edges, usually requiring terabytes of reminiscence. Nvidia’s press launch on the DGX GH200 emphasises that NVLink change mixed with a number of superchips allows 144 TB of shared reminiscence for coaching advice techniques. This reminiscence capability is essential for fashions like Deep Studying Suggestion Mannequin 3 (DLRM‑v3) or GNNs utilized in social networks and information graphs. GH200 can drastically scale back coaching time and enhance scaling.

Scientific HPC and Exascale Simulations

Outdoors AI, the GH200 performs a job in scientific HPC. The European JUPITER supercomputer, anticipated to exceed 90 exaflops, employs 24 000 GH200 superchips interconnected by way of InfiniBand, with every node utilizing 288 Arm cores and 896 GB of reminiscence. The excessive reminiscence and compute density speed up local weather fashions, physics simulations and drug discovery. Equally, the Helios and DGX GH200 techniques join a whole lot of superchips by way of NVLink switches to type unified supernodes with exascale efficiency.

Skilled Insights

  • RAG is reminiscence‑sure – RAG workloads usually fail on smaller GPUs attributable to restricted reminiscence for embeddings and indices; GH200 solves this by providing unified reminiscence and close to‑zero copy entry.
  • Video era wants massive temporal context – GH200’s reminiscence allows storing a number of frames and have maps for prime‑decision video synthesis, lowering I/O overhead.
  • Graph workloads thrive on reminiscence bandwidth – Analysis on GNN coaching exhibits GH200 gives 4×–7× speedups for graph neural networks in contrast with conventional GPUs, due to its reminiscence capability and NVLink community.

Deployment Choices and Ecosystem

Fast Abstract: The place are you able to entry GH200 in the present day? – GH200 is on the market by way of on‑premises DGX techniques, cloud suppliers like AWS, Azure and Google Cloud, specialist GPU clouds (Lambda, Baseten, Fluence) and decentralised marketplaces. Clarifai presents enterprise‑grade GH200 internet hosting with options like good autoscaling, GPU fractioning and cross‑cloud orchestration. NVLink change techniques permit a number of superchips to behave as a single GPU with large shared reminiscence.

On‑Premise DGX Techniques

Nvidia’s DGX GH200 makes use of NVLink change to attach as much as 256 superchips, delivering 1 exaflop of efficiency and 144 TB unified reminiscence. Organisations like Google, Meta and Microsoft had been early adopters and plan to make use of DGX GH200 techniques for big mannequin coaching and AI analysis. For enterprises with strict information‑sovereignty necessities, DGX bins supply most management and excessive‑pace NVLink interconnects.

Hyperscaler Cases

Main cloud suppliers now supply GH200 situations. On AWS, Azure and Google Cloud, you’ll be able to lease GH200 nodes at roughly $4–$6 per hour. Pricing varies relying on area and configuration; the unified reminiscence reduces the necessity for multi‑GPU clusters, doubtlessly reducing general prices. Cloud situations are sometimes obtainable in restricted areas attributable to provide constraints, so early reservation is advisable.

Specialist GPU Clouds and Decentralised Markets

Firms like Lambda Cloud, Baseten and Fluence present GH200 rental or hosted inference. Fluence’s information compares pricing throughout suppliers and notes that specialist clouds might supply extra aggressive pricing and higher software program help than hyperscalers. Baseten’s experiments present the best way to run Llama 3 on GH200 for inference with 32 % higher throughput than H100. Decentralised GPU marketplaces equivalent to Golem or GPUX permit customers to lease GH200 capability from people or small information centres, though options like NVLink pairing could also be restricted.

Clarifai Compute Platform

Clarifai stands out by providing enterprise‑grade GH200 internet hosting with strong orchestration instruments. Key options embody:

  • Good autoscaling: robotically scales GH200 assets based mostly on mannequin demand, making certain low latency whereas optimising value.
  • GPU fractioning: splits a GH200 into smaller logical partitions, permitting a number of workloads to share the reminiscence pool and compute items effectively.
  • Cross‑cloud flexibility: run workloads on GH200 {hardware} throughout a number of clouds or on‑premises, simplifying migration and failover.
  • Unified management & governance: handle all deployments by Clarifai’s console or API, with monitoring, logging and compliance inbuilt.

These capabilities let enterprises undertake GH200 with out investing in bodily infrastructure and guarantee they solely pay for what they use.

Skilled Insights

  • NVLink change vs InfiniBand – NVLink change presents decrease latency and better bandwidth than InfiniBand, enabling a number of GH200 modules to behave like a single GPU.
  • Cloud availability is restricted – Attributable to excessive demand and restricted provide, GH200 situations could also be scarce on public cloud; working with specialist suppliers or Clarifai ensures precedence entry.
  • Compute orchestration simplifies adoption – Utilizing Clarifai’s orchestration options permits engineers to concentrate on fashions moderately than infrastructure, bettering time‑to‑market.

Choice Information: GH200 vs H100/H200 vs B200/Rubin

Fast Abstract: How do you determine which GPU to make use of? – The selection relies on reminiscence necessities, bandwidth, software program help, energy funds and value. GH200 presents unified reminiscence (96–144 GB HBM + 480 GB LPDDR) and excessive bandwidth (900 GB/s NVLink‑C2C), making it best for reminiscence‑sure duties. H100 and H200 are higher for compute‑sure workloads or when utilizing x86 software program stacks. B200 (Blackwell) and upcoming Rubin promise much more reminiscence and value effectivity, however availability might lag. Clarifai’s orchestration can combine and match {hardware} to satisfy workload wants.

Reminiscence Capability & Bandwidth

  • H100 – 80 GB HBM and a pair of TB/s reminiscence bandwidth (HBM3). Reminiscence is native to the GPU; information have to be moved from CPU by way of PCIe.
  • H200 – 141 GB HBM3e and 4.8 TB/s bandwidth. A drop‑in alternative for H100 however nonetheless requires PCIe or NVLink bridging. Appropriate for compute‑sure duties needing extra GPU reminiscence.
  • GH200 – 96 GB HBM3 or 144 GB HBM3e plus 480 GB LPDDR5X accessible by way of 900 GB/s NVLink‑C2C, yielding a unified 624 GB pool.
  • B200 (Blackwell) – Rumoured to supply 208 GB HBM3e and 10 TB/s bandwidth; lacks unified CPU reminiscence, so nonetheless reliant on PCIe or NVLink connections.
  • Rubin platform – Will function an 88‑core CPU with 1.5 TB of LPDDR5X and 1.8 TB/s NVLink‑C2C bandwidth. NVL72 racks will drastically scale back inference value.

Software program Stack & Structure

  • GH200 makes use of an ARM structure (Grace CPU). Many AI frameworks help ARM, however some Python libraries and CUDA variations might require recompilation. Clarifai’s native runner solves this by offering containerised environments with the fitting dependencies.
  • H100/H200 run on x86 servers and profit from mature software program ecosystems. In case your codebase closely relies on x86‑particular libraries, migrating to GH200 might require further effort.

Energy Consumption & Cooling

GH200 techniques can draw as much as 1 000 W per node because of the mixed CPU and GPU. Guarantee satisfactory cooling and energy infrastructure. H100 and H200 nodes sometimes eat much less energy individually however might require extra nodes to match GH200’s reminiscence capability.

Price & Availability

GH200 {hardware} is costlier than H100/H200 upfront, however the lowered variety of nodes required for reminiscence‑intensive workloads can offset value. Pricing information suggests GH200 leases value about $4–$6 per hour. H100/H200 could also be cheaper per hour however want extra items to host the identical mannequin. Blackwell and Rubin are usually not but extensively obtainable; early adopters might pay premium pricing.

Choice Matrix

  • Select GH200 when your workloads are reminiscence‑sure (LLM inference, RAG, GNNs, large embeddings) or require unified reminiscence for environment friendly pipelines.
  • Select H100/H200 for compute‑sure duties like convolutional neural networks, transformer pretraining, or when utilizing x86‑dependent software program. H200 provides extra HBM however nonetheless lacks unified CPU reminiscence.
  • Anticipate B200/Rubin for those who want even bigger reminiscence or higher value effectivity and might deal with delayed availability. Rubin’s NVL72 racks could also be revolutionary for exascale AI.
  • Leverage Clarifai to combine {hardware} varieties inside a single pipeline, utilizing GH200 for reminiscence‑heavy levels and H100/B200 for compute‑heavy phases.

Skilled Insights

  • Unified reminiscence modifications the calculus – Contemplate reminiscence capability first; the unified 624 GB on GH200 can exchange a number of H100 playing cards and simplify scaling.
  • ARM software program is maturing – Instruments like PyTorch and TensorFlow have improved help for ARM; containerised environments (e.g., Clarifai native runner) make deployment manageable.
  • HBM3e is a powerful bridge – H200’s HBM3e reminiscence gives a few of GH200’s capability advantages with out new CPU structure, providing an easier improve path.

Challenges, Limitations and Mitigation

Fast Abstract: What are the pitfalls of adopting GH200 and how will you mitigate them? – Key challenges embody software program compatibility on ARM, excessive energy consumption, cross‑die latency, provide chain constraints and greater value. Mitigation methods contain utilizing containerised environments (Clarifai native runner), proper‑sizing assets (GPU fractioning), and planning for provide constraints.

Software program Ecosystem on ARM

The Grace CPU makes use of an ARM structure, which can require recompiling libraries or dependencies. PyTorch, TensorFlow and CUDA help ARM, however some Python packages depend on x86 binaries. Lambda’s weblog warns that PyTorch have to be compiled for ARM, and there could also be restricted prebuilt wheels. Clarifai’s native runner addresses this by packaging dependencies and offering pre‑configured containers, making it simpler to deploy fashions on GH200.

Energy and Cooling Necessities

A GH200 superchip can eat as much as 900 W for the GPU and 1000 W for the total system. Information centres should guarantee satisfactory cooling, energy supply and monitoring. Utilizing good autoscaling to spin down unused nodes reduces power utilization. Contemplate the environmental affect and potential regulatory necessities (e.g., carbon reporting).

Latency & NUMA Results

Whereas NVLink‑C2C presents excessive bandwidth, cross‑die reminiscence entry has greater latency than native HBM. Chips and Cheese’s evaluation notes that the common latency will increase when accessing CPU reminiscence vs HBM. Builders ought to design algorithms to prioritise information locality: hold often accessed tensors in HBM and use CPU reminiscence for KV caches and often accessed information. Analysis is ongoing to optimise information placement and scheduling. explores LLVM OpenMP offload optimisations on GH200, offering insights for HPC workloads.

Provide Chain & Pricing

Excessive demand and restricted provide imply GH200 situations will be scarce. Fluence’s pricing comparability highlights that GH200 might value greater than H100 per hour however presents higher efficiency for reminiscence‑heavy duties. To mitigate provide points, work with suppliers like Clarifai that reserve capability or use decentrised markets to dump non‑vital workloads.

Skilled Insights

  • Embrace hybrid structure – Use each H100/H200 and GH200 the place applicable; unify them by way of container orchestration to beat provide and software program limitations.
  • Optimise information placement – Maintain compute‑intensive kernels on HBM; offload caches to LPDDR reminiscence. Monitor reminiscence bandwidth and latency utilizing profiling instruments.
  • Plan for lengthy lead occasions – Pre‑order GH200 {hardware} or cloud reservations. Develop software program in moveable frameworks to ease transitions between architectures.

Rising Traits & Future Outlook

Fast Abstract: What’s subsequent for reminiscence‑centric AI {hardware}? – Traits embody HBM3e reminiscence, Blackwell (B200/GB200) GPUs, Rubin CPU platforms, NVLink‑6 and NVL72 racks, and the rise of exascale supercomputers. These improvements intention to additional scale back inference value and power consumption whereas growing reminiscence capability and compute density.

HBM3e and Blackwell

The HBM3e revision of GH200 already will increase reminiscence capability to 144 GB and bandwidth to 4.9 TB/s. Nvidia’s subsequent GPU structure, Blackwell, options the B200 and server configurations like GB200 and GB300. These chips will enhance HBM capability to round 208 GB, present improved compute throughput and should incorporate the Hopper or Rubin CPU for unified reminiscence. In accordance with Medium analyst Adrian Cockcroft, GH200 pairs an H200 GPU with the Grace CPU and might join 256 modules utilizing shared reminiscence for improved efficiency.

Rubin Platform and NVLink‑6

Nvidia’s Rubin platform pushes reminiscence‑centric design additional by introducing an 88‑core CPU with 1.5 TB LPDDR5X and 1.8 TB/s NVLink‑C2C bandwidth. Rubin’s NVL72 rack techniques will scale back inference value by 10× and the variety of GPUs wanted for coaching by in contrast with Blackwell. We will anticipate mainstream adoption round 2026–2027, though early entry could also be restricted to massive cloud suppliers.

Exascale Supercomputers & International AI Infrastructure

Supercomputers like JUPITER and Helios display the potential of GH200 at scale. JUPITER makes use of 24 000 GH200 superchips and is anticipated to ship greater than 90 exaflops. These techniques will energy analysis into local weather change, climate prediction, quantum physics and AI. As generative AI functions equivalent to video era and protein folding require extra reminiscence, these exascale infrastructures might be essential.

Business Collaboration and Ecosystem

Nvidia’s press releases emphasise that main tech firms (Google, Meta, Microsoft) and integrators like SoftBank are investing closely in GH200 techniques. In the meantime, storage and networking distributors are adapting their merchandise to deal with unified reminiscence and excessive‑throughput information streams. The ecosystem will proceed to develop, bringing higher software program instruments, reminiscence‑conscious schedulers and cross‑vendor interoperability.

Skilled Insights

  • Reminiscence is the brand new frontier – Future platforms will emphasise reminiscence capability and bandwidth over uncooked flops; algorithms might be redesigned to use unified reminiscence.
  • Rubin and NVLink 6 – These will probably allow multi‑rack clusters with unified reminiscence measured in petabytes, remodeling AI infrastructure.
  • Put together now – Constructing pipelines that may run on GH200 units you as much as undertake B200/Rubin with minimal modifications.

Clarifai Product Integration & Finest Practices

Fast Abstract: How does Clarifai leverage GH200 and what are greatest practices for customers? – Clarifai presents enterprise‑grade GH200 internet hosting with options equivalent to good autoscaling, GPU fractioning, cross‑cloud orchestration, and a native runner for ARM‑optimised deployment. To maximise efficiency, use bigger batch sizes, retailer key–worth caches on CPU reminiscence, and combine vector databases with Clarifai’s RAG APIs.

Clarifai’s GH200 Internet hosting

Clarifai’s compute platform makes the GH200 accessible with no need to buy {hardware}. It abstracts complexity by options:

  • Good autoscaling provisions GH200 situations as demand will increase and scales them down throughout idle intervals.
  • GPU fractioning lets a number of jobs share a single GH200, splitting reminiscence and compute assets to maximise utilisation.
  • Cross‑cloud orchestration permits workloads to run on GH200 throughout varied clouds and on‑premises infrastructure with unified monitoring and governance.
  • Unified management & governance gives centralised dashboards, auditing and function‑based mostly entry, vital for enterprise compliance.

Clarifai’s RAG and embedding APIs are optimised for GH200 and help vector search and summarisation. Builders can deploy LLMs with massive context home windows and combine exterior information sources with out worrying about reminiscence administration. Clarifai’s pricing is clear and sometimes tied to utilization, providing value‑efficient entry to GH200 assets.

Finest Practices for Deploying on GH200

  1. Use massive batch sizes – Leverage the unified reminiscence to extend batch sizes for inference; this reduces overhead and improves throughput.
  2. Offload KV caches to CPU reminiscence – Retailer key–worth caches in LPDDR reminiscence to unlock HBM for compute; NVLink‑C2C ensures low‑latency entry.
  3. Combine vector databases – For RAG, join Clarifai’s APIs to vector shops; hold indices in unified reminiscence to speed up search.
  4. Monitor reminiscence bandwidth – Use profiling instruments to detect reminiscence bottlenecks. Information placement issues; excessive‑frequency tensors ought to keep in HBM.
  5. Undertake containerised environments – Use Clarifai’s native runner to deal with ARM dependencies and keep reproducibility.
  6. Plan cross‑{hardware} pipelines – Mix GH200 for reminiscence‑intensive levels with H100/B200 for compute‑heavy levels, orchestrated by way of Clarifai’s platform.

Skilled Insights

  • Reminiscence‑conscious design – Rethink your algorithms to use unified reminiscence: pre‑allocate massive buffers, scale back information copies and tune for NVLink bandwidth.
  • GPU sharing boosts ROI – Fractioning a GH200 throughout a number of workloads will increase utilisation and lowers value per job; that is particularly helpful for startups.
  • Clarifai’s cross‑cloud synergy – Working workloads throughout a number of clouds prevents vendor lock‑in and ensures excessive availability.

Regularly Requested Questions

Q1: Is GH200 obtainable in the present day and the way a lot does it value? – Sure. GH200 techniques can be found by way of cloud suppliers and specialist GPU clouds. Rental costs vary from $4–$6 per hour relying on supplier and area. Clarifai presents utilization‑based mostly pricing by its platform.

Q2: How does GH200 differ from H100 and H200? – GH200 fuses a CPU and GPU on one module with 900 GB/s NVLink‑C2C, making a unified reminiscence pool of as much as 624 GB. H100 is a standalone GPU with 80 GB HBM, whereas H200 upgrades the H100 with 141 GB HBM3e. GH200 is healthier for reminiscence‑sure duties; H100/H200 stay robust for compute‑sure workloads and x86 compatibility.

Q3: Will I must rewrite my code to run on GH200? – Most AI frameworks (PyTorch, TensorFlow, JAX) help ARM and CUDA. Nevertheless, some libraries may have recompilation. Utilizing containerised environments (e.g., Clarifai native runner) simplifies the migration.

This fall: What about energy consumption and cooling? – GH200 nodes can eat round 1 000 W. Guarantee satisfactory energy and cooling. Good autoscaling reduces idle consumption.

Q5: When will Blackwell/B200/Rubin be extensively obtainable? – Nvidia has introduced B200 and Rubin platforms, however broad availability might arrive in late 2026 or 2027. Rubin guarantees 10× decrease inference value and 4× fewer GPUs in comparison with Blackwell. For many builders, GH200 will stay a flagship selection by 2026.

Conclusion

The Nvidia GH200 marks a turning level in AI {hardware}. By fusing a 72‑core Grace CPU with a Hopper/H200 GPU by way of NVLink‑C2C, it delivers a unified reminiscence pool as much as 624 GB and eliminates the bottlenecks of PCIe. Benchmarks present as much as 1.8× extra efficiency than the H100 and massive enhancements in value per token for LLM inference. These features stem from reminiscence: the flexibility to maintain whole fashions, key–worth caches and vector indices on chip. Whereas GH200 isn’t good—software program on ARM requires adaptation, energy consumption is excessive and provide is restricted—it presents unparalleled capabilities for reminiscence‑sure workloads.

As AI enters the period of trillion‑parameter fashions, reminiscence‑centric computing turns into important. GH200 paves the way in which for Blackwell, Rubin and past, with bigger reminiscence swimming pools and extra environment friendly NVLink interconnects. Whether or not you’re constructing chatbots, producing video, exploring scientific simulations or coaching recommender techniques, GH200 gives a strong platform. Partnering with Clarifai simplifies adoption: their compute platform presents good autoscaling, GPU fractioning and cross‑cloud orchestration, making the GH200 accessible to groups of all sizes. By understanding the structure, efficiency traits and greatest practices outlined right here, you’ll be able to harness the GH200’s potential and put together for the subsequent wave of AI innovation.



ShinyHunters declare to be behind SSO-account information theft assaults

0


The ShinyHunters extortion gang claims it’s behind a wave of ongoing voice phishing assaults focusing on single sign-on (SSO) accounts at Okta, Microsoft, and Google, enabling menace actors to breach company SaaS platforms and steal firm information for extortion.

In these assaults, menace actors impersonate IT assist and name staff, tricking them into coming into their credentials and multi-factor authentication (MFA) codes on phishing websites that impersonate firm login portals.

As soon as compromised, the attackers acquire entry to the sufferer’s SSO account, which might present entry to different related enterprise functions and companies.

Wiz

SSO companies from Okta, Microsoft Entra, and Google allow corporations to hyperlink third-party functions right into a single authentication move, giving staff entry to cloud companies, inside instruments, and enterprise platforms with a single login. 

These SSO dashboards usually listing all related companies, making a compromised account a gateway into company programs and information.

Platforms generally related by SSO embrace Salesforce, Microsoft 365, Google Workspace, Dropbox, Adobe, SAP, Slack, Zendesk, Atlassian, and plenty of others.

Microsoft Entra single sign-on (SSO) dashboard
Microsoft Entra single sign-on (SSO) dashboard
Supply: Microsoft

Vishing assaults used for information theft

As first reported by BleepingComputer, menace actors have been finishing up these assaults by calling staff and posing as IT employees, utilizing social engineering to persuade them to log into phishing pages and full MFA challenges in actual time.

After having access to a sufferer’s SSO account, the attackers browse the listing of related functions and start harvesting information from the platforms out there to that consumer.

BleepingComputer is conscious of a number of corporations focused in these assaults which have since obtained extortion calls for signed by ShinyHunters, indicating that the group was behind the intrusions.

BleepingComputer contacted Okta earlier this week in regards to the breaches, however the firm declined to touch upon the information theft assaults.

Nevertheless, Okta launched a report yesterday describing the phishing kits utilized in these voice-based assaults, which match what BleepingComputer has been advised.

In response to Okta, the phishing kits embrace a web-based management panel that permits attackers to dynamically change what a sufferer sees on a phishing website whereas talking to them on the telephone. This permits menace actors to information victims by every step of the login and MFA authentication course of.

If the attackers enter stolen credentials into the actual service and are prompted for MFA, they’ll show new dialog containers on the phishing website in actual time to instruct a sufferer to approve a push notification, enter a TOTP code, or carry out different authentication steps.

Phishing kit letting attackers display different dialogs while calling victims
A phishing package lets attackers show totally different dialogs whereas calling victims
Supply: Okta

ShinyHunters declare accountability

Whereas ShinyHunters declined to touch upon the assaults final evening, the group confirmed to BleepingComputer this morning that it’s accountable for a number of the social engineering assaults.

“We affirm we’re behind the assaults,” ShinyHunters advised BleepingComputer. “We’re unable to share additional particulars at the moment, moreover the truth that Salesforce stays our main curiosity and goal, the remainder are benefactors.”

The group additionally confirmed different elements of BleepingComputer’s reporting, together with particulars in regards to the phishing infrastructure and domains used within the marketing campaign. Nevertheless, it disputed {that a} screenshot of a phishing package command-and-control server shared by Okta was for its platform, claiming as an alternative that theirs was constructed in-house.

ShinyHunters claimed it’s focusing on not solely Okta but additionally Microsoft Entra and Google SSO platforms.

Microsoft stated it has nothing to share at the moment, and Google stated it had no proof its merchandise have been being abused within the marketing campaign.

“Right now, now we have no indication that Google itself or its merchandise are affected by this marketing campaign,” a Google spokesperson advised BleepingComputer.

ShinyHunters claims to be utilizing information stolen in earlier breaches, such because the widespread Salesforce information theft assaults, to establish and phone staff. This information contains telephone numbers, job titles, names, and different particulars used to make the social-engineering calls extra convincing.

Final evening, the group relaunched its Tor information leak website, which at the moment lists breaches at SoundCloud, Betterment, and Crunchbase.

SoundCloud beforehand disclosed an information breach in December 2025, whereas Betterment confirmed this month that its electronic mail platform had been abused to ship cryptocurrency scams and that information was stolen.

Crunchbase, which had not beforehand disclosed a breach, confirmed right this moment that information was stolen from its company community.

“Crunchbase detected a cybersecurity incident the place a menace actor exfiltrated sure paperwork from our company community,” an organization spokesperson advised BleepingComputer. “No enterprise operations have been disrupted by this incident. We now have contained the incident and our programs are safe.”

“Upon detecting the incident we engaged cybersecurity consultants and contacted federal legislation enforcement. We’re reviewing the impacted data to find out if any notifications are required in line with relevant authorized necessities.”

Whether or not you are cleansing up outdated keys or setting guardrails for AI-generated code, this information helps your staff construct securely from the beginning.

Get the cheat sheet and take the guesswork out of secrets and techniques administration.

A Hashish Customary Measure, Weird Lifeforms, And Extra! : ScienceAlert

0


This week in science: Scientists suggest new commonplace measures for hashish use; mysterious historical lifeforms do not comfortably match on the tree of life; the world’s oldest rock artwork found; and rather more!

Scientists Figured Out a Customary Measure For Hashish Use

(Terrance Barksdale/Canva)

UK scientists have calculated a normal measure for hashish as THC models, which will help customers and docs monitor consumption.

A 0.45-gram joint of sturdy natural hashish would possibly comprise 12.78 commonplace THC models, whereas weaker, seeded natural hashish can comprise simply 3.78 THC models, in response to the brand new estimates.

Learn the total story right here.

Mysterious Giants May Be a Entire New Type of Life That No Longer Exists

Mysterious Giants Could Be a Whole New Kind of Life That No Longer Exists
Prototaxites does not examine with another life kind we all know of. (Loron et al., Science, 2025)

Prototaxites – 8-m (26-ft) tall organisms that lived 400 million years in the past – does not belong to any recognized class of life at the moment, a brand new examine finds.

By a evaluate of microscopic anatomy and chemical evaluation of its tubular constructions, the workforce of researchers systematically eradicated every candidate group, leaving no trendy organism with which it would share some type of ancestral relationship.

Fungi? Rejected due to the distinctive manner its anatomy connects.

A plant or algae? Not going given its chemical composition.

A mixture of the 2, resembling a lichen? Not with that anatomy.

Some weird animal? Cell partitions say no likelihood.

Learn the total story right here.

Research Confirms Why Some Individuals Get Drunk With out Touching Alcohol

Study Confirms Why Some People Get Drunk Without Touching Alcohol
Klebsiella pneumoniae, one in every of two bacterial species implicated. (Callista Photos/Getty Photos)

E. coli and Ok. pneumoniae have been recognized as the 2 principal micro organism behind a uncommon syndrome the place alcohol brews within the intestine after consuming.

Audition now for ScienceAlert's Casting Call

The findings counsel reduction for sufferers would possibly lie in selling or introducing, by means of dietary adjustments, stool transplants or probiotics, different strains of intestine micro organism that readily metabolize ethanol.

Learn the total story right here.

World’s Oldest Rock Artwork Found in Indonesian Cave

World's Oldest Rock Art Discovered in Indonesian Cave
The information of the fingers had been pointed, a mode solely seen in Sulawesi, and which can have been associated to animals. (Ahdi Agus Oktaviana)

The world’s oldest recognized rock artwork has been found in a collapse Indonesia, dated to a minimum of 67,800 years in the past.

“What we’re seeing in Indonesia might be not a sequence of remoted surprises, however the gradual revealing of a a lot deeper and older cultural custom that has merely been invisible to us till lately,” archaeologist Maxime Aubert of Griffith College in Australia, who co-led the analysis, instructed ScienceAlert.

Learn the total story right here.

Scientists Discovered a Sugar That is Candy, Low-Calorie, And Would not Spike Insulin

Scientists Found a Sugar That's Sweet, Low-Calorie, And Doesn't Spike Insulin
(Uma Shankar sharma/Getty Photos)

Scientists have discovered a easy technique to produce a pure sugar known as tagatose, 92% as candy as sucrose however with solely 30% of the energy.

What’s notably thrilling about it’s that it doesn’t spike insulin ranges like sucrose or high-intensity synthetic sweeteners – making it a doubtlessly engaging possibility for these with diabetes or blood glucose points.

Learn the total story right here.

Shingles Vaccine Linked to Slower Organic Growing older, Research Finds

A reddish colored virus particle seen under a microscope
Transmission electron microscope picture of the varicella zoster virus, which causes shingles. (Science Photograph Library – HEATHER DAVIES/Getty Photos)

The shingles vaccine has been linked to slower growing old and fewer irritation, in a examine of three,800 folks over the age of 70.

The findings counsel that the vaccine could have “broad” and lingering results on “aging-related processes”, in response to the authors, gerontologists Jung Ki Kim and Eileen Crimmins from the College of Southern California.

Learn the total story right here.

jQuery 4.0.0 JavaScript library options trusted varieties

0

Model 4.0.0 of the still-widely-used jQuery JavaScript library is now accessible. Celebrated as the primary main launch in almost 10 years, jQuery 4.0.0 options help for trusted varieties and a brand new, slimmer construct.

Introduced January 17, the most recent model of the jQuery JavaScript library might be downloaded from jquery.com.  Trusted varieties in jQuery 4.0.0 be certain that HTML within the TrustedHTML interface might be enter to jQuery manipulation strategies in compliance with a browser’s Content material Safety Coverage (CSP) required-trusted-types-for directive. As well as, whereas some AJAX requests already had been utilizing to keep away from any CSP errors brought on by inline scripts. There nonetheless are just a few instances the place XHR is used for asynchronous script requests, corresponding to when the "headers" possibility is handed, however