Close Menu
  • Home
  • AI
  • Big Data
  • Cloud Computing
  • iOS Development
  • IoT
  • IT/ Cybersecurity
  • Tech
    • Nanotechnology
    • Green Technology
    • Apple
    • Software Development
    • Software Engineering

Subscribe to Updates

Get the latest technology news from Bigteetechhub about IT, Cybersecurity and Big Data.

    What's Hot

    Microsoft named a Leader in 2026 Gartner® Magic Quadrant™ for Integration Platform as a Service

    April 23, 2026

    Bluesky now supports better quality photos

    April 23, 2026

    Enhancing Web Design: Recognizing Accessibility Issues Now

    April 23, 2026
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    Big Tee Tech Hub
    • Home
    • AI
    • Big Data
    • Cloud Computing
    • iOS Development
    • IoT
    • IT/ Cybersecurity
    • Tech
      • Nanotechnology
      • Green Technology
      • Apple
      • Software Development
      • Software Engineering
    Big Tee Tech Hub
    Home»Big Data»Using Apache Sedona with AWS Glue to process billions of daily points from a geospatial dataset
    Big Data

    Using Apache Sedona with AWS Glue to process billions of daily points from a geospatial dataset

    big tee tech hubBy big tee tech hubApril 23, 20260018 Mins Read
    Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email Telegram WhatsApp
    Follow Us
    Google News Flipboard
    Using Apache Sedona with AWS Glue to process billions of daily points from a geospatial dataset
    Share
    Facebook Twitter LinkedIn Pinterest Email Copy Link


    Data strategy can use geospatial data to provide organizations with insights for decision-making and operational optimization. By incorporating geospatial data (such as GPS coordinates, points, polygons and geographic boundaries), businesses can uncover patterns, trends, and relationships that might otherwise remain hidden across multiple industries, from aviation and transportation to environmental studies and urban planning. Processing and analyzing this geospatial data at scale can be challenging, especially when dealing with billions of daily observations.

    In this post, we explore how to use Apache Sedona with AWS Glue to process and analyze massive geospatial datasets.

    Introduction to geospatial data

    Geospatial data is information that has a geographic component. It describes objects, events, or phenomena along with their location on the Earth’s surface. This data includes coordinates (latitude and longitude), shapes (points, lines, polygons), and associated attributes (such as the name of a city or the type of road).

    Key types of geospatial geometries (and examples of each in parentheses) include:

    • Point – Represents a single coordinate (a weather station).
    • MultiPoint – A collection of points (bus stops in a city).
    • LineString – A series of points connected in a line (a river or a flight path).
    • MultiLineString – Multiple lines (multiple flight routes).
    • Polygon – A closed area (the boundary of a city).
    • MultiPolygon – Multiple polygons (national parks in a country).

    Geospatial datasets come in different formats, each designed to store and represent different types of geographic information. Common formats for geospatial data are vector formats (Shapefile, GeoJSON), raster formats (GeoTIFF, ESRI Grid), GPS formats (GPX, NMEA), web formats (WMS, GeoRSS) among others.

    Core concepts of Apache Sedona

    Apache Sedona is an open-source computing framework for processing large-scale geospatial data. Built on top of Apache Spark, Sedona extends Spark’s capabilities to handle spatial operations efficiently. At its core, Sedona introduces several key concepts that enable distributed spatial processing. These include Spatial Resilient Distributed Datasets (SRDDs), which allow for the distribution of spatial data across a cluster, and Spatial SQL, which provides a familiar SQL-like interface for spatial queries. Some of the core capabilities of Apache Sedona are:

    • Efficient spatial data types like points, lines and polygons.
    • Spatial operations and functions such as ST_Contains (check if point is inside of a polygon), ST_Intersects (check if point is inside of a polygon), ST_H3CellIDs (geospatial indexing system developed by Uber, return the H3 cell ID(s) that contain the given point at the specified resolution).
    • Spatial joins to combine different spatial datasets.
    • Integration with Spark SQL (geospatial functions to run spatial SQL queries).
    • Spatial indexing techniques, such as quad-trees and R-trees, to optimize query performance.

    For more information about the functions available in Apache Sedona, visit the official Sedona Functions documentation.

    Use case

    This use case consists of a global air traffic visualization and analysis platform that processes and displays real-time or historical aircraft tracking data on an interactive world map. Using unique aircraft identifiers from the International Civic Aviation Organization (ICAO), the system ingests trajectory records containing information such as geographic position (latitude and longitude), altitude, speed, and flight direction, then transforms this raw data into two complementary visual layers. The Flight Tracks Layer plots the routes traveled by each aircraft individually, allowing for the analysis of specific trajectories and navigation patterns. The Flight Density Layer uses hexagonal spatial indexing (H3) to aggregate and identify regions of higher air traffic concentration worldwide, revealing busy air corridors, aviation hubs, and high-density flight zones.

    The dataset used for this use case is historical flight tracker data from ADSB.lol. ADSB.lol provides unfiltered flight tracker with a focus on open data. Data is also freely available via the API. The data contains a file per aircraft, a JSON gzip file containing the data for that aircraft for the day.

    This is a JSON trace file format sample:

    {
        icao: "0123ac", // hex id of the aircraft
        timestamp: 1609275898.495, // unix timestamp in seconds since epoch (1970)
        trace: [
            [ seconds after timestamp,
                lat,
                lon,
                altitude in ft or "ground" or null,
                ground speed in knots or null,
                track in degrees or null, (if altitude == "ground", this will be true heading instead of track)
                flags as a bitfield: (use bitwise and to extract data)
                    (flags & 1 > 0): position is stale (no position received for 20 seconds before this one)
                    (flags & 2 > 0): start of a new leg (tries to detect a separation point between landing and takeoff that separates flights)
                    (flags & 4 > 0): vertical rate is geometric and not barometric
                    (flags & 8 > 0): altitude is geometric and not barometric
                 ,
                vertical rate in fpm or null,
                aircraft object with extra details or null,
                type / source of this position or null,
                geometric altitude or null,
                geometric vertical rate or null,
                indicated airspeed or null,
                roll angle or null
            ],
        ]
    }

    For this use case, this is a simplified schema of the dataset after processing:

    • icao - Unique aircraft identifier
    • timestamp - Epoch timestamp of the observation (converted to readable format)
    • trace.lat / trace.lon - Latitude and longitude of the aircraft
    • trace.altitude - Aircraft altitude
    • trace.ground_speed - Ground speed
    • geometry - Geospatial geometry of the observation point (Point)

    Solution overview

    This solution enables aircraft tracking and analysis. The data can be visualized on maps and used for aviation management and safety applications. The process begins with data acquisition, extracting the compressed JSON files from TAR archives, then transforms this raw data into geospatial objects, aggregating them into H3 cells for efficient analysis. The processed data schema includes ICAO aircraft identifiers, timestamps, latitude/longitude coordinates, and derived fields such as H3 cell identifiers and point counts per cell. This structure allows detailed tracking of individual flights and aggregate analysis of traffic patterns. For visualization, you can generate density maps using the H3 grid system and create visual representations of individual flight tracks. The architecture data flow is as follows:

    • Data ingestion – Aircraft observation data stored as JSON compressed files in Amazon Simple Storage Service (Amazon S3).
    • Data processing – AWS Glue jobs using Apache Sedona for geospatial processing.
    • Data visualization – Spark SQL with Sedona’s spatial functions to extract insights and export data to visualize the information in a map on Kepler.gl.

    The following figure illustrates this solution.

    AWS architecture diagram showing a geospatial data processing pipeline.

    Prerequisites

    You will need the following for this solution:

    • An AWS Account and a user with AWS Console access.
    • Access to a Linux terminal and the AWS Command Line Interface (AWS CLI).
    • An IAM role for AWS Glue with list, read, and write permissions for Amazon S3 buckets.
    • An Amazon S3 Bucket for flight files. For this example, name the bucket blog-sedona-nessie--, using your account number and region.
    • An Amazon S3 bucket for artifacts and Sedona libraries. For this example, name the bucket blog-sedona-artifacts--, using your account number and region.
    • Download a day of historical data from ADSB.lol. In our examples, we used v2025.05.29-planes-readsb-prod-0tmp.tar.aa and v2025.05.29-planes-readsb-prod-0tmp.tar.ab.
    • Download the Apache Sedona libraries. The example was created using sedona-spark-shaded-3.5_2.12-1.7.1.jar and geotools-wrapper-1.7.1-28.5.jar.
    • Download the AWS Glue script from AWS Sample to process the geospatial data.
    • Review the AWS Glue security best practices, especially IAM least-privilege, encryption for sensitive data at rest and in transit, and configuring VPC Endpoints to prevent data from routing through the public internet.

    Solution walkthrough

    From now on, executing the next steps will incur costs on AWS. This step-by-step walkthrough demonstrates an approach to processing and analyzing large-scale geospatial flight data using Apache Sedona and Uber’s H3 spatial indexing system, using AWS Glue for distributed processing and Apache Sedona for efficient geospatial computations. It explains how to ingest raw flight data, transform it using Sedona’s geospatial functions, and index it with H3 for optimized spatial queries. Finally, it also demonstrates how to visualize the data using Kepler.gl. For data processing, it is possible to use both Glue scripts and Glue notebooks. In this post, we will focus only on Glue scripts.

    Upload the Apache Sedona libraries to Amazon S3

    1. Open your OS terminal command line.
    2. Create a folder to download the Sedona libraries and name it jar.
      
      	# Create a directory for the Sedona libraries (JARs files)
      	mkdir jar
      	# Go to the folder JARs folder
      	cd jar
      	
    3. Download the Apache Sedona libraries.
      
      	# Download required Sedona libraries (JARs files)
      	wget 
      	wget 
      	
    4. Upload the Sedona libraries (JARs files) to Amazon S3. In this example, we use the S3 path s3://aws-blog-post-sedona-artifacts/jar/.
      
      	# Upload the JARs files to Amazon S3 bucket
      	aws s3 cp . s3://blog-sedona-artifacts--/jar/ --recursive
      	
    5. Your Amazon S3 folder should now look similar to the following image:

    Amazon S3 console screenshot displaying the jar folder contents in blog-sedona-artifacts bucket.

    Download and upload the geospatial data to Amazon S3

    1. Open your OS terminal command line.
    2. Create a folder to download the flight files and name it adsb_dataset.
      		# Create a directory for download the geospatial flight files
      		mkdir adsb_dataset
      		# Go to the folder for geospatial flight files
      		cd adsb_dataset
      	
    3. Download the flight files data from adsblol GitHub repository.
      	# Download the geospatial flight files in the folder created
      	wget 
      	wget 
      	
    4. Extract the flight files.
      	# Combine the two the tar files together
      	cat v2025.05.29* >> combined.tar
      	# Extract the json flight files from the tar file
      	tar xf combined.tar
      	
    5. Copy the flight files to Amazon S3. In this case, we are using the S3 folder: s3://blog-sedona-nessie--/raw/adsb-2025-05-28/traces/.
      	# Copy the json flight files to Amazon S3
      	aws s3 cp ./traces/ s3://blog-sedona-nessie--/raw/adsb-2025-05-28/traces/ --recursive
      	
    6. Your Amazon S3 folder should now look similar to the following image.

    Amazon S3 console showing JSON trace files in the path raw/adsb-2025-05-28/traces/00/.

    Create an AWS Glue job and set up the job

    Now, we are ready to define the AWS Glue job using Apache Sedona to read the geospatial data files. To create a Glue job:

    1. Open the AWS Glue console.
    2. On the Notebooks page, choose Script editor.

    AWS Glue Studio jobs creation interface showing three job creation methods: Visual ETL with data flow interface, Notebook for interactive coding, and Script editor for code authoring

    1. On the Script screen, for the engine, choose Spark, then select the option Upload script.
    2. Choose Choose file. Find the process_sedona_geo_track.py file, then choose Create script.

    Script creation dialog box with Spark engine selected. Upload script option is active, showing successfully uploaded file process_sedona_geo_track.py.

    1. Rename the job from Untitled to process_sedona_geo_track.
    2. Choose Save.
    3. Now, let’s set up the AWS Glue job. Choose Job Details.
    4. Choose the IAM Role created to be used with Glue. For this example, we use blog-glue.
    5. Set the Glue version to Glue 5.0 and the Worker type as needed. For this example, G.1X is sufficient, but we use G.2X to speed up processing.

    AWS Glue job details configuration page for process_sedona_geo_track.

    1. Now, let’s import the libraries for Apache Sedona.
    2. In the Dependent JARs path, type the path of the JAR files for Apache Sedona that you uploaded in the preceding steps. For this example, we used s3://blog-sedona-artifacts--/jar/sedona-spark-shaded-3.5_2.12-1.7.1.jar,s3://blog-sedona-artifacts--/jar/geotools-wrapper-1.7.1-28.5.jar
    3. In Additional Python modules path, enter the modules for Apache Sedona: apache-sedona==1.7.1,geopandas==0.13.2,shapely==2.0.1,pyproj==3.6.0,fiona==1.9.5,rtree==1.2.0

    ob libraries configuration section showing Dependent JARs path pointing to S3 bucket.

    1. In the Job parameters section, in the Key field, type —BUCKET_NAME. For its Value, enter your bucket name. In this example, ours is blog-sedona-nessie--.

    ob parameters configuration interface showing key-value pair with --BUCKET_NAME parameter.

    1. Choose Save.

    Processing the geospatial flights data

    Before we run the job, let’s understand how the code works. First, import the Apache Sedona libraries:

    import json 
    import gzip 
    from sedona.spark import SedonaContext

    Next, initialize the Sedona context using an existing Spark session:

    sedona = SedonaContext.create(spark)

    After that, create a function for handling compressed JSON data:

    def parse_gzip_json(byte_content):
            try:
                decompressed = gzip.decompress(byte_content)
                return json.loads(decompressed.decode('utf-8'))
            except Exception as e:
                print(f"Error during gzip parse: {str(e)}")
                return None

    Add a function to transform raw tracking data into a structured format suitable for a valid coordinates process:

    def flatten_records(json_obj):
        records = []
        if "trace" in json_obj and isinstance(json_obj["trace"], list):
            for point in json_obj["trace"]:
                if len(point) >= 3:
                    lat, lon = float(point[1]), float(point[2])
                    if -90 <= lat <= 90 and -180 <= lon <= 180:
                        records.append(Row(
                            icao=json_obj.get("icao", None),
                            timestamp=json_obj.get("timestamp", None),
                            lat=lat,
                            lon=lon
                        ))
        return records

    The flat_rdd variable applies these functions to the structured data from the original gzipped JSON. Each element in this RDD is a Row object representing a single data point from an aircraft’s trace, with fields for ICAO, timestamp, latitude, and longitude.

    flat_rdd = raw_rdd.map(lambda x: parse_gzip_json(x[1])).filter(lambda x: x is not None).flatMap(flatten_records)

    The ADSB trace files contain a deeply nested JSON structure where the trace field holds an array of mixed-type arrays, compressed in Gzip format. For this specific case, developing a UDF represented one of the most practical and efficient solutions. Since Gzip is a non-splittable format, Spark is unable to parallelize processing, constraining both methods to a single worker per file and processing the data multiple times across JVM decompression, full JSON parsing, and subsequent re-parsing operations. The UDF bypasses all of this by reading raw bytes and doing everything in a single Python pass: decompress → parse → extract → validate, returning only the small set of needed fields directly to Spark.

    The Spark SQL query processes geographic trace data using the H3 hexagonal grid system, converting point data into a regularized hexagonal grid that can help identify areas of high point density. A resolution of 5 was adopted, producing hexagons of approximately 253 km² (roughly the same size as the city of Edinburgh, Scotland, which is approximately 264 km²), for its ability to effectively capture route density patterns at the city and metropolitan level.

    h3_traces_df = spark.sql("""
    WITH base_h3 AS (
        SELECT
            ST_H3CellIDs(geometry, 5, false)[0] AS h3_index,
            lat,
            lon
        FROM traces
    )
    SELECT
        COUNT(*) AS num, -- Count points in each H3 cell
        h3_index,
        AVG(lon) AS center_lon,
        AVG(lat) AS center_lat
    FROM base_h3
    GROUP BY h3_index
    """)
    

    Finally, this code prepares the datasets for visualization purposes. The first dataset is based on the aircraft unique identifier. The complete dataset for a single day can contain more than 80 million data points. A random sampling rate of 0.1% was applied, which proves sufficient to illustrate route density patterns without overwhelming the Kepler.gl browser renderer. The second dataset aggregates trace points into hexagonal spatial cells (result from the query above).

    points_viz_sampled = df_points.select(
        col("icao"), # Aircraft unique identifier (24-bit address)
        col("timestamp").cast("double").alias("timestamp"),
        col("lat").cast("double").alias("lat"),
        col("lon").cast("double").alias("lon")
    ).sample(False, 0.001)
    
    h3_viz_csv = h3_traces_df.select(
        col("num").alias("point_count"),
        col("h3_index").cast("string").alias("h3_index"),
        col("center_lon"),
        col("center_lat")
    )

    Now that we understand the code, let’s run it.

    1. Open the AWS Glue console.
    2. On the ETL jobs >> Notebooks page, choose the job name process_sedona_geo_track.
    3. Choose Run.

    Python script editor showing import statements for process_sedona_geo_track job.

    1. Now, it is possible to monitor the job by choosing the Runs tab.
    2. It may take a few minutes to run the entire job. It took nearly 8 minutes to process approximately 2.50 GB (67,540 compressed files) with 20 DPUs. After the job is processed, you should see your job with the status Succeeded.

    Job runs monitoring dashboard showing successful execution on June 5, 2025, running from 12:28:03 to 12:36:37 with 8 minutes 19 seconds duration.

    Now your data should be saved for a preview visualization demo in a folder named s3://blog-sedona-nessie--/visualization/.

    Performance insights

    The workload characterization of this job reveals a CPU-intensive profile, primarily because of the processing of small binary files with GZIP compression and subsequent JSON parsing. Given the inherent nature of this pipeline, which includes Python UDF serialization and partial single-partition write stages, linear scaling does not yield proportional performance gains. The following table presents an analysis of AWS Glue configurations, evaluating the trade-off between computational capacity, execution duration, and associated costs:

    Duration Capacity (DPUs) Worker type Glue version Estimated Cost*
    10 m 7 s 32 DPUs G.1X 5 $2.34
    11 m 50 s 10 DPUs G.1X 5 $0.88
    19 m 7 s 4 DPUs G.1X 5 $0.59
    8 m 19 s 20 DPUs G.2X 5 $1.32

    *Estimated Cost = DPUs x Duration (hours) x $0.44 per DPU-hour (us-east-1)

    Visualizing and analyzing geospatial data with Kepler.gl

    Kepler.gl is an open-source geospatial analysis tool developed by Uber with code available at Github. Kepler.gl is designed for large-scale data exploration and visualization, offering multiple map layers, including point, arc, heatmap, and 3D hexagon. It supports various file formats like CSV, GeoJSON, and KML. In this use case, we will use Kepler.gl to present interactive visualizations that illustrate flight patterns, routes, and densities across global airspace.

    Downloading the geospatial files

    Before we can view the graph, we will need to download the flight files to our local machine, unzip them, and rename them (to make it easier to identify the files).

    1. Open your OS terminal command line.
    2. Create the folders to download the data processed in the steps before. In this case, we create kepler and kepler_csv.
      	#create kepler folders: first folder is to download the files,
      	#second folder is to organize the files to use in the next step
      	mkdir kepler
      	mkdir kepler_csv
      	
    3. Replace the bracketed variables with your account and directory information, then download all the CSV files.
      	#copy the files from Amazon S3 to local machine
      	aws s3 cp s3://blog-sedona-nessie--/visualization/ //kepler --recursive
      	
    4. Extract the files, rename them, and move them to another folder.
      	# Extract the files processed by Spark and Sedona
      	gzip -d ./kepler/kepler_h3_density/*.gz
      	gzip -d ./kepler/kepler_track_points_sample/*.gz
      	
      	# Rename the Spark output files to more readable names
      	cd ./kepler/kepler_h3_density/
      	ls
      	mv part-00000-*.csv kepler_h3_density.csv
      	cd ..
      	
      	cd ./kepler/kepler_track_points_sample/
      	ls
      	mv part-00000-*.csv kepler_track_points_sample.csv
      	cd ..
      	
      	# Ensure the output folder exists
      	mkdir -p ../kepler_csv
      	
      	# Copy the renamed CSV files to the folder that will be used as input in kepler.gl
      	cp ./kepler/kepler_h3_density/*.csv ../kepler_csv
      	cp ./kepler/kepler_track_points_sample/*.csv ../kepler_csv
      	
    5. Your kepler_csv folder should look similar to the return of the command below.
      	#list the files in the kepler_csv directory
      	ls -l
      	total 11684
      	-rw-rw-r-- 1 ec2-user ec2-user 8630110 Jun 12 14:47 kepler_h3_density.csv
      	-rw-rw-r-- 1 ec2-user ec2-user 3331763 Jun 12 14:47 kepler_track_points_sample.csv
      	

    Visualizing the data in a graph

    Now that you have saved the data to your local machine, you can analyze the flight data through interactive map graphics. To import the data into the Kepler.gl web visualization tool:

    1. Open the Kepler.gl Demo web application.
    2. Load data into Kepler.gl:
      1. Choose Add Data in the left panel.
      2. Drag and drop both CSV files (flight_points and h3_density) into the upload area.
      3. Confirm that both datasets are loaded successfully.
    3. Delete all layers.
    4. Create the Flight Density Layer:
      1. Choose Add Layer in the left panel.
      2. In Basic, choose H3 as the layer type, then add the following configuration:
        1. Layer Name: Flight Density
        2. Data Source: kepler_h3_density.csv
        3. Hex ID: h3_index
      3. In the Fill Color section:
        1. Color: point_count
        2. Color Scale: Quantile.
        3. Color Range: Choose a blue/green gradient.
      4. Set Opacity to 0.7.
      5. In the Coverage section, set it to 0.9.
    5. Create the Flight Tracks Layer:
      1. Choose Add Layer in the left panel.
      2. In Basic, choose Point as the layer type, then add the following configuration:
        1. Layer Name: Flight Tracks
        2. Data Source: kepler_track_points_sample.csv
        3. Columns:
          1. Latitude: lat
          2. Longitude: lon
      3. In the Fill Color section:
        1. Solid Color: Orange
        2. Opacity: 0.3
      4. Set the Point’s Radius to 1
    6. The layers should look similar to the following figure.

    Kepler.gl layer configuration panel for Flight Density H3 layer using kepler_h3_density.csv data source.

    1. The graph visualization should now show flight density through color-coded hexagons, with individual flight tracks visible as orange points:

    Kepler.gl interactive map visualization displaying global flight density heatmap. High-density areas shown in yellow over North America, particularly the United States.

    There you go! Now that you have knowledge about geospatial data and have created your first use case, take the opportunity to do some analysis and learn some interesting facts about flight patterns.

    It is possible to experiment with other interesting types of analysis in Kepler.gl, such as Time Playback.

    Clean up

    To clean up your resources, complete the following tasks:

    1. Delete the AWS Glue job process_sedona_geo_track.
    2. Delete content from the Amazon S3 buckets: blog-sedona-artifacts-- and blog-sedona-nessie--.

    Conclusion

    In this post, we showed how processing geospatial data can present significant challenges due to its complex nature (from big data to data structure format). For this use case of flight trackers, it involves vast amounts of information across multiple dimensions such as time, location, altitude, and flight paths, however, the combination of Spark’s distributed computing capabilities and Sedona’s optimized geospatial functions helps overcome those challenges. The spatial partitioning and indexing features of Sedona, coupled with Spark’s framework, enable us to perform complex spatial joins and proximity analyses efficiently, simplifying the overall data processing workflow.

    The serverless nature of AWS Glue eliminates the need for managing infrastructure while automatically scaling resources based on workload demands, making it an ideal platform for processing growing volumes of flight data. As the volume of flight data grows or as processing requirements fluctuate, with AWS Glue, you can quickly adjust resources to meet demand, ensuring optimal performance without the need for cluster management.

    By converting the processed results into CSV format and visualizing them in Kepler.gl, it is possible to create interactive visualizations that reveal patterns in flight paths, and you can efficiently analyze air traffic patterns, routes, and other insights. This end-to-end solution demonstrates how a modern data strategy in AWS with the support of open-source tools can transform raw geospatial data into actionable insights.


    About the authors

    Ruan

    Ruan Roloff is a Lead GTM Specialist Architect for Analytics and AI at AWS. During his time at AWS, he was responsible for the data journey and AI product strategy of customers across a range of industries, including finance, oil and gas, manufacturing, digital natives, public sector, and startups. He has helped these organizations achieve multi-million dollar use cases. Outside of work, Ruan likes to assemble and disassemble things, fish on the beach with friends, play SFII, and go hiking in the woods with his family.

    Lucas

    Lucas Vitoreti is a ProServe Data & Analytics Specialist at AWS with 12+ years in the data domain. Architects and delivers solutions for data warehouses, lakes, lakehouses, and meshes, helping organizations transform their data strategies and achieve business outcomes. Expertise in scalable data architectures and guiding data-driven transformations. He balances professional life with weightlifting, music, and family time.

    Denys

    Denys Gonzaga is a ProServe Consultant at AWS, he is an experienced professional with over 15 years of working across multiple technical domains, with a strong focus on development and data analytics. Throughout his career, he has successfully applied his skills in various industries, including aerospace, finance, telecommunications, and retail. Outside of AWS, Denys enjoys spending time with his family and playing video games.



    Source link

    Apache AWS Billions Daily Dataset geospatial Glue Points Process Sedona
    Follow on Google News Follow on Flipboard
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
    tonirufai
    big tee tech hub
    • Website

    Related Posts

    AWS Weekly Roundup: Claude Opus 4.7 in Amazon Bedrock, AWS Interconnect GA, and more (April 20, 2026)

    April 22, 2026

    How conversational analytics removes the BI bottleneck

    April 22, 2026

    Britain’s empty buildings could unlock billions, advocacy group tells treasury

    April 22, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Editors Picks

    Microsoft named a Leader in 2026 Gartner® Magic Quadrant™ for Integration Platform as a Service

    April 23, 2026

    Bluesky now supports better quality photos

    April 23, 2026

    Enhancing Web Design: Recognizing Accessibility Issues Now

    April 23, 2026

    Eric Tschetter on Decoupling Observability – Software Engineering Radio

    April 23, 2026
    Timer Code
    15 Second Timer for Articles
    20
    About Us
    About Us

    Welcome To big tee tech hub. Big tee tech hub is a Professional seo tools Platform. Here we will provide you only interesting content, which you will like very much. We’re dedicated to providing you the best of seo tools, with a focus on dependability and tools. We’re working to turn our passion for seo tools into a booming online website. We hope you enjoy our seo tools as much as we enjoy offering them to you.

    Don't Miss!

    Microsoft named a Leader in 2026 Gartner® Magic Quadrant™ for Integration Platform as a Service

    April 23, 2026

    Bluesky now supports better quality photos

    April 23, 2026

    Subscribe to Updates

    Get the latest technology news from Bigteetechhub about IT, Cybersecurity and Big Data.

      • About Us
      • Contact Us
      • Disclaimer
      • Privacy Policy
      • Terms and Conditions
      © 2026 bigteetechhub.All Right Reserved

      Type above and press Enter to search. Press Esc to cancel.