Latest Tech Papers: Confidential Computing, Serverless & Containers

by Admin 68 views
Latest 15 Papers - November 02, 2025

Hey guys! Here's the latest scoop on some cool tech papers. I've broken it down into three hot topics: Confidential Computing, Serverless, and Containers. Check it out!

Confidential Computing

Let's dive into Confidential Computing, shall we? This field is all about keeping your data safe and sound, even when it's being processed. It's like having a super-secure vault for your digital secrets.

Cross-Chain Sealed-Bid Auctions Using Confidential Compute Blockchains

This paper explores how to use confidential computing to make cross-chain sealed-bid auctions more secure. This is super important because it helps keep the bidding process fair and private. Imagine a world where your bids are secret until the very end โ€“ pretty cool, right? This leverages confidential compute blockchains to ensure that sensitive auction data remains protected throughout the process.

sNVMe-oF: Secure and Efficient Disaggregated Storage

Next up, we've got sNVMe-oF, which is all about making storage secure and efficient. It's like having a super-fast and reliable hard drive that's also extra safe.

Agora: Trust Less and Open More in Verification for Confidential Computing

This paper delves into making verification in confidential computing more trustworthy and open. The idea is to reduce the need to trust third parties, making the whole system more transparent. Think of it as building a house where you can see all the pipes and wires โ€“ you know exactly what's going on.

Confidential LLM Inference: Performance and Cost Across CPU and GPU TEEs

Here, the focus is on confidential LLM (Large Language Model) inference. The research explores the performance and cost of running these models on both CPUs and GPUs within Trusted Execution Environments (TEEs). Essentially, how can we use AI in a way that's both powerful and secure?

Dstack: A Zero Trust Framework for Confidential Containers

Dstack presents a Zero Trust approach for confidential containers. This means that no one is automatically trusted, and every access request is verified. It's like having a security guard at every door, checking credentials constantly.

Characterizing Trust Boundary Vulnerabilities in TEE Containers

This paper examines the weaknesses in TEE containers, specifically the trust boundaries. It's all about figuring out where the system might be vulnerable and how to patch those holes.

Characterization of GPU TEE Overheads in Distributed Data Parallel ML Training

This research looks at the extra work (overheads) that GPUs have to do when they're running in a TEE environment, especially when training machine learning (ML) models. The main goal here is to optimize performance by understanding and minimizing these overheads.

Confidential Serverless Computing

This paper discusses the concept of confidential serverless computing, where serverless functions are executed in a secure, confidential environment. It combines the benefits of serverless (like ease of use and scalability) with the security of confidential computing.

HSM and TPM Failures in Cloud: A Real-World Taxonomy and Emerging Defenses

This paper takes a look at Hardware Security Modules (HSMs) and Trusted Platform Modules (TPMs), two crucial security components in the cloud. It identifies the ways these systems can fail and proposes new defenses. This is vital for maintaining the integrity and security of cloud infrastructure.

Distilled Large Language Model in Confidential Computing Environment for System-on-Chip Design

This research explores using a smaller, more efficient Large Language Model (LLM) within a confidential computing environment for System-on-Chip (SoC) design. The goal is to provide a secure and efficient way to use AI in designing hardware.

Careful Whisper: Attestation for peer-to-peer Confidential Computing networks

This work focuses on attestation in peer-to-peer (P2P) confidential computing networks. Attestation is the process of verifying the integrity and security of a computing environment. This paper aims to ensure that the nodes in a P2P network are trustworthy.

NVIDIA GPU Confidential Computing Demystified

This paper is all about understanding NVIDIA GPUs in the context of confidential computing. It's like a guide that clarifies how these powerful processors can be used securely.

NanoZone: Scalable, Efficient, and Secure Memory Protection for Arm CCA

This paper introduces NanoZone, a system for memory protection in Arm CCA (Confidential Compute Architecture). NanoZone aims to provide a way to securely isolate and protect memory regions.

OpenCCA: An Open Framework to Enable Arm CCA Research

OpenCCA is an open-source framework designed to help researchers explore and develop Arm CCA technologies. It's like a toolkit that helps people build and test new security features.

Performance of Confidential Computing GPUs

This paper takes a look at how well GPUs perform in a confidential computing environment. This helps in understanding the trade-offs between security and performance.

Serverless

Alright, let's switch gears and talk about Serverless. Think of it as building with pre-made blocks. You don't have to worry about the underlying infrastructure; you just focus on your code.

Odyssey: An End-to-End System for Pareto-Optimal Serverless Query Processing

This paper discusses Odyssey, a system designed to optimize serverless query processing. It aims to find the best balance between performance and cost. It's all about making serverless applications run faster and cheaper.

ProFaaStinate: Delaying Serverless Function Calls to Optimize Platform Performance

This research explores delaying serverless function calls to improve overall platform performance. It's like scheduling your tasks to make the most of available resources.

GeoFF: Federated Serverless Workflows with Data Pre-Fetching

Here, the focus is on federated serverless workflows and data pre-fetching. This approach is aimed at improving the efficiency of serverless applications that work with data from different sources.

Serverless GPU Architecture for Enterprise HR Analytics: A Production-Scale BDaaS Implementation

This paper talks about using a serverless GPU architecture for HR analytics in an enterprise setting. It's a real-world example of how serverless can be used to solve business problems.

The Hidden Dangers of Public Serverless Repositories: An Empirical Security Assessment

This study looks at the security risks associated with using public serverless repositories. It's about identifying potential vulnerabilities and how to mitigate them. This looks at the security risks associated with the usage of public serverless repositories. It's crucial for developers to be aware of and address these risks to secure their applications.

Object as a Service: Simplifying Cloud-Native Development through Serverless Object Abstraction

This paper discusses using serverless object abstraction to simplify cloud-native development. The idea is to make it easier for developers to build and manage applications in the cloud.

FlexPipe: Adapting Dynamic LLM Serving Through Inflight Pipeline Refactoring in Fragmented Serverless Clusters

This work looks at FlexPipe, a method for adapting Large Language Model (LLM) serving in fragmented serverless clusters. It aims to make LLMs run smoothly even when resources are limited.

Multi-Event Triggers for Serverless Computing

This paper explores the use of multi-event triggers in serverless computing. It allows serverless functions to respond to multiple events simultaneously, making applications more flexible and responsive.

Towards Energy-Efficient Serverless Computing with Hardware Isolation

This research focuses on making serverless computing more energy-efficient using hardware isolation. It's like finding ways to save energy while keeping your applications secure.

Scalable Cosmic AI Inference using Cloud Serverless Computing

This paper discusses using cloud serverless computing for cosmic AI inference. It's an example of how serverless can be used for large-scale scientific computations.

Dynamic Function Configuration and its Management in Serverless Computing: A Taxonomy and Future Directions

This paper provides a detailed look at dynamic function configuration and its management in serverless computing. It's like a guide that helps you understand how serverless functions can be configured and managed effectively.

Lumos: Performance Characterization of WebAssembly as a Serverless Runtime in the Edge-Cloud Continuum

This work examines the performance of WebAssembly as a serverless runtime in the edge-cloud continuum. It's about understanding how WebAssembly can be used to run serverless applications in different environments.

Demystifying Serverless Costs on Public Platforms: Bridging Billing, Architecture, and OS Scheduling

This paper aims to clarify the costs associated with serverless computing on public platforms. It connects billing, architecture, and OS scheduling to provide a clear understanding of the financial implications.

Code once, Run Green: Automated Green Code Translation in Serverless Computing

This research focuses on green code translation in serverless computing. It's about automating the process of making code more energy-efficient, contributing to sustainability.

HydraServe: Minimizing Cold Start Latency for Serverless LLM Serving in Public Clouds

This paper presents HydraServe, a method for reducing cold start latency in serverless LLM serving on public clouds. Cold start latency is the delay when a serverless function starts for the first time.

Container

Finally, let's talk about Containers. Think of them as pre-packaged software units that include everything needed to run an application. They're like self-contained boxes that ensure your software works the same way everywhere.

Fast and Robust Point Containment Queries on Trimmed Surface

This paper focuses on how to quickly and reliably determine if a point is inside a trimmed surface. This is important in areas like computer graphics and 3D modeling.

Dynamic Dimensioning of Frequency Containment Reserves: The Case of the Nordic Grid

This research explores the dynamic dimensioning of frequency containment reserves, specifically in the Nordic power grid. Itโ€™s about ensuring the stability of the power grid by managing the available resources to contain frequency deviations.

The Atomic Instruction Gap: Instruction-Tuned LLMs Struggle with Simple, Self-Contained Directives

This paper discusses the limitations of instruction-tuned LLMs when dealing with simple, self-contained directives. It's like finding the weaknesses in how LLMs understand instructions.

SBOMproof: Beyond Alleged SBOM Compliance for Supply Chain Security of Container Images

This paper explores how to improve supply chain security for container images using SBOMs (Software Bill of Materials). It's about making sure you know what's in your software and that it's secure.

Towards Carbon-Aware Container Orchestration: Predicting Workload Energy Consumption with Federated Learning

This research is about making container orchestration more carbon-aware. This involves predicting the energy consumption of workloads using Federated Learning. This is about making cloud computing more environmentally friendly.

Optimizing Container Loading and Unloading through Dual-Cycling and Dockyard Rehandle Reduction Using a Hybrid Genetic Algorithm

This paper looks at how to optimize the process of loading and unloading containers using a hybrid genetic algorithm. This means improving efficiency at ports and terminals.

A Benchmark Study of Deep Reinforcement Learning Algorithms for the Container Stowage Planning Problem

This research uses Deep Reinforcement Learning algorithms to tackle the container stowage planning problem. It aims to find the best way to arrange containers on ships for optimal efficiency.

gh0stEdit: Exploiting Layer-Based Access Vulnerability Within Docker Container Images

This paper focuses on the vulnerabilities in Docker container images. It explores how attackers can exploit the layer-based access to compromise the security of these containers.

Direct Token Optimization: A Self-contained Approach to Large Language Model Unlearning

This research focuses on the