I've recorded my jots and experiences. Pick a card!
Behind the scenes of Amazon EBS innovation and operational excellence
KEYNOTE by AWS CEO guy Adam Selipsky
Advanced AWS CDK: Lessons learned from 4 years of use
Optimizing storage price and performance with Amazon S3
AWS storage: The backbone for your data-driven business
Compute innovation for any application, anywhere
Morning keynote: why AI is sweet and you shouldn’t think about anything else ever
Hyperscaling databases on Amazon Aurora
Innovation talk: Innovate faster with generative AI
Future-proofing your applications with AWS databases
Networking hang: Mike Sullivan, CTO/2nd-in-command of Landit in NYC
Keynote: Robert Vogel (CTO/VP of Amazon.com)
Innovation talk: emerging tech
Doing serverless event driven architecture stuffs
Innovation talk: “building without limits”
re:play
×
Behind the Scenes of Amazon EBS Innovation and Operational Excellence
Overview of EBS: EBS is more than just a storage disk; it's a fully managed, distributed system.
Global Data Transfer: EBS handles a massive 13 exabytes of data transfers every day worldwide.
Scaling Challenges: EBS presents scaling difficulties in terms of both size and performance, making it less ideal for databases.
Initial Limitations: Initially, EBS was available only in the US East (N. Virginia) region (us-east-1). EBS snapshots were stored in Amazon Simple Storage Service (S3), leading to suboptimal performance for small, random Input/Output (I/O) operations due to hard drive limitations.
Minimizing the Blast Radius: EBS stores data across multiple Availability Zones for enhanced reliability and disaster recovery.
Control Plane Functions: The control plane is responsible for managing and configuring data plane activities like creating, attaching, and deleting EBS volumes.
Data Plane Operations: The data plane deals with direct customer data handling, including reading and writing bytes and interfacing with Amazon Elastic Compute Cloud (EC2).
Physalia/Cells: These are smaller than an Availability Zone, featuring a small number of clients per data, simplified management, and reduced blast radius. Each cell functions as a 7-node Paxos cluster.
Nitro Card for EBS: Includes an NVMe (Non-Volatile Memory Express) controller and NVMe to remote storage protocol, optimized for EBS by default.
Io2 Volumes: Offers five nines (99.999%) durability, making them 2000 times more reliable than standard commodity disk drives.
Rethinking EBS Volumes: Volumes are split into shards, removing constraints imposed by individual storage nodes and enabling quicker replacements in case of failure.
Io2 Block Express Volumes: A new type of volume offering improved performance.
Eliminating Dependency Bottlenecks: Transition from TCP (Transmission Control Protocol) connections to SRD (Scalable Reliable Datagram) for multi-path network communication.
SRD Advantages: SRD facilitates multi-pathing and microsecond-level retries, operating on Nitro architecture. It results in 90% lower tail latency.
Io2 Block Express Availability: Now accessible on all EC2 instances equipped with Nitro, offering four times the max IOPS throughput and size compared to io1 volumes, and 100 times more durability.
Focus on Reducing Outlier Latency: Utilizing Deming’s Cycle (Plan, Do, Check, Act) to identify and address issues leading to high latency.
Focus on Reducing Outlier Latency: Utilizing Deming’s Cycle (Plan, Do, Check, Act) to identify and address issues leading to high latency. Part of this reflective approach includes the "5 Whys" analysis to understand and correct the root causes of problems:
Why is the outlier latency high? Because certain data requests are delayed.
Why are these requests delayed? Due to network congestion.
Why is there network congestion? Because multiple requests are routed through the same network path.
Why are requests routed through the same path? The current routing protocol doesn't support multi-pathing.
Why doesn’t the protocol support multi-pathing? Because it's an older standard not optimized for current cloud infrastructure.
Reflective Approach to Problem-Solving: Following a structured approach to identify the root causes of high outlier latency and address them effectively.
Resiliency Features in EBS: Includes EBS pause I/O action in Fault Injection Simulator (FIS), Attached EBS status check, instance lifecycle enhancements, and EBS stalled I/O check.
Key Takeaways: Emphasizing the importance of continuous improvement and resiliency in architectural planning and risk management.
×
Keynote by AWS CEO guy Adam Selipsky
Guy wears jeans, sneakers and a sport jacket over collared shirt with no tie, because he is “cool”
Data Centers, Services, and Features: AWS leads with 3x more data centers, 60% more services, and 40% more features than the next cloud provider.
Amazon S3 Express One Zone: Introduced as a new S3 class, allowing co-location of data and compute, managing millions of access requests per minute, and providing single-digit millisecond latency at 50% lower costs than S3 Standard.
Graviton4 Chip: Announced as AWS's most powerful and energy-efficient chip, surpassing Graviton3 with significant speed improvements for database and Java applications. Accompanied by the introduction of R8G instances for EC2/Graviton4.
Generative AI Stack: Focused on applications that leverage foundational models (FMs), providing tools to build with large language models (LLMs), and offering infrastructure for training and inference.
EFA (Elastic Fabric Adapter): Capable of scaling up to 20,000 GPUs in a single cluster, equivalent to supercomputer capabilities.
NVIDIA Collaboration: NVIDIA's CEO discusses the integration of NVIDIA processors/GPUs with AWS, introducing new models like L4 and H200. He wears leather jacket, black T shirt, jeans and also white sneakers, because he is “cool.”
Earth2 and Project Shiba: Earth2 is a digital twin of Earth, while Project Shiba connects 16,384 GPUs into a giant AI supercomputer, aimed at training next-generation LLMs.
Amazon EC2 Capacity Blocks for ML: The introduction of EC2 ultraClusters for machine learning workloads, offering the capability to reserve hundreds of GPUs. These are ideal for training and fine-tuning foundational models (FMs) for short-duration workloads.
AWS Trainium2 Chip: A specially designed chip for generative AI and ML training. It's optimized to train FMs with hundreds of billions to trillions of parameters, delivering training speeds 4x faster than previous AWS solutions and offering 65 exaflops of on-demand supercomputing performance.
AWS Neuron Support: AWS Neuron is now supporting popular machine learning frameworks such as TensorFlow and PyTorch, enhancing its capabilities in AI and ML applications.
Bedrock for Generative AI: Bedrock is introduced as a tool to build and scale generative AI applications using large language models (LLMs) and other FMs. It emphasizes that there is no single model that fits all use cases, highlighting the need for diverse and specific solutions.
Partnership with Anthrop\c: Announcement of a new partnership aimed at advancing the field of AI and ML.
AGENTS for Bedrock: A new feature for Bedrock that allows users to select FMs, provide basic instructions, specify Lambda functions, and choose relevant data sources for running tasks. AGENTS in Bedrock generate a sequence of steps and required information, including which APIs to call, and then execute the plan without the need for additional training of the FM. This feature is compliant with stringent standards like GDPR and HIPAA.
Machine Learning in Cancer Research: Raising a thought-provoking question, "What happens when machine learning uncovers cancer’s weakness?" indicating the potential impact of AI and ML in medical research and treatment.
Amazon Q: A new generative AI-powered assistant designed to be an expert resource for businesses. Amazon Q is capable of providing interactive answers, solving problems, generating content, and taking action. It understands company-specific information, coding practices, and system architectures. The assistant personalizes interactions based on user roles and permissions and is designed with security and privacy as priorities. Amazon Q is positioned as an invaluable tool for AWS, assisting across the AWS console, Integrated Development Environments (IDEs), and documentation, and guiding users in selecting the right tools for their applications and projects. Trained on 17 years of AWS knowledge, Amazon Q represents a significant leap in AI-powered business assistance.
Language Upgrades and Code Transformation: Addressing the common issue of outdated Java versions in projects, AWS introduces a solution for rapid and efficient language upgrades. This initiative, aimed at enhancing security posture and performance, accelerates the migration from older Java versions. The service notably upgraded 1,000 Java applications in just two days, showcasing its efficiency and potential to significantly reduce the time and resources typically required for such upgrades.
Amazon Q Code Transformation: This service facilitates comprehensive Java language upgrades in a significantly reduced timeframe, aiming to enhance security and performance. It accelerates migration from older Java versions, exemplified by the rapid upgrade of 1,000 Java applications in just two days, a feat that received considerable applause from the audience.
Diverse Database Offerings: AWS showcases its wide array of database types, including relational, key-value, document, graph, time series, ledger, wide column, and memory databases. This diversity caters to a broad range of data management needs and application scenarios.
Zero-ETL Integrations: AWS announced zero-ETL (Extract, Transform, Load) integrations for several of its database services, including Aurora, PostgreSQL, DynamoDB, and MySQL with Redshift, as well as Amazon OpenSearch. This advancement is aimed at simplifying data integration and management processes, enabling more efficient and seamless data handling and analytics.
Amazon DataZone: A new service that leverages Machine Learning to enhance data catalogs by adding metadata. Amazon DataZone is designed to enrich data understanding and utilization, thereby improving data-driven decision-making and insights.
CDK: Cloud Development Kit Insights and Best Practices
CloudFormation and CDK: CDK typically generates CloudFormation apps with multiple stacks. An alternative approach is using Projen, which is similar to CDK but generates files instead of CloudFormation templates.
Safety with Curl: Using curl as a safety command in CDK implementations.
AWS CodePipeline vs GitHub Actions: Comparison between AWS's native CI/CD pipeline and GitHub's Actions for pipeline management.
Single Stack Deployment:
Pros: No worries about cross-stack dependencies and offers one atomic update.
Cons: Can hit resource limits, lead to longer updates, create noisy stack differences, and encounter atomic block problems where a small issue in a large application can halt all changes.
Multi Stack Deployment:
Pros: Less likely to hit resource limits, allows faster updates, cleaner stack differences, and avoids atomic blocks.
Cons: Requires handling cross-stack dependencies.
Construct Layout: Constructs in CDK should model business functionality rather than just encapsulate resources. The approach favors more discrete constructs, and refactoring often involves changing logical IDs.
AWS Secrets Manager Secrets Construct: This construct includes a function called generateSecretString to enhance secret management.
Risks with Changes: Introducing changes in the system can easily break it, suggesting caution and thorough testing.
External Data Integration: Not recommended due to obfuscated inputs leading to nondeterminism and potential system failures.
Importance of Determinism: It's crucial to have code that produces the same outputs from the same inputs to ensure reliability and predictability.
Aspects in CDK: Useful features like Datadog enrollment, logical ID mapper, and tagging.
Stack Best Practices:
Handle stateful resources in their own stacks.
Always provide environmental context (account/region) to stacks.
Refactor early and often.
Conduct snapshot tests of stacks.
Avoid having more than seven constructs and refrain from reusing synthesized templates.
Construct Best Practices:
Constructs should represent business functions and be small and discrete.
Early and frequent refactoring is encouraged.
Use fine-grained assertions.
Avoid using imports in your constructs to maintain clarity and simplicity.
End-to-End Testing: Use triggers module or cdk-intrinsic-validator for comprehensive testing of your CDK applications.
×
Data Growth and Cost Optimization Strategies
Explosive Data Growth: Data is growing at an unprecedented pace, emphasizing the need for effective management and cost optimization strategies.
Pillars of Cost Optimization:
Define your workload requirements.
Understand your storage needs and develop insights.
Optimize and measure to ensure efficient use of resources.
Visibility is Key: "If you can’t see it, you can’t change it!" — Understanding and visualizing data is essential for effective management.
Cost Optimization Techniques: Identifying which data prefixes have become 'cold' or less frequently accessed to optimize storage costs.
S3 Storage Lens: A tool to discover infrequently accessed data, incomplete multipart uploads, and concurrent versions. Integration with Amazon CloudWatch offers expanded insights.
S3 Inventory Report: Provides object-level analysis, feeding into services like Amazon Athena for detailed data assessment.
Patterns of Cost Optimization: Differentiating between data with known/predictable access patterns and data with unknown/changing access patterns. Workloads with predictable patterns often exhibit low retrieval rates over time.
Amazon S3 Storage Classes:
S3 Standard: Ideal for frequently accessed data.
S3 Standard-IA (Infrequent Access): For data that is accessed less frequently.
Glacier Instant Retrieval: Suitable for rarely accessed data.
Glacier Flexible Retrieval: Designed for archived data, offering flexible retrieval options.
S3 Glacier Deep Archive: Best for long-term archiving of data.
Selecting the Right Storage Class: Considerations include the frequency of access, duration of storage, and performance requirements.
Addressing Unknown Data Patterns: The majority of data often has unknown access patterns, necessitating versatile storage solutions.
Characteristics of Storage Performance: Key performance metrics include request rate, request latency, and throughput.
Understanding 503 Slowdowns: A 503 slowdown in S3 usually signals a high volume of requests to your bucket, indicating the need for performance tuning and optimization.
Strategies for Controlling Latencies: Optimize your S3 usage by tuning timeouts to match your environment's needs. Implement retries for slow requests and parallelize them for efficiency. Additionally, employing multiple connections, such as through byte-range GETs and multipart uploads, can enhance throughput. Utilizing multi-value DNS allows opening multiple connections, and monitoring these connections for performance, closing underperforming ones, is crucial. Timely tuning of client timeouts to cancel slow requests and initiate retries plays a vital role in maintaining optimal performance.
What if Manual Tuning Isn't Your Thing? Enter AWS Common Runtime (CRT): This tool embodies S3's performance best practices, delivering high data transfer rates without the manual hassle. It automatically manages timeouts, retries, DNS load balancing, and request parallelization, simplifying the optimization process.
Mountpoint for S3: Designed for scenarios requiring repeated data access, Mountpoint for S3 dramatically optimizes performance. It caches data on EC2 instance storage, memory, or EBS volumes, thus reducing the cost of requests and response times for frequently accessed data.
Pro Tips for S3 Optimization:
Optimize your object size to balance performance and storage efficiency.
Horizontally scale your operations across multiple prefixes to distribute load and improve accessibility.
Embrace the AWS Common Runtime; it's your ally in seamless and effective S3 management.
Adopt an asynchronous mindset: thinking asynchronously can lead to more efficient designs and usage patterns, especially in high-load environments.
15 Years of EBS: A Deep Dive into Performance and Security
Remarkable Scale of EBS: Over 15 years, EBS has achieved a staggering 100 trillion daily I/O operations, with more than 390 million volumes created daily, and over 13 exabytes of data transferred by customers each day.
Nitro's Role in Data Security: Within the EBS infrastructure, the Nitro system encrypts data before it ever leaves the host, ensuring AWS never has access to customer data. This highlights AWS's commitment to data security and privacy.
Challenges in Storage Latency: Factors such as read request queuing, authentication, authorization, metadata lookups, and network latency all contribute to storage latency. The goal is to speed up these processes to improve overall performance.
Impact of Latency Reduction: Lowering latency is particularly beneficial for workloads that are interactive or have critical data dependencies. Faster data access can significantly enhance the performance and responsiveness of these applications.
S3 Express One Zone: A new advancement in S3, Express One Zone offers a single-availability zone architecture, enabling the co-location of storage and compute. This setup, combined with high-performance storage media and tailored optimizations, aims to reduce request latency. Additionally, Directory Buckets are introduced as a new bucket type for S3 Express, further enhancing storage options.
Amazon FSx Performance Boost: Amazon FSx has seen a significant performance increase, with a ninefold enhancement in read throughput (from 4 GB/s to 36 GB/s) and a sixfold increase in write throughput (from 1 GB/s to 6.6 GB/s). These improvements underscore AWS's ongoing efforts to enhance storage solutions for demanding workloads.
S3's Dominance in Data Lakes: S3 has become the go-to choice for building data lakes, with over 90 million data lakes currently hosted on S3. This widespread adoption highlights S3's reliability and scalability in handling large datasets.
The Concept of an Iceberg: Iceberg represents open table formats where objects are organized into tables. This forms the basis of the 'lakehouse' architecture, introducing new capabilities such as serializable isolation, enhanced performance, schema evolution, and the ability to perform time travel operations. A prime use case for this technology is the management and analysis of log data.
Open Table Catalog: The catalog layer, metadata layer, and data layer work in tandem to provide a comprehensive structure. The catalog layer (e.g., db.customer_iceberg) points to current metadata, the metadata layer consists of metadata files and lists, and the data layer includes data files and delete files, among others. This structure streamlines data organization and accessibility in data lakes.
Optimizing Transactional Data Lakes: AWS introduces automatic compaction of Apache Iceberg tables, specifically designed to optimize transactional data lakes. This feature improves data management efficiency and overall system performance.
Power of Data Lakes: Data lakes serve as a powerful organizational tool, offering great velocity and value to businesses. Their ability to manage vast amounts of diverse data effectively makes them an invaluable asset in the data-driven world.
S3 Connector for PyTorch: This connector provides PyTorch-specific data loading and checkpointing primitives, achieving a 40% faster machine learning data loading process. This enhancement reflects AWS's commitment to integrating and optimizing its services for popular ML frameworks.
Mountpoint for S3 and S3 Express One Zone: These features optimize repeated data access requests by caching data in various storage mediums like EC2 instance storage, memory, or EBS volumes. This leads to reduced request costs and faster response times for repeated data access, signifying a 6x increase in throughput.
×
Innovation Talk: Compute Innovation for Any Application, Anywhere
Live Migration Needs: Live migration is highlighted as an example of applications requiring substantial compute power.
AWS Nitro Enclaves: These enclaves provide additional isolation for data in use. EC2 instances, encompassing users, third-party libraries, apps, and operating systems, connect to Nitro enclaves through a secure local channel. This setup ensures encryption of data and plaintext processing, along with CPU and memory isolation.
EC2 U7i: A powerful instance with 32TB of DDR5 memory and 896 CPUs, showcasing the immense computational capabilities of AWS's infrastructure.
Lambda Snapstart: Aimed at improving cold starts for Java-based functions. This technology leverages serverless Firecracker technology, enabling functions to start 10 times faster without necessitating changes to existing code.
Graviton4 Advancements: Graviton4 boasts 50% more cores, double the L2 cache, and 75% more memory bandwidth. It also incorporates DRAM, Nitro cards, and a coherent link, underlining significant enhancements in processing power and efficiency.
World's Fastest AI Supercomputer: AWS and Nvidia collaborate to create the fastest AI supercomputer, hosted in the DGX cloud on AWS. Featuring EFA interconnect and Nitro, it's set to become available to customers in the DGX Cloud next year.
EC2 Capacity Blocks for ML: Users can reserve Amazon EC2 p5 instances, specifically designed for machine learning workloads, highlighting the emphasis on providing tailored solutions for ML applications.
AWS Console to Code: This feature enables a smoother transition from prototyping to production. It allows users to generate well-architected code using tools like CloudFormation, CDK, SDK, TerraForm, etc., streamlining the development process.
Compute Badge Introduction: The introduction of a new "COMPUTE" badge, underscoring AWS's focus and achievements in the realm of computational power and innovation.
×
Morning Keynote: Why AI is Sweet and You Shouldn’t Think About Anything Else Ever
Cheesy inspirational open video saying “THIS (a bunch of 1’s and 0’s) is data…but is THIS (showing videos of humans doing cool things) data!?? What will YOU create?”
Building AI Applications: Essential elements for AI app development include access to various foundational models, private environments for leveraging data, user-friendly tools for building and deploying applications, and specialized ML infrastructure.
AI Stack Components: The AI stack is comprised of GPUs, Trainium, Inferentia, SageMaker, UltraClusters, EFA (Elastic Fabric Adapter), EC2 capacity blocks, Nitro, and Neuron, showcasing AWS's extensive resources for AI development.
Applications Leveraging Foundational Models: Various applications such as Amazon Q, Q in Quicksight, Q in Amazon Connect, and Codewhisperer are highlighted for their use of foundational models in AI.
Tools for Building with FMs and LLMs: Bedrock is presented as a key tool, with features like guardians, agents, and customizations, for building applications with foundational models and large language models.
Model Diversity: The keynote emphasizes that no single model can address all use cases, underscoring the need for a diverse range of AI models.
Vector Embedding: This technique is highlighted for its role in enhancing information accuracy within AI systems.
Amazon Titan Image Generator: A tool capable of generating realistic, studio-quality images using natural language prompts, representing a significant advancement in image generation AI.
Invisible Watermarks: The concept of "invisible watermarks" is introduced as a measure to support responsible AI practices and safeguard intellectual property.
Model Updating through Continued Pre-training: The importance of updating AI models through ongoing pre-training is emphasized to ensure continued relevance and accuracy.
Fine-Tuning vs. Pre-Training: The distinction between fine-tuning (maximizing accuracy for specific tasks with a small amount of data) and pre-training (maintaining model accuracy for a particular domain with large datasets) is clarified, highlighting different approaches to model optimization.
RAG Implementation Challenges: Implementing Retrieval-Augmented Generation (RAG) is resource-intensive and time-consuming, involving the conversion of data into embeddings, storing these embeddings, and integrating with vector databases. To address this, Amazon Bedrock now includes knowledge bases for streamlined operations.
Distributed Training Complexities: The process of distributed training involves breaking down datasets into chunks, allocating them to a training cluster, and then building the model. This method is resource-intensive, with checkpoints causing delays and node failures leading to downtime.
Strong Data Foundation: The keynote emphasizes the importance of a data foundation that is comprehensive, integrated, and governed, underscoring the key principles for effective data management and utilization.
Storing Vectors and Data Together: The approach of storing vectors and data together is highlighted for its ability to use familiar tools, reduce licensing and management overhead, offer faster experiences to end-users, and minimize the need for data synchronization and movement.
Enabling Vector Search: AWS introduces vector search capabilities across various services, including Aurora PostgreSQL, Amazon RDS for PostgreSQL, Amazon OpenSearch, and Amazon OpenSearch Serverless, enhancing search functionalities in these databases.
Aurora Optimized Reads with pgvector: A significant 20x improvement in queries per second is achieved with pgvector in Aurora, demonstrating enhanced read performance.
Amazon MemoryDB for Redis: This service is introduced as a Redis-compatible, durable, in-memory database service designed for ultra-fast performance.
Vector Search Expansion: AWS extends vector search to additional databases like Amazon DocumentDB, Amazon DynamoDB, and Amazon Neptune, creating more robust and versatile search capabilities across its database offerings. However, MySQL is notably absent from this expansion.
Commitment to Zero-ETL Future: AWS reiterates its commitment to a zero-Extract, Transform, Load (ETL) future, showcasing integrations like MySQL directly into Amazon Redshift and Amazon OpenSearch directly into Amazon DynamoDB, simplifying data integration processes.
S3 Data Lake to OpenSearch Integration: Demonstrating the seamless integration of S3 Data Lakes into Amazon OpenSearch, enhancing data analytics and search functionalities.
AWS Clean Rooms: AWS introduces the concept of 'clean rooms' where users can collaborate with partners without sharing raw data, creating a secure and efficient environment for data analysis and collaboration.
Eliza – A Nostalgic Reference: The keynote humorously references Eliza, a mock Rogerian psychotherapist program created in 1966, drawing a contrast to the advanced AI capabilities available today.
Data Abundance Irony: Reflecting on the irony that despite the abundance of data available today, much of the effort in the data industry is still focused on just accessing and managing the necessary data effectively.
PartyRock – App Development Simplified: Introducing 'PartyRock', a tool that allows users to describe what they want their app to do, simplifying the app development process and making it more accessible.
×
Hyperscaling Databases on Amazon Aurora
Database-per-Service Approach: The session advocates for creating an Aurora cluster for each service, involving cloning the database plus replication, and then repeating this process for each service. This approach aims to shrink the monolithic database and grants DevOps teams full ownership of their respective service's database.
Sharding as a Scaling Solution: Sharding is presented as a solution to challenges in scaling beyond 128TB and for improving write operations. The process involves selecting a partition key to query a specific shard. However, multi-shard challenges include complexity in locating data across shards, maintaining consistency across clusters, and ensuring performance.
Shard Mapping Logic: The session discusses options for implementing shard mapping logic. This can be embedded within the application module/library, such as using active record sharding in Rails, or it can be managed through a sharing tier or a routing database.
Shard Key Importance: Emphasis is placed on using a shard key, not just an index. The application must know which shard to access before reaching the index level. A 'materialized global index' is suggested for effective sharding.
Aurora Limitless Database: Introducing Aurora's limitless database feature, which simplifies sharding. It offers horizontally scalable writes and reads, declarative table shading, integrated sharding logic and data movement, high availability, auto-scaling, and transactional consistency.
Shard Group Introduction: With the limitless database, a shard group becomes a part of the Aurora cluster, enhancing data management and scalability.
Index Optimization and Query Tuning: The session highlights the importance of optimizing indexes and refining queries to support database scalability.
Refactoring Data Model: It's recommended to continually refactor the data model to support scaling as the database grows.
Scaling Up and Out: The session concludes with strategies for upscale, including database per service, sharding, and leveraging Aurora's limitless database capabilities.
×
Innovate Faster with Generative AI
Key Metrics for AI Applications: The session emphasized the importance of metrics such as cost-effectiveness, low hallucination rates, low latency, completeness, and conciseness in AI applications.
Use Case - RyanAir: RyanAir's use of AI was highlighted, demonstrating its application in various areas:
Dynamic Pricing: Implemented through PageMaker, a tool for adjusting prices based on demand and other factors.
Paperless Cockpit: Using Extract, a tool to digitize and manage cockpit documents and information.
Predictive Maintenance: Leveraging SageMaker to predict maintenance needs and schedule timely interventions.
Schedule & Roster Building: AI tools are used to optimize staff scheduling and roster management.
In-App COVID Wallet: An application developed using AWS Lambda to manage COVID-related health data for travelers.
Refund Processing: Automating and streamlining the refund process for efficiency and customer satisfaction.
AWS Inferentia2: Introduction of AWS Inferentia2, described as a price accelerator, which is expected to enhance the cost-effectiveness and efficiency of AI applications.
Trainium: Trainium, an AWS tool, was also discussed as part of the suite of tools enabling faster and more efficient AI application development and deployment.
×
Investing in a Zero ETL Future
Challenges of Moving Data: The session discussed the complexities and costs associated with moving data around nodes, highlighting how changes in data structure can lead to a brittle system requiring extensive development, testing, and production cycles to manage.
Focus on Data, Not Plumbing: Emphasis was placed on the importance of focusing on the data itself rather than the infrastructure ("plumbing") required to manage it.
Amazon Datazone: Introduction of Amazon Datazone as a solution that liberates data, allowing customers to connect with datasets more efficiently and effectively.
Graviton in AWS Services: Announcement that AWS Graviton is now available across various AWS services, including RDS (Relational Database Service), Aurora, Elasticache, and MemoryDB for Redis, enhancing performance and efficiency.
New Features in AWS Services: The session introduced new features and updates:
GuardDuty with RDS Protection: AWS GuardDuty now includes protection for Amazon RDS, enhancing security for database services.
RDS Multi-AZ with Dual Standbys: RDS now supports Multi-AZ deployments with two standby instances, enabling faster and more reliable failovers. Minor version upgrades can be performed in under 30 seconds, significantly improving maintenance and update efficiency.
Aurora IO Optimization: Aurora has been enhanced for IO optimization, improving performance and efficiency.
RDS Custom for SQL Server BYOM: Introduction of a new feature for Amazon RDS custom that supports Bring Your Own Model (BYOM) for SQL Server, allowing more flexibility and customization.
Vector Database on Aurora and RDS for PostgreSQL: Announcement of pgvector, a vector database, now available on both Aurora and RDS for PostgreSQL, enhancing database capabilities for complex queries.
Aurora Optimized Reads: Aurora has been improved to optimize read operations, enhancing data retrieval efficiency.
DynamoDB Incremental Export to S3: DynamoDB now features incremental export functionality to Amazon S3, allowing for more efficient data backups and transfers.
Elasticache Performance Boost: Elasticache can now handle up to 1 million requests per second per node, significantly increasing its scalability and performance.
AWS Limitless Aurora PostgreSQL: This new service enables scaling to millions of writes per second and petabytes of data in a single database. It combines serverless architecture with fast scaling capabilities.
Elasticache Serverless: Introducing a serverless option for Elasticache, providing a highly available cache in under a minute without the need for infrastructure management. It scales instantly, offers 99.9999% high availability for Redis/Memcached, and operates on a pay-for-what-you-use model.
Conclusion: The session underscored the importance of data as a key differentiator for businesses and emphasized the notion that "Serverless is the new normal," highlighting the shift towards more efficient, scalable, and managed cloud services.
After watching the raunchy and wildly entertaining show "Absinthe," we grabbed sushi at the Bellagio, where Mike shared various intriguing stories and insights:
$3 Fire Sticks as Computer Replacements: Mike was impressed by the potential of $3 Fire Sticks to replace $700 computers, especially in synergy with AWS Workspaces. He highlighted the strategic move by AWS to develop Workspaces compatible with these affordable devices, foreseeing significant adoption in sectors like banking due to the convenience and cost-effectiveness of having a closet full of these Fire Sticks as backups. In his words, "Amazon played the long game, and now they're gonna make a fortune."
AI Image Model Training with 'Noise': Mike described an interesting AI technique where image models are trained by repeatedly adding and removing 'noise' from an image. This method was tested with a picture of an AWS employee. The model became adept at recreating the individual's image, but struggled when the AWS logo was included, showcasing both the capabilities and current limitations of AI in image recognition and generation. He also noted that general models tend to be slower due to being trained on excessive amounts of data.
Spear Phishing Awareness: Sharing an incident at his company, Mike mentioned how a new admin lady nearly fell victim to Nick Conrad's favorite gift card scam. In response, he conducted a spear phishing campaign that surprisingly ensnared 50% of the company's users. This served as a powerful reminder of the importance of continuous vigilance and training in cybersecurity.
Super savings, bro: Mike also provided a handy tip for saving money on the conference: he said get in touch with your AWS rep a month or two before the conference starts. They will "accidentally" find a discount pass for re:invent to give to you. He only had to pay $800!
×
Cost and Sustainability in System Design
Cost as a Proxy for Sustainability: Emphasizes that cost management is closely linked to sustainable practices in system design and operation.
Non-Functional Requirements: Highlights the importance of various non-functional aspects such as cost, sustainability, security, compliance, accessibility, performance, availability, scalability, and maintainability in system architecture.
Cost Consideration in Design: Advises to consider cost implications at every stage of the design process, underlining the role of cost as a critical non-functional requirement.
Aligning Cost with Business Objectives: Stresses the significance of aligning architectural costs with business goals and revenue models. The statement "Find the dimension you’re going to make money over, then make sure that the architecture follows the money" underscores this alignment.
Building Evolvable Architecture: The necessity of creating architectures that can evolve over time to meet changing business and technological needs.
Paying Off Technical Debt: Encourages addressing and resolving technical debt as part of aligning system costs with business objectives.
Architectural Trade-Offs: Acknowledges that architectural decision-making often involves balancing various trade-offs to achieve the optimal outcome for the business and system requirements.
Amazon's System Cost Metrics: Introduces innovative metrics such as cost per request, transitive cost per request, and conversion rate per request, offering a new lens to evaluate system efficiency and cost-effectiveness.
MyApplications on AWS Management Console: A groundbreaking new feature that empowers users to monitor and manage the cost, health, security posture, and performance of their applications, seamlessly integrating critical insights directly into the console.
Amazon CloudWatch Application Signals: A transformative new tool that automates instrumentation and operation of applications. This feature tracks application performance against key business objectives, offering an unparalleled view of how applications are delivering on business goals.
Cost-Aware Architectures: Emphasizes the importance of cost-awareness in system design, noting that unobserved systems can lead to unknown and uncontrolled costs. It advocates for implementing robust cost controls within architectural designs.
Eliminating Digital Waste: The session spotlights the need to eliminate inefficiencies and redundant processes in digital architectures, coining the term "eliminate digital waste" as a mantra for efficient system design.
Challenging Conventional Wisdom: The statement "The most dangerous phrase in the English language is: we’ve always done it this way" challenges traditional approaches and encourages innovative thinking in architectural design.
Resource for Frugal Architects: Directs attendees to visit thefrugalarchitect.com, a resource for architects seeking to design cost-effective and sustainable systems.
AI-Powered Decision Support: Highlights the role of AI in predicting trends and assisting professionals in making informed decisions, reinforcing the synergy between AI insights and human expertise.
Inspiration for Builders: Concludes with a stupid and corny rallying call to action: "It’s never been a better time to be a builder. NOW GO BUILD", delivered with a bombastic tone, allegedly inspiring attendees to embrace their role as innovative builders in the tech world.
×
Innovation Talk: Emerging Technology and the Digital Twin Concept
The Flywheel Concept: Presents an iterative cycle - real & synthetic data leads to software-defined solutions, creating digital twins, enabling simulation & testing, optimization with Machine Learning, pushing developments to production, and then connecting & collecting more data to feed back into the cycle. This concept accelerates innovation and efficiency in digital solutions.
Software-Defined Everything: Emphasizes the critical importance of having software-defined systems and processes. The talk posits that failing to adopt this approach could result in falling behind in the rapidly evolving technological landscape.
AWS IoT TwinMaker: Introduces the ability to create a full-fidelity digital twin of any real-world data model. This powerful tool allows for the replication and analysis of real-world systems in a virtual environment, enhancing understanding and enabling advanced simulations.
Contextualizing Data: Discusses the significance of contextualizing data to make it useful in the cloud. By defining the role and function of data, it can be effectively utilized for cloud-based applications and insights.
Operational Technology Applications: Details the ease of building and deploying IT applications for operational technology. Highlights the new product, AWS IoT Sidewise Edge, designed to organize and analyze data efficiently. Also introduces Mendix, a low-code environment that facilitates taking data to the shop floor, enhancing operational efficiency.
Quantum Computing and Qubits: Delves into the realm of quantum computing, explaining the concept of qubits that simultaneously represent 1’s and 0’s. This section covers phenomena like entanglement and superposition, fundamental to quantum computing, and illustrates concepts like bit flip and phase flip.
Accelerating the Flywheel: Concludes with the goal of making the flywheel spin ever faster, symbolizing the continuous and rapid innovation in the field of emerging technologies, driven by advancements in digital twins, IoT, and quantum computing.
×
Understanding Event-Driven Architecture
Basics of Event-Driven Architecture: EDA involves a flow from the producer to the broker and then to the consumer. The session emphasizes that adopting EDA is a gradual process, not an overnight transformation.
Events vs. Tables: Events in EDA narrate the story of what's happening, while tables (such as databases) reflect the current state of the system.
Benefits of EDA: One of the key benefits highlighted is loose coupling, which allows for more flexibility and scalability in system design and integration.
Challenges with Domain Language: A common problem in EDA is the confusion arising from unclear domain language. Different stakeholders, like developers, insurers, and underwriters, may have varying interpretations of key terms, leading to a 'big bowl of mud' situation.
Misunderstandings in Development: The session points out that often, it's the developers' misunderstanding, not the domain experts' knowledge, that gets released into production. This highlights the need for better communication and understanding between different teams.
Event Storming: This technique involves collaboration between domain experts and development teams to discover domains, identify system boundaries, recognize events, and share knowledge. It's applicable to both new and legacy projects.
Message Types: In EDA, messages are categorized as commands, which represent an intent or action and can be rejected, or events, which are immutable facts that have already occurred and can be ignored by consumers.
Messaging Patterns: The session covers various patterns in EDA, including point-to-point messaging, the publish/subscribe (pub/sub) pattern, and event streaming.
Asynchronous Point-to-Point Messaging: This involves using queues to manage messages, which helps decrease temporal coupling and allows the receiver to control the consumption rate. Consumers pull messages from the queue, allowing for scalability in consuming services.
Amazon SQS: Amazon Simple Queue Service (SQS) is highlighted as a fully managed message queue service that can scale almost infinitely. It features a simple, easy-to-use API and supports both standard and FIFO queue options, along with dead letter queue support.
Amazon EventBridge: EventBridge is presented as a fully managed event bus that enables publishing events to downstream consumers. It offers capabilities to filter and transform events, targets 28+ AWS services, and extends beyond AWS with API destinations and Software as a Service (SaaS) integrations.
Content Enricher Pattern: This pattern is used to enrich data for downstream consumers. It involves adding more information to an event before it reaches the consumer, with the aim of fulfilling the consumer's additional information requirements. This approach delegates contract management to the enrichment process.
Accidental Coupling Through Contracts: The discussion focuses on understanding bounded context mappings and how they can lead to accidental coupling in EDA. Three options are presented to tackle these challenges:
Conformist Approach: In this approach, one service (B) consumes messages as-is, conforming to the payload defined by another service (A). This leads to coupling through the payload structure.
Anti-Corruption Layer: This involves transforming data before consumption. An isolation layer for domains is created to decouple services through an interface, thus avoiding direct dependency on the payload structure of other services.
Open Host Service: This option suggests using a public or common language for communication and transforming data before publishing. It aims to hide implementation details and provide a unified interface for external services.
Private vs. Public Information in Events: The concept of private and public events is explored in the context of EDA. Private events are internal messages used within a private interface, including implementation details and internal contracts. Conversely, public events are external messages with a public interface, designed to avoid implementation details and create a public contract for external services.
Event-Driven Architecture and Domain-Driven Design: The keynote emphasizes that EDA is a collection of patterns that are effectively complemented by domain-driven design (DDD), a methodology that focuses on modeling software based on the business domain.
×
Exploring Amazon Bedrock AI and CodeWhisperer
Partyrock - Amazon Bedrock AI Playground: An invitation to use Partyrock, a playground for experimenting with Amazon Bedrock AI. It allows users to try different models, have fun, and test various scenarios in an AI environment.
CodeWhisperer for Individual Developers: Encouragement to try CodeWhisperer, especially since it's free for individual developers. CodeWhisperer offers capabilities such as scanning for vulnerabilities and suggesting open-source tools.
Internal Repository Integration with CodeWhisperer: CodeWhisperer now has the capability to generate code recommendations based on your internal repositories. This feature promises to provide better and more relevant code suggestions and helps onboard developers faster. It is emphasized that your content is never used to train the underlying models of CodeWhisperer.
Command Line Support in CodeWhisperer: CodeWhisperer now supports the command line, enabling developers to generate code recommendations directly from their CLI. This includes features like CLI completions, inline documentation, and AI-powered natural language to code translation.
Natural Language Command Interpretation: CodeWhisperer can interpret natural language commands like “copy all files in this directory to S3” and generate the corresponding command, demonstrating its advanced AI capabilities in understanding and translating user intents into actionable code.
Developer Time Allocation: Highlighting the distribution of developer time, with 73% spent on running and maintaining applications and only 27% on innovation and creating new stuff.
Amazon Q's Extensive Training: Amazon Q, an AI-powered assistant, is trained on 17 years of AWS knowledge. It assists users across various AWS interfaces like the console, IDE, and documentation. Q engages in conversations to help users explore new AWS capabilities, learn unfamiliar technologies, and find architectural solutions.
Integration of Q in IDE with CodeWhisperer: This integration enables developers to understand program logic in unfamiliar codebases, identify and fix bugs quickly, and generate functional tests, thereby enhancing the coding process.
Feature Development with Amazon Q: Amazon Q now includes a feature development capability that significantly reduces the time required to ship new features. It allows developers to go from a natural language prompt to an actionable feature implementation plan, focusing on end-to-end feature development. This capability is available directly in the IDE and through AWS CodeCatalyst.
Troubleshooting with Q: Amazon Q's troubleshooting capability enables developers to diagnose and troubleshoot errors much more rapidly, streamlining the error resolution process and improving development efficiency.
Q Code Transformation: Amazon Q's code transformation feature enables rapid upgrades and transformations of applications. It facilitates complete language upgrades in a fraction of the time, enhances security posture and performance, and will soon accelerate migration from Windows to Linux.
Optimizing Applications with Q: In a scenario where you're new to a Python application integrating with DynamoDB, Amazon Q can provide an overview of the app, identify endpoints, and offer optimization suggestions, even if you're unfamiliar with the codebase.
CodeCatalyst with Custom Blueprints: AWS CodeCatalyst now includes custom blueprints to codify best practices across organizations, allowing for more standardized and efficient development processes.
Q in CodeCatalyst: Integration with CodeCatalyst enables Amazon Q to summarize source repositories, develop coding approaches, author code, create pull requests, and monitor workflows while resolving errors, streamlining the entire development lifecycle.
Introducing AWS Amplify: AWS Amplify is a scalable front-end framework backed by AWS services. It allows developers to focus on application code rather than infrastructure. With every Git push, it deploys the front end and backend globally and scales serverlessly with native AWS services, offering a robust solution for modern application development.
×
×
×
×
×
The final party
Endless electronic music. So many beats. Lights that could give seizures seizures. Ultimately a failure because last call for drinks was at ELEVEN PM. U guyz are 4 real??? I left in abhorrence. Not because the party wasn't LIT – it was a neon dream in binary – but because when you say 'party till you drop,' I didn't know you meant 'drop by eleven.' Next year, AWS, let's sync our party clocks a bit better, MAN.