Market report February 2024
Trends on our radar
This month has been a continuation of AI and an upwards trend on FinOps and CyberSecurity. On top of those domains, there has been smaller, but very notable trends on the growth of Software as a Service (SaaS) or MicroSaaS’ have rebound and created significant growth for serverless services, as a means to avoid startup and initial running costs. Likewise, AI has been changing how we work and how we design systems to enrich workflows and create more naturally interactive apps.
Serverless computing: Redefining efficiency and scalability
AWS’s low latency runtime (LLRT)
Amazon Web Services (AWS) introduced the Low Latency Runtime (LLRT), a game-changer for serverless applications, promising over 10x faster startup times and up to 2x lower operational costs compared to conventional JavaScript runtimes on AWS Lambda. Engineered in Rust and utilizing the QuickJS JavaScript engine, LLRT is specifically optimized for the unique demands of serverless environments, eschewing a JIT compiler to reduce complexity and resource usage. This strategic optimization not only streamlines the runtime architecture but also significantly enhances performance, setting a new standard for serverless computing efficiency.
The FLAME pattern: Simplifying scalability
The FLAME pattern represents a novel approach to serverless computing, enabling effortless scaling by encapsulating any existing application code within a function. This innovation promises to simplify serverless application development and deployment, offering a more versatile and accessible computing model.
AI and machine learning: The dawn of intuitive programming
Nvidia’s vision: Beyond coding to system engineering
Jensen Huang, Nvidia’s CEO, challenges the traditional coding education paradigm, advocating for a shift towards system engineering. In an era dominated by AI advancements, Huang’s perspective underscores the importance of designing systems over writing code, suggesting that AI technologies like GPT models could democratize programming through natural language prompts.
OpenAI’s Sora: A new frontier in video generation
OpenAI introduces Sora, a text-to-video model capable of producing up to one-minute videos from textual prompts. Leveraging a diffusion model and transformer architecture, Sora epitomizes the fusion of GPT and DALL·E methodologies, offering unparalleled visual content creation capabilities. This innovation not only enhances creative possibilities but also signifies a major leap in AI-generated visual media.
MobileLLM and TestGen-LLM: Meta’s forward-thinking AI applications
Meta’s MobileLLM and TestGen-LLM illustrate the company’s commitment to optimizing AI for specific uses. MobileLLM’s compact yet powerful model showcases the feasibility of advanced AI capabilities on mobile devices, while TestGen-LLM’s application in automated testing for Meta’s platforms highlights AI’s role in enhancing software quality and development efficiency.
The era of 1-bit LLMs: A leap towards efficiency
A groundbreaking method has been proposed, utilizing 1.58 bits per parameter in large language models (LLMs), a significant reduction from the traditional 16 or more bits. This innovative approach, which assigns parameters values of {-1, 0, 1}, matches the performance of models up to 3B parameters, heralding a new era of efficiency in AI. While no code or models have been released yet, this advancement could dramatically reduce the computational cost and energy consumption of running LLMs, making powerful AI tools more accessible and sustainable.
FinOps: Approach to economic pressure
The FinOps domain is experiencing a strategic shift towards waste reduction and the management of commitment-based discounts in response to global economic pressures. This change emphasizes the importance of optimizing cloud expenditures, forecasting cloud spend, and assessing the impact of AI/ML costs. Additionally, an increase in collaboration between FinOps and sustainability teams is anticipated, highlighting a comprehensive approach to financial and environmental stewardship in cloud computing.
Cybersecurity: A call for memory safety
The White House Office of the National Cyber Director’s report on cybersecurity underscores the critical need for memory-safe programming languages. By advocating for the adoption of languages like C#, Go, Java, Python, Rust, and Swift, the report highlights the importance of mitigating vulnerabilities inherent in non-memory-safe languages like C and C++.
References
- Whitehouse press release technical report
- Key priorities shift in 2024
- MobileLLM: Optimizing sub-billion parameter language models for on-device use cases
- The era of 1-bit LLMs: All large language models are in 1.58 bits
- Automated unit test improvement using large language models at Meta
- Testgen LLM AI writes testing case Meta
- Rethinking serverless with flame
- AWSLabs
- Nvidia CEO: Don’t learn to code
- Sora