Home
Your GPUs are not lazy —Your Scheduler is: ACB makes GPUs work harder — not cost more

● Runtime-aware GPU allocation: Monitors AI framework to allocate GPUs needed ● Full Memory Access: Active program has access to the full GPU memory ● Advanced Scheduling: Employs techniques like backfill to optimize job placement and maximize aggregate utilization ● Fast Deployment: No code changes required in user programs
Unlock all the power your GPUs can offer, free, for 30 days

Do any of the following apply to you?
- Your workloads involve alternating phases of CPU and GPU tasks
- You have multiple GPU-capable jobs that can be queued or batched
- Your workloads can utilize full GPU memory intermittently
- You need to host multiple domain-specific LLMs with uneven demand on shared GPU infrastructure
|
|
|
|
Technology Overview
Target Industry/Users
AI Infrastructure teams, Enterprise ML and Data teams, Cloud/SaaS AI Service providers, AI Research labs and Startups (Biotech, Finance, Analytics), GPU Hosting Providers
Challenges in Target Industry and Operations
Soaring prices and shortages of GPUs, underutilization of current GPU infrastructure
Technical Challenges
Runtime inefficiencies, fragmented toolchains, hardware scaling constraints
Fujitsu's Technological Advantage
- Runtime aware GPU allocation
- GPU memory oversubscription management
- Integration without code changes in the user programs
- See technical white paper for a detailed introduction (linked at the bottom)
Use Cases
- LLM inference
- Installing multiple AI applications on a single inference server to provide diverse services
- AI Training and Inference
- Improving development throughput by enabling simultaneous execution of multiple learning processes during model parameter exploration
Case Studies
- AlphaFold2 Throughput Optimization
- ACB dynamically shares GPU resources between multiple AlphaFold2 jobs, maintaining throughput increasing GPU utilization by 45%
- Multi-LLM Hosting
- ACB manages multiple LLMs on a single GPU server, optimizing memory usage and enabling concurrent model serving without delays
- FX Risk Models
- Improved training throughput for financial risk models by multiplexing GPU usage, enhancing efficiency by over 2-fold
- Cloud Services
- Host several AI models on a single GPU node and manage five times the physical GPU memory capacity
- AI models for Retail
- Achieved 2× throughput in object recognition model training for IaaS applications
(see press release and technical white paper linked at the bottom for more details)
Technical Trial
- Join Private Beta: Get 30 days of full access. Test, explore, and shape what is next
- A Proof of Concept is possible
Related Information
- Press release October 22, 2024
- Demo Videos:
Documents (coming soon)
English vesion | Japanese version |
---|---|
ACB Walking Deck | ACB Walking Deck |
Technical White Paper | Technical White Paper |
User Manual | User Manual |