A comprehensive initiative to democratize AI development through two complementary projects: a high-performance programming language and an AI-optimized operating system. This project aims to make advanced AI capabilities accessible to developers worldwide by providing both the tools to write AI applications and the platform to run them efficiently.
A modern programming language designed specifically for AI development, featuring:
- Native code compilation with LLVM for C/C++-level performance
- Built-in tensor operations and neural network primitives
- First-class CUDA support for GPU acceleration
- Automatic differentiation
- Python-like syntax with powerful type inference
A microkernel-based operating system optimized for AI workloads, offering:
- AI-optimized architecture for efficient tensor operations
- Hardware acceleration support (GPU, NPU)
- Memory-safe implementation in Rust
- Native AArch64 and Apple Silicon support
- AI-aware scheduling and resource management
The combination of these projects creates a powerful ecosystem for AI development:
- End-to-End Optimization: The language's compiled code runs directly on an OS optimized for AI workloads
- Hardware Acceleration: Both projects support GPU/NPU acceleration, working together seamlessly
- Memory Safety: Rust-based implementation across both projects ensures security and reliability
- Developer Experience: Clean syntax and robust tooling, backed by OS-level AI optimizations
- Rust (nightly)
- LLVM
- CUDA Toolkit (optional, for GPU support)
- QEMU (for OS development)
- ARM development tools
# Clone the repository with submodules
git clone
cd
# Set up development environment
./Singularity-Research1/Domocratizing-AI-OS/scripts/setup.sh
# Build the language
cd Singularity-Research1/Democratizing-AI-Lang
cargo build --release
# Build the OS
cd ../Domocratizing-AI-OS
make
// Example combining language features with OS capabilities
use democratising::nn::{Model, Dense};
use democratising::tensor::Tensor;
fn main() -> Result<()> {
// Create and train a model using language features
let mut model = Model::new();
model.add_layer(Dense::new(784, 256));
// Leverage OS-level optimizations automatically
// The OS handles efficient memory allocation and hardware acceleration
let x = Tensor::randn(&[100, 784])?;
let y = Tensor::randn(&[100, 10])?;
// Training automatically utilizes available hardware accelerators
model.fit(&x, &y, epochs=10, batch_size=32)?;
Ok(())
}
We welcome contributions to both projects! Please see:
- Basic language and OS implementation
- Tensor operations and neural network primitives
- CUDA support and hardware acceleration
- AI-optimized memory management
- Distributed training capabilities
- Enhanced hardware support
- Advanced AI model optimization
- Improved developer tooling
- Extended platform support
Both projects are licensed under the MIT License - see the LICENSE file for details.
If you use these projects in your research, please cite:
@software{democratizing_ai_2025,
author = {Allahyar, Sahibzada},
title = {Democratizing AI: Language and Operating System},
year = {2025},
institution = {University of Cambridge},
publisher = {GitHub},
url =
}
- The Rust community for their excellent tools and documentation
- The LLVM team for their compiler infrastructure
- The OS development community
- All contributors who have helped with both projects