EGMO Token Airdrops are coming

Model Support

EdgeMob is designed to provide developers with maximum flexibility in how they run and integrate AI models. The platform supports both open-source and custom models, ensuring that developers can experiment, deploy, and scale AI workloads without depending on closed ecosystems.

Open-Source Model Integration

  • EdgeMob natively supports popular open-source models such as LLaMA, Mistral, Qwen, and CodeLLaMA, along with smaller models optimized for on-device execution.

  • Developers can load these models directly into the EdgeMob app using built-in import tools or via the model registry.

  • Continuous updates to EdgeMob’s runtime ensure compatibility with new model formats and community-driven improvements.

Custom Model Integration

  • Developers are not limited to pre-approved models. They can package and load their own custom models into EdgeMob.

  • The platform provides SDKs and documentation for preparing model weights, quantization, and deployment pipelines.

  • Custom models can be exposed locally (for testing) or through the API Gateway (for broader application integration).

Extensible Runtime

  • The EdgeMob orchestrator, built in Rust and C++, integrates with frameworks like Candle and llama.cpp, enabling efficient inference across multiple architectures.

  • This extensibility ensures that future frameworks or model families can be adopted with minimal friction.

Model Registry & Marketplace

  • EdgeMob includes a model registry where developers can publish, share, and access models.

  • In future iterations, a marketplace will allow developers to monetize models, while node operators can choose which models to host based on demand.


By supporting both open-source and custom models, EdgeMob lowers barriers to innovation and empowers developers to build AI applications that are cost-efficient, privacy-preserving, and tailored to their unique needs.

Last updated