/claim #82 /fixes #82
This PR implements an offline DeepSeek model loader for inference as requested in the feature request.
Created a modular architecture with separate components:
DeepSeekLoader: Core loading functionality with memory optimizationDeepSeekTokenizer: Text encoding/decodingDeepSeekWrapper: High-level interface following project patternsImplemented memory optimization techniques:
Added dynamic model discovery:
Created comprehensive tests:
Added example usage for easy integration
DeepSeekLoader: Handles model loading with memory optimizationDeepSeekTokenizer: Handles tokenization for input/outputDeepSeekWrapper: Provides a unified interfaceThis implementation avoids high-level libraries like transformers as requested.
Kunal Darekar
@Kunal-Darekar
IntelliNode
@intelligentnode