Metadata-Version: 2.4
Name: jang
Version: 1.0.0
Summary: JANG — Adaptive Mixed-Precision Quantization for Apple Silicon. The GGUF equivalent for MLX.
Author-email: Jinho Jang <eric@jangq.ai>
License: Apache-2.0
Project-URL: Homepage, https://jangq.ai
Project-URL: Repository, https://github.com/jjang-ai/jangq
Project-URL: Documentation, https://github.com/jjang-ai/jangq#readme
Project-URL: Bug Tracker, https://github.com/jjang-ai/jangq/issues
Project-URL: HuggingFace, https://huggingface.co/JANGQ-AI
Keywords: quantization,llm,apple-silicon,metal,mlx,jang,moe,mixed-precision
Classifier: Development Status :: 4 - Beta
Classifier: Intended Audience :: Science/Research
Classifier: Intended Audience :: Developers
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.11
Classifier: Programming Language :: Python :: 3.12
Classifier: Programming Language :: Python :: 3.13
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Classifier: Operating System :: MacOS
Requires-Python: >=3.11
Description-Content-Type: text/markdown
Requires-Dist: safetensors>=0.4
Requires-Dist: numpy>=1.24
Requires-Dist: tqdm>=4.60
Requires-Dist: huggingface_hub>=0.20
Provides-Extra: mlx
Requires-Dist: mlx>=0.22; extra == "mlx"
Requires-Dist: mlx-lm>=0.20; extra == "mlx"
Provides-Extra: torch
Requires-Dist: torch>=2.0; extra == "torch"
Requires-Dist: transformers>=4.40; extra == "torch"
Provides-Extra: all
Requires-Dist: jang[mlx]; extra == "all"
Requires-Dist: jang[torch]; extra == "all"
