Say hello to my little friends! I just unboxed this trio of HP Z2 G1a!
Three is always better than one! 3x AMD Ryzen AI Max+ Pro 395 384GB RAM 24TB of RAID storage Ubuntu 24.04 ROCm 7.0.2 llama cpp, vLLM and Aibrix
Small, cheap GPUs are about to become the Raspberry Pi of edge AI inference. Sprinkle some kubectl fairy dust on top, and suddenly it's a high-availability, self-healing, cloud-native, enterprise-grade AI cluster camping in a closet.
Make sure you own your AI. AI in the cloud is not aligned with you; itβs aligned with the company that owns it.
A new lightweight model to do machine translation from English to Ukrainian using recently published LFM2 model. Use demo Yehor/en-uk-translator to test it.
Facts: - Fine-tuned with 40M samples (filtered by quality metric) from ~53.5M for 1.4 epochs - 354M params - Requires 1 GB of RAM to run with bf16 - BLEU on FLORES-200: 27.24 - Tokens per second: 229.93 (bs=1), 1664.40 (bs=10), 8392.48 (bs=64) - License: lfm1.0
Happy to announce the release of Shisa V2, our latest generation of our bilingual Japanese-English language models. After hundreds of ablations and months of work, we're releasing some of the strongest open Japanese models at 7B, 8B, 12B, 14B, 32B and 70B! Full announcement here https://shisa.ai/posts/shisa-v2/ or visit the Shisa V2 HF collection: shisa-ai/shisa-v2-67fc98ecaf940ad6c49f5689