Lamini, in collaboration with AMD's ROCm open software ecosystem, has unveiled the LLM Superstation, a finely tuned supercomputer with 128 AMD Instinct GPUs. This allows customers to train large LLMs, including Meta AI's Llama 2, making their AI models proprietary.
Lamini incorporates sophisticated optimisations such as PEFT (LoRA), RLHF, and toolformer, enabling data isolation across 4,266x models on a single server, accelerating model switching, compressing models, and integrating LLMs with enterprise APIs seamlessly.
The platform facilitates hosting 200 billion parameter models on a singular server, handling 10,000 fine-tuned language models, managing 12,800 concurrent requests, and processing over 3.5 million queries per day on a single node.
Lamini's solution also offers a cost advantage, claiming to be 10 times less expensive than AWS, without the lengthy wait time associated with NVIDIA H100s. AMD's Instinct MI250X, used as the foundation for Lamini, reportedly runs larger models than NVIDIA's A100s.
Lamini has been deployed in AMD's internal Kubernetes cluster, emphasising the company's commitment to creating models trained on AMD's code base across various components for specific developer tasks.
Read the full story here.
Busting Benchmark Myths
There’s a stiff competition among companies to excel on different benchmarks and especially, the MMLU (massive multitask language understanding), which is designed to measure knowledge by evaluating models exclusively in zero-shot and few-shot settings. However, this approach has been massively criticised as the marketing gimmick to showcase performance on the paper. However, the real evaluation of the model only happens when it goes in the hands of the people and its real-world use cases are discovered.
That's where Hugging Face’s Open LLM leaderboard comes into picture which gives a complete picture listing different models and their compute costs. In an exclusive interview with AIM, Philipp Schmid, a Technical Lead at Hugging Face, busts many myths about LLM benchmarks and gives new insights.
Read the full story here.
Apple’s Health Game
Apple recently rolled out iOS 17.2, enabling spatial video recording for iPhone 15 Pro and iPhone 15 Pro Max users. The upcoming Apple Vision Pro, expected in early 2024, will bring these videos to life. Reports suggest Apple's 2024 focus will shift heavily to wearables, with minor iPhone upgrades. The healthcare sector takes centre stage, featuring in the launch of advanced AirPods and watches.
The latter will include enhanced health detection capabilities, covering issues like hypertension and apnea. AirPods may incorporate brain wave-detecting sensors for measuring various physiological parameters, representing Apple's deep dive into health-related technology advancements.
Read the full story here.
A Celestial Showdown
In the competitive realm of satellite internet, Amazon's Project Kuiper aims to challenge SpaceX's Starlink. Jeff Bezos expects multiple winners in space, which is like the diverse success of internet entities. Recently, Amazon achieved a breakthrough in Kuiper, testing optical mesh networks in low-Earth orbit, achieving 100 Gbps links between prototype satellites using infrared lasers.
Kuiper's optical inter-satellite links (OISL) enable simultaneous connections with multiple spacecraft, forming a mesh network that moves data 30% faster than terrestrial fibre optic cables. Despite progress, Kuiper trails behind Starlink, which boasts over 2 million customers and significant government contracts.
Read the full story here.
Комментариев нет:
Отправить комментарий
Примечание. Отправлять комментарии могут только участники этого блога.