Skip to content

AWS Inferentia2 builds on AWS Inferentia1 by delivering 4x higher throughput and 10x lower latency Samir Araujo AWS Machine Learning Blog

  • by

​ The size of the machine learning (ML) models––large language models (LLMs) and foundation models (FMs)––is growing fast year-over-year, and these models need faster and more powerful accelerators, especially for generative AI. AWS Inferentia2 was designed from the ground up to deliver higher performance while… Read More »AWS Inferentia2 builds on AWS Inferentia1 by delivering 4x higher throughput and 10x lower latency Samir Araujo AWS Machine Learning Blog

Deploy Falcon-40B with large model inference DLCs on Amazon SageMaker James Park AWS Machine Learning Blog

  • by

​ Last week, Technology Innovation Institute (TII) launched TII Falcon LLM, an open-source foundational large language model (LLM). Trained on 1 trillion tokens with Amazon SageMaker, Falcon boasts top-notch performance (#1 on the Hugging Face leaderboard at time of writing) while being comparatively lightweight and… Read More »Deploy Falcon-40B with large model inference DLCs on Amazon SageMaker James Park AWS Machine Learning Blog

Rendered.ai Integrates NVIDIA Omniverse for Synthetic Data Generation Katja Reitemeyer – Archives Page 1 | NVIDIA Blog

  • by

​ Rendered.ai is easing AI training for developers, data scientists and others with its platform-as-a-service for synthetic data generation, or SDG. Training computer vision AI models requires massive, high-quality, diverse and unbiased datasets. These can be challenging and costly to obtain, especially with increasing demands… Read More »Rendered.ai Integrates NVIDIA Omniverse for Synthetic Data Generation Katja Reitemeyer – Archives Page 1 | NVIDIA Blog

Microsoft AI Introduces Orca: A 13-Billion Parameter Model that Learns to Imitate the Reasoning Process of LFMs (Large Foundation Models) Niharika Singh Artificial Intelligence Category – MarkTechPost

  • by

​ The remarkable zero-shot learning capabilities demonstrated by large foundation models (LFMs) like ChatGPT and GPT-4 have sparked a question: Can these models autonomously supervise their behavior or other models with minimal human intervention? To explore this, a team of Microsoft researchers introduces Orca, a… Read More »Microsoft AI Introduces Orca: A 13-Billion Parameter Model that Learns to Imitate the Reasoning Process of LFMs (Large Foundation Models) Niharika Singh Artificial Intelligence Category – MarkTechPost

Researchers from Princeton Introduce MeZO: A Memory-Efficient Zeroth-Order Optimizer that can Fine-Tune Large Language Models (LLMs) Tanya Malhotra Artificial Intelligence Category – MarkTechPost

  • by

​ Large Language Models are rapidly advancing with the huge success of Generative Artificial Intelligence in the past few months. These models are contributing to some remarkable economic and societal transformations, the best example of which is the well-known ChatGPT developed by OpenAI, which has… Read More »Researchers from Princeton Introduce MeZO: A Memory-Efficient Zeroth-Order Optimizer that can Fine-Tune Large Language Models (LLMs) Tanya Malhotra Artificial Intelligence Category – MarkTechPost

Build custom chatbot applications using OpenChatkit models on Amazon SageMaker Vikram Elango AWS Machine Learning Blog

  • by

​ Open-source large language models (LLMs) have become popular, allowing researchers, developers, and organizations to access these models to foster innovation and experimentation. This encourages collaboration from the open-source community to contribute to developments and improvement of LLMs. Open-source LLMs provide transparency to the model… Read More »Build custom chatbot applications using OpenChatkit models on Amazon SageMaker Vikram Elango AWS Machine Learning Blog

Superhuman Performance on the Atari 100K Benchmark: The Power of BBF – A New Value-Based RL Agent from Google DeepMind, Mila, and Universite de Montreal Niharika Singh Artificial Intelligence Category – MarkTechPost

  • by

​ Deep reinforcement learning (RL) has emerged as a powerful machine learning algorithm for tackling complex decision-making tasks. To overcome the challenge of achieving human-level sample efficiency in deep RL training, a team of researchers from Google DeepMind, Mila, and Universite de Montreal has introduced… Read More »Superhuman Performance on the Atari 100K Benchmark: The Power of BBF – A New Value-Based RL Agent from Google DeepMind, Mila, and Universite de Montreal Niharika Singh Artificial Intelligence Category – MarkTechPost