Blog

Your blog category

Frugality meets Accuracy: Cost-efficient training of GPT NeoX and Pythia models with AWS Trainium

Large language models (or LLMs) have become a topic of daily conversations. Their quick adoption is evident by the amount of time required to reach a 100 million users, which has gone from “4.5yrs by facebook” to an all-time low of mere “2 months by ChatGPT.” A generative pre-trained transformer (GPT) uses causal autoregressive updates […]

Frugality meets Accuracy: Cost-efficient training of GPT NeoX and Pythia models with AWS Trainium Read More »

Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models – The Berkeley Artificial Intelligence Research Blog

TL;DR: Text Prompt -> LLM -> Intermediate Representation (such as an image layout) -> Stable Diffusion -> Image. Recent advancements in text-to-image generation with diffusion models have yielded remarkable results synthesizing highly realistic and diverse images. However, despite their impressive capabilities, diffusion models, such as Stable Diffusion, often struggle to accurately follow the prompts when

Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models – The Berkeley Artificial Intelligence Research Blog Read More »

Create a web UI to interact with LLMs using Amazon SageMaker JumpStart

The launch of ChatGPT and rise in popularity of generative AI have captured the imagination of customers who are curious about how they can use this technology to create new products and services on AWS, such as enterprise chatbots, which are more conversational. This post shows you how you can create a web UI, which

Create a web UI to interact with LLMs using Amazon SageMaker JumpStart Read More »

Acer’s Predator Triton Neo 16 gaming laptop has Intel Core Ultra processors

Acer just unveiled the Predator Triton Neo 16 gaming laptop which features Intel’s long-awaited Core Ultra processors, formerly referred to internally as Meteor Lake. These CPUs boast a dedicated Neural Processing Unit (NPU) for advanced AI performance. The chips prioritize efficiency and speed, with Intel promising that graphics will be twice as fast when it

Acer’s Predator Triton Neo 16 gaming laptop has Intel Core Ultra processors Read More »

Generating 3D Molecular Conformers via Equivariant Coarse-Graining and Aggregated Attention – The Berkeley Artificial Intelligence Research Blog

Figure 1: CoarsenConf architecture. Molecular conformer generation is a fundamental task in computational chemistry. The objective is to predict stable low-energy 3D molecular structures, known as conformers, given the 2D molecule. Accurate molecular conformations are crucial for various applications that depend on precise spatial and geometric qualities, including drug discovery and protein docking. We introduce

Generating 3D Molecular Conformers via Equivariant Coarse-Graining and Aggregated Attention – The Berkeley Artificial Intelligence Research Blog Read More »

Build an end-to-end MLOps pipeline using Amazon SageMaker Pipelines, GitHub, and GitHub Actions

Machine learning (ML) models do not operate in isolation. To deliver value, they must integrate into existing production systems and infrastructure, which necessitates considering the entire ML lifecycle during design and development. ML operations, known as MLOps, focus on streamlining, automating, and monitoring ML models throughout their lifecycle. Building a robust MLOps pipeline demands cross-functional

Build an end-to-end MLOps pipeline using Amazon SageMaker Pipelines, GitHub, and GitHub Actions Read More »

Intel unveils Core Ultra, its first chips with NPUs for AI work

Intel today is entering the “AI PC” era with the launch of its new Core Ultra notebook chips. Originally codenamed “Meteor Lake,” these are Intel’s first processors to include an NPU, or neural processing unit, for accelerating AI tasks. The launch comes a week after AMD revealed its upcoming Ryzen 8040 hardware, its second batch

Intel unveils Core Ultra, its first chips with NPUs for AI work Read More »

On the Stepwise Nature of Self-Supervised Learning – The Berkeley Artificial Intelligence Research Blog

Figure 1: stepwise behavior in self-supervised learning. When training common SSL algorithms, we find that the loss descends in a stepwise fashion (top left) and the learned embeddings iteratively increase in dimensionality (bottom left). Direct visualization of embeddings (right; top three PCA directions shown) confirms that embeddings are initially collapsed to a point, which then

On the Stepwise Nature of Self-Supervised Learning – The Berkeley Artificial Intelligence Research Blog Read More »

Automate PDF pre-labeling for Amazon Comprehend

Amazon Comprehend is a natural-language processing (NLP) service that provides pre-trained and custom APIs to derive insights from textual data. Amazon Comprehend customers can train custom named entity recognition (NER) models to extract entities of interest, such as location, person name, and date, that are unique to their business. To train a custom model, you

Automate PDF pre-labeling for Amazon Comprehend Read More »

Scroll to Top