We collect cookies to analyze our website traffic and performance; we never collect any personal data. Cookie Policy
Accept
NEW YORK DAWN™NEW YORK DAWN™NEW YORK DAWN™
Notification Show More
Font ResizerAa
  • Home
  • Trending
  • New York
  • World
  • Politics
  • Business
    • Business
    • Economy
    • Real Estate
  • Crypto & NFTs
  • Tech
  • Lifestyle
    • Lifestyle
    • Food
    • Travel
    • Fashion
    • Art
  • Health
  • Sports
  • Entertainment
Reading: ​​IBM needs to be the enterprise LLM king with its new open-source Granite 3.1 fashions
Share
Font ResizerAa
NEW YORK DAWN™NEW YORK DAWN™
Search
  • Home
  • Trending
  • New York
  • World
  • Politics
  • Business
    • Business
    • Economy
    • Real Estate
  • Crypto & NFTs
  • Tech
  • Lifestyle
    • Lifestyle
    • Food
    • Travel
    • Fashion
    • Art
  • Health
  • Sports
  • Entertainment
Follow US
NEW YORK DAWN™ > Blog > Technology > ​​IBM needs to be the enterprise LLM king with its new open-source Granite 3.1 fashions
​​IBM needs to be the enterprise LLM king with its new open-source Granite 3.1 fashions
Technology

​​IBM needs to be the enterprise LLM king with its new open-source Granite 3.1 fashions

Last updated: December 19, 2024 11:38 am
Editorial Board Published December 19, 2024
Share
SHARE

IBM is staking its declare on the high of the open-source AI leaderboard with its new Granite 3.1 collection out in the present day.

The Granite 3.1 giant language fashions (LLMs) supply enterprise customers prolonged context size of 128K tokens, new embedding fashions, built-in hallucination detection and improved efficiency. In line with IBM, the brand new Granite 8B Instruct mannequin tops open-source rivals of the identical measurement together with Meta Llama 3.1, Qwen 2.5 and Google Gemma 2. IBM ranked its fashions throughout a collection of educational benchmarks included within the OpenLLM Leaderboard. 

The brand new fashions are a part of the accelerated launch cadence of IBM’s Granite open-source fashions. Granite 3.0 was simply launched in October. On the time, IBM claimed that it has a $2 billion guide of enterprise associated to generative AI. With the Granite 3.1 replace, IBM is specializing in packing extra functionality into smaller fashions. The fundamental concept is that smaller fashions are simpler for enterprises to run and are extra cost-efficient to function.

“We’ve also just boosted all the numbers — all the performance of pretty much everything across the board has improved,” David Cox, VP for AI fashions at IBM Analysis, advised VentureBeat. “We use Granite for many different use cases, we use it internally at IBM for our products, we use it for consulting, we make it available to our customers and we release it as open source, so we have to be kind of good at everything.”

Why efficiency and smaller fashions matter for enterprise AI

There are any variety of methods an enterprise can consider the efficiency of an LLM with benchmarks.

The path that IBM is taking is to run fashions by way of a gamut of educational and real-world assessments. Cox emphasised that IBM examined and skilled its fashions to be optimized for enterprise use circumstances. Efficiency isn’t nearly some summary measure of pace, both; fairly, it’s a considerably extra nuanced measure of effectivity.

One facet of effectivity that IBM is aiming to push ahead helps customers spend much less time to get desired outcomes.

“You should spend less time fiddling with prompts,” mentioned Cox. “So, the stronger a model is in an area, the less time you have to spend engineering prompts.”

Effectivity can be about mannequin measurement. The bigger a mannequin, the extra compute and GPU assets it sometimes requires, which additionally means extra value.

“When people are doing minimum viable prototype kind of work, they often jump to very large models, so you might go to a 70 billion parameter model or a 405 billion parameter model to build your prototype,” mentioned Cox. “But the reality is that many of those are not economical, so the other thing we’ve been trying to do is drive as much capacity as possible into the smallest package possible.”

Context issues for enterprise agentic AI

Other than the promise of improved efficiency and effectivity, IBM has dramatically expanded Granite’s context size.

With the preliminary Granite 3.0 launch, the context size was restricted to 4k. In Granite 3.1, IBM has prolonged  that to 128k, permitting for the processing of for much longer paperwork. The prolonged context is a major improve for enterprise AI customers, each for retrieval-augmented technology (RAG) and for agentic AI.

Agentic AI methods and AI brokers typically must course of and purpose over longer sequences of data, similar to bigger paperwork, log traces or prolonged conversations. The elevated 128k context size permits these agentic AI methods to have entry to extra contextual info, enabling them to raised perceive and reply to advanced queries or duties.

IBM can be releasing a collection of embedding fashions to assist speed up the method of changing information into vectors. The Granite-Embedding-30M-English mannequin can obtain efficiency of 0.16 seconds per question, which IBM claims is quicker than rival choices together with Snowflake’s Arctic.

How IBM has improved Granite 3.1 to serve enterprise AI wants

So how did IBM handle to enhance its efficiency for Granite 3.1? It wasn’t anybody particular factor, however fairly a collection of course of and technical improvements, Cox defined.

IBM has developed more and more superior multi-stage coaching pipelines, he mentioned. This has allowed the corporate to extract extra efficiency from fashions. Additionally, a crucial a part of any LLM coaching is information. Relatively than simply specializing in growing the amount of coaching information, IBM has put a powerful emphasis on bettering the standard of information used to coach the Granite fashions.

“It’s not a quantity game,” mentioned Cox. “It’s not like we’re going to go out and get 10 times more data and that’s magically going to make models better.”

Lowering hallucination immediately within the mannequin

A typical strategy to decreasing the chance of hallucinations and errant outputs in LLMs is to make use of guardrails. These are sometimes deployed as exterior options alongside an LLM.

With Granite 3.1, IBM is integrating hallucination safety immediately into the mannequin. The Granite Guardian 3.1 8B and 2B fashions now embrace a function-calling hallucination detection functionality.

“The model can natively do its own guardrailing, which can give different opportunities to developers to catch things,” mentioned Cox. 

He defined that performing hallucination detection within the mannequin itself optimizes the general course of. Inner detection means fewer inference calls, making the mannequin extra environment friendly and correct.

How enterprises can use Granite 3.1 in the present day, and what’s subsequent

The brand new Granite fashions are all now freely obtainable as open supply to enterprise customers. The fashions are additionally obtainable through IBM’s Watsonx enterprise AI service and will probably be built-in into IBM’s business merchandise.

The corporate plans on retaining an aggressive tempo for updating the Granite fashions. Trying ahead, the plan for Granite 3.2 is so as to add multimodal performance that may debut in early 2025. 

“You’re gonna see us over the next few point releases, adding more of these kinds of different features that are differentiated, leading up to the stuff that we’ll announce at the IBM Think conference next year,” mentioned Cox.

Day by day insights on enterprise use circumstances with VB Day by day

If you wish to impress your boss, VB Day by day has you coated. We provide the inside scoop on what corporations are doing with generative AI, from regulatory shifts to sensible deployments, so you may share insights for optimum ROI.

An error occured.

You Might Also Like

AI denial is turning into an enterprise threat: Why dismissing “slop” obscures actual functionality positive factors

GAM takes purpose at “context rot”: A dual-agent reminiscence structure that outperforms long-context LLMs

The 'reality serum' for AI: OpenAI’s new technique for coaching fashions to admit their errors

Anthropic vs. OpenAI pink teaming strategies reveal completely different safety priorities for enterprise AI

Inside NetSuite’s subsequent act: Evan Goldberg on the way forward for AI-powered enterprise methods

TAGGED:enterpriseGraniteIBMKingLLMmodelsopensource
Share This Article
Facebook Twitter Email Print

Follow US

Find US on Social Medias
FacebookLike
TwitterFollow
YoutubeSubscribe
TelegramFollow
Popular News
Amid Jan. 6 Revelations, Election Lies Still Dominate the G.O.P.
Politics

Amid Jan. 6 Revelations, Election Lies Still Dominate the G.O.P.

Editorial Board June 18, 2022
South Africa’s Corruption Inquiry Leaves Few of the Nation’s Powerful Unscathed
Emmy nominations 2025: Checklist of nominees
Required Studying
A trio of latest TV thrillers can present some motion and escapism this Thanksgiving

You Might Also Like

Nvidia's new AI framework trains an 8B mannequin to handle instruments like a professional
Technology

Nvidia's new AI framework trains an 8B mannequin to handle instruments like a professional

December 4, 2025
Gong examine: Gross sales groups utilizing AI generate 77% extra income per rep
Technology

Gong examine: Gross sales groups utilizing AI generate 77% extra income per rep

December 4, 2025
AWS launches Kiro powers with Stripe, Figma, and Datadog integrations for AI-assisted coding
Technology

AWS launches Kiro powers with Stripe, Figma, and Datadog integrations for AI-assisted coding

December 4, 2025
Workspace Studio goals to unravel the true agent drawback: Getting staff to make use of them
Technology

Workspace Studio goals to unravel the true agent drawback: Getting staff to make use of them

December 4, 2025

Categories

  • Health
  • Sports
  • Politics
  • Entertainment
  • Technology
  • Art
  • World

About US

New York Dawn is a proud and integral publication of the Enspirers News Group, embodying the values of journalistic integrity and excellence.
Company
  • About Us
  • Newsroom Policies & Standards
  • Diversity & Inclusion
  • Careers
  • Media & Community Relations
  • Accessibility Statement
Contact Us
  • Contact Us
  • Contact Customer Care
  • Advertise
  • Licensing & Syndication
  • Request a Correction
  • Contact the Newsroom
  • Send a News Tip
  • Report a Vulnerability
Term of Use
  • Digital Products Terms of Sale
  • Terms of Service
  • Privacy Policy
  • Cookie Settings
  • Submissions & Discussion Policy
  • RSS Terms of Service
  • Ad Choices
© 2024 New York Dawn. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?