Close Menu
The AI Book
    Facebook X (Twitter) Instagram
    The AI BookThe AI Book
    • Home
    • Categories
      • AI Media Processing
      • AI Language processing (NLP)
      • AI Marketing
      • AI Business Applications
    • Guides
    • Contact
    Subscribe
    Facebook X (Twitter) Instagram
    The AI Book
    Daily AI News

    Combat-happy Meta CEO Aims for Responsible and Safe AI with New Llama 2 Model

    18 July 2023No Comments4 Mins Read

    [ad_1]

    Meta CEO Mark Zuckerberg wants to punch Elon Musk in the face but is also looking to beat down AI rivals with his company’s latest large-language model.

    The Llama 2 model, which was announced by Meta on Tuesday, joins a rash of open-source software AI available for users to download.

    The model is available for free, and the model weights and tokenizers are available for download. But users need to fill out a download request as Meta wants to keep this AI model away from the hands of evildoers.

    EnterpriseAI made a download request, but Meta first evaluates the request. We received a download link within two hours of filling out the forms. But once Meta approves the request, it can be downloaded via Github or HuggingFace.

    The Llama 2 model is like ChatGPT – it was trained on available information on the Internet and can answer questions. Llama 2 is exclusively a chatbot, and once compiled, provides a prompt in which users can ask questions and compile stories.

    (AVIcon/Shutterstock)

    But Meta, which is the king of offering chat interfaces through its Facebook, WhatsApp, and newly developed Threads application, isn’t offering a direct interface for users to try Llama 2 out.

    “We’re not offering Llama in a chat interface. Researchers or organizations would have to build their own interface on top of the model,” a Meta spokesperson told EnterpriseAI.

    Meta may have refused to offer a chat interface for Llama 2 as it may not want to deal with the backlash of hallucinating and biased chatbots presented by OpenAI and Microsoft, which had to put safety mechanisms in place. Meta has a history of political backlash for its controversial role in trying to sway public dialog.

    Meta’s goal was to produce a transformer model that was safe at the outset. Custom implementations of Llama 2 on PCs will not be censored by Meta. Users can report problematic output of Llama 2 to Meta via the company’s website.

    Users can also use Llama 2 via cloud services like Amazon Web Services and Microsoft. On Amazon, it is available via SageMaker JumpStart, which is widely used by Fortune 500 companies to test and develop AI models.

    Meta claims the new transformer model is smarter than its predecessor, Llama 1, as it can reason better and provide more relevant answers. Llama 2 was trained on 40% more data than Llama 1, which reduces the instances of hallucinating or incorrect answers.

    “We have taken measures to increase the safety of these models, using safety-specific data annotation and tuning, as well as conducting red-teaming and employing iterative evaluations,” Meta researchers said in a paper outlining Llama 2.

    (Ole.CNX/Shutterstock)

    It took six months to train Llama 2, which included the pre-training and inputs based on human feedback. The training surprisingly relied heavily on supervised training at various levels of finetuning, which is a thumbs down on unsupervised training techniques, which was used to train GPT-4.

    “The tuned versions use supervised fine-tuning and reinforcement learning with human feedback to align to human preferences for helpfulness and safety,” Meta researchers wrote in the paper.

    The model training also included techniques reminiscent of decision tree steps, in which less important data is rejected as part of the finetuning process.

    The model comes with parameters ranging from 7 billion to 70 billion parameters. Meta claimed Llama 2 had better reasoning capabilities than open-source AI transformer models including Falcon and MosaicML’s MPT with a comparable number of parameters.

    But Llama 2 is not necessarily better than closed-source AI transformer models like GPT-4.

    “They also appear to be on par with some of the closed-source models, at least on the human evaluations we performed,” the researchers wrote in the paper.

    The model was trained at Meta’s Research Super Cluster on Nvidia A100 GPUs with 80GB of memory. The training was completed in 3.3 million GPU hours.

    Related

    [ad_2]

    Source link

    Previous ArticleFacebook parent Meta unveils LLaMA 2 open-source AI model for commercial use 
    Next Article SAP invests directly in three AI startups: Cohere, Anthropic and Aleph Alpha
    The AI Book

    Related Posts

    Daily AI News

    Adobe Previews New GenAI Tools for Video Workflows

    16 April 2024
    Daily AI News

    Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

    15 April 2024
    Daily AI News

    8 Reasons to Make the Switch

    15 April 2024
    Add A Comment
    Leave A Reply Cancel Reply

    • Privacy Policy
    • Terms and Conditions
    • About Us
    • Contact Form
    © 2025 The AI Book.

    Type above and press Enter to search. Press Esc to cancel.