Connect with us

Startup

Meta rolls out its biggest open source AI model Llama 3.1 405B

Published

on


Tech firm Meta has unveiled its largest open-source AI model, Llama 3.1 405B.

The model boasts 405 billion parameters, making it one of the most advanced and capable models in the open-source AI space, said the company.

The model will be available free of charge, enabling widespread access to the developer community. It was trained using 16,000 NVIDIA‘s H100 GPU (graphic processing unit) and leverages advanced training and development techniques.

Meta claims the model can outperform OpenAI‘s GPT-4o and Anthropic’s Claude 3.5 Sonnet on several benchmarks. Developers can run inference with Llama 3.1 405B on their own infrastructure at approximately half the cost of using other proprietary models such as GPT-4o—for both user-facing and offline inference applications, it said.

Towards an open and positive AI future

In an open letter, Meta CEO Mark Zuckerberg emphasised the importance of open source for a positive AI future. 

He stated that open-source AI models are not only catching up but are already advancing faster than proprietary models, akin to how Linux evolved to become industry standard foundation, for both cloud computing and the operating systems that run most mobile devices. 

“I believe that open source is necessary for a positive AI future. AI has more potential than any other modern technology to increase human productivity, creativity, and quality of life–and to accelerate economic growth while unlocking progress in medical and scientific research.

“Open source will ensure that more people around the world have access to the benefits and opportunities of AI, that power isn’t concentrated in the hands of a small number of companies, and that the technology can be deployed more evenly and safely across society,” said Zuckerberg. 

The company is actively building partnerships with companies such as Scale.AI, Dell, and Deloitte, which aim to assist enterprises in adopting Llama and train custom models using their own data.

Llama 3.1 405B is available on cloud platforms such as AWS, Azure, and Google Cloud.

“With past Llama models, Meta developed them for ourselves and then released them, but didn’t focus much on building a broader ecosystem. We’re taking a different approach with this release. We’re building teams internally to enable as many developers and partners as possible to use Llama, and we’re actively building partnerships so that more companies in the ecosystem can offer unique functionality to their customers as well,” said Zuckerberg.

He further stated that Meta’s safety process involves rigorous testing and red teaming to identify potential harms from their models, aiming to mitigate risks before release. Since the models are open-source, external parties can also conduct their own tests. 

Zuckerburg also drew a comparison between Meta’s investment in open-source AI and its previous Open Compute Project and claimed to have saved the company “billions”.

“Meta has a long history of open-source projects and successes. We’ve saved billions of dollars by releasing our server, network, and data center designs with Open Compute Project and having supply chains standardize on our designs. This approach has consistently worked for us when we stick with it over the long term,” he said.





Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Startup

Thesys secures $4M funding led by Together Fund

Published

on

By


AI startup Thesys bags $4 million funding in a round led by Together Fund. The round also saw participation from 8VC, the company said in a statement.

The startup will use the funding to bridge the gap of user experience with AI agents. As a visual collaboration tool, the company will also provide a platform that will enable businesses to ideate, visualise, and ship intelligent experiences at scale.

“The way we engage with technology is changing faster than ever. Static interfaces simply don’t meet the demands of today’s AI-capabilities…At Thesys, we’re building tools that make it possible for businesses to adapt and thrive in this new era,” said Parikshit Deshmukh, Co-founder, Thesys. 

This evolution is about unlocking the full potential of AI-driven interactions and delivering unparalleled user experiences, he added.

“The future of AI relies as much on intuitive, adaptive interfaces as it does on backend capabilities. Thesys’ vision for Generative UI aligns perfectly with Together Fund’s commitment to enabling founders who are redefining the user experience,” said Manav Garg, Co-founder and managing partner of Together Fund.

“By empowering teams to create real-time, personalized interactions, Thesys is setting a new standard for AI-driven interfaces. We’re excited to support their journey in transforming the role of design and development tools for the next generation of AI applications,” he added.

The company, founded by Rabi Shanker Guha and Parikshit Deshmukh this year, emerged from the understanding of the need to provide support in the shift towards AI-driven interfaces, it said.

“Thesys envisions a future where all interfaces dynamically adjust to each user’s behavior, preferences, and needs—driven by what the company calls “Generative UI”. Unlike traditional static interfaces that rely on predefined paths, Generative UI uses AI to create unique, adaptive user interfaces on-the-fly, allowing businesses to provide truly personalized digital experiences,” the company added.

The company plans to launch a UI SDK that is set to enable developers to seamlessly integrate Generative UI into their applications. Additionally, post its closed beta launch, the company plans a general availability (GA) with its product within the next quarter positioning itself as the go-to product toolkit for businesses looking to stay ahead in the AI revolution.

“Thesys is pioneering a transformative shift in UI design workflows by integrating AI-driven adaptability… Their Generative UI approach aligns with our commitment to investing in technologies that drive innovation in user experiences,” said Bhaskar Ghosh, partner at 8VC.





Source link

Continue Reading

Startup

BrowserStack launches AI-driven Low Code Automation tool

Published

on

By


Software testing platform BrowserStack has rolled out Low Code Automation, a solution to simplify test automation for quality assurance teams, developers, and non-technical users. 

The newly launched solution will address challenges faced by software teams, including manual testing delays and complex automation frameworks, BrowserStack said in a statement. 

While traditional test automation requires coding expertise by often limiting non-technical testers to contribute, this tool allows user—irrespective of their technical background—to create and manage AI-driven automated tests without writing code. Users can also use BrowserStack’s cloud infrastructure for reliable test execution.

“(The AI-powered Low-Code Automation (LCA) simplifies the process of building and maintaining test automation suites compared to traditional tools like Selenium. It reduces the steep learning curve and complexity often associated with automation projects, leading to a quicker return on investment (ROI),” Chintan Doshi, Director of Product Management at BrowserStack, told YourStory

To support development teams worldwide, Low Code Automation speeds up testing cycles, boosts product quality, and enhances user experience by reducing technical barriers. 

“Citizen testers—such as business analysts, product managers, and customer support teams—can easily add validations and create automated tests with the test recorder, without requiring coding skills. This reduces their dependency on developers and QAs and empowers them to actively contribute to testing efforts,” Doshi explained.

Founded in 2011 by Ritesh Arora and Nakul Aggarwal, BrowserStack provides a cloud-based platform for developers to test websites and mobile apps across devices, operating systems, and browsers on demand.

With headquarters in San Francisco and Mumbai, the company has expanded its product line to include over 15 products, of which 10 were launched in the past 18 months.

In August, the Accel-backed firm acquired Berlin-based Bird Eats Bug, an advanced bug-reporting tool. The acquisition aims to address the existing gaps in bug reporting and streamline fragmented testing workflows.





Source link

Continue Reading

Startup

Flipkart’s delivery arm Instakart reports widening losses, lower revenue in FY24

Published

on

By


Flipkart’s delivery service arm Instakart’s FY24 losses increased multifold to Rs 1718.4 crore, from Rs 324.6 crore in the previous year, hurt by higher expenses and marginally lower revenues. 

The company, which is in the logistics, warehouse, courier and allied services business, clocked an operating revenue of Rs 12,115.3 crore in FY24, 5% lower than Rs 12,787.4 crore it posted a year ago, according to filings made with Toefler. 

During the period, the company’s total expenses increased 6% to Rs 14,149.4 crore, mainly driven by employee benefit and other expenses. 

Logistics services accounted for the majority (about 78%) of Instakart’s total operating revenues, with Rs 9,429.8 crore, marginally lower than what it collected in the previous year.

Warehousing services, which accounted for about 10% of total operating revenues, witnessed a 28.4% drop in revenue, while collection services, which accounted for 12%, remained stable. 

Just a week ago, Flipkart Internet reported a 21% rise in FY24 revenue at Rs 17,907.3 crore helped by rising income from its advertising services.

Flipkart India Ltd, which is Flipkart’s business-to-business (B2B) arm, reported a 26.4% rise in revenue from operations at Rs 70,541.9 crore in FY24. 





Source link

Continue Reading

Trending

Copyright © 2017 Zox News Theme. Theme by MVP Themes, powered by WordPress.