AI trends 2025

AI is developing all the time. Here are some picks from several articles what is expected to happen in AI and around it in 2025. Here are picks from various articles, the texts are picks from the article edited and in some cases translated for clarity.

AI in 2025: Five Defining Themes
https://news.sap.com/2025/01/ai-in-2025-defining-themes/
Artificial intelligence (AI) is accelerating at an astonishing pace, quickly moving from emerging technologies to impacting how businesses run. From building AI agents to interacting with technology in ways that feel more like a natural conversation, AI technologies are poised to transform how we work.
But what exactly lies ahead?
1. Agentic AI: Goodbye Agent Washing, Welcome Multi-Agent Systems
AI agents are currently in their infancy. While many software vendors are releasing and labeling the first “AI agents” based on simple conversational document search, advanced AI agents that will be able to plan, reason, use tools, collaborate with humans and other agents, and iteratively reflect on progress until they achieve their objective are on the horizon. The year 2025 will see them rapidly evolve and act more autonomously. More specifically, 2025 will see AI agents deployed more readily “under the hood,” driving complex agentic workflows.
In short, AI will handle mundane, high-volume tasks while the value of human judgement, creativity, and quality outcomes will increase.
2. Models: No Context, No Value
Large language models (LLMs) will continue to become a commodity for vanilla generative AI tasks, a trend that has already started. LLMs are drawing on an increasingly tapped pool of public data scraped from the internet. This will only worsen, and companies must learn to adapt their models to unique, content-rich data sources.
We will also see a greater variety of foundation models that fulfill different purposes. Take, for example, physics-informed neural networks (PINNs), which generate outcomes based on predictions grounded in physical reality or robotics. PINNs are set to gain more importance in the job market because they will enable autonomous robots to navigate and execute tasks in the real world.
Models will increasingly become more multimodal, meaning an AI system can process information from various input types.
3. Adoption: From Buzz to Business
While 2024 was all about introducing AI use cases and their value for organizations and individuals alike, 2025 will see the industry’s unprecedented adoption of AI specifically for businesses. More people will understand when and how to use AI, and the technology will mature to the point where it can deal with critical business issues such as managing multi-national complexities. Many companies will also gain practical experience working for the first time through issues like AI-specific legal and data privacy terms (compared to when companies started moving to the cloud 10 years ago), building the foundation for applying the technology to business processes.
4. User Experience: AI Is Becoming the New UI
AI’s next frontier is seamlessly unifying people, data, and processes to amplify business outcomes. In 2025, we will see increased adoption of AI across the workforce as people discover the benefits of humans plus AI.
This means disrupting the classical user experience from system-led interactions to intent-based, people-led conversations with AI acting in the background. AI copilots will become the new UI for engaging with a system, making software more accessible and easier for people. AI won’t be limited to one app; it might even replace them one day. With AI, frontend, backend, browser, and apps are blurring. This is like giving your AI “arms, legs, and eyes.”
5. Regulation: Innovate, Then Regulate
It’s fair to say that governments worldwide are struggling to keep pace with the rapid advancements in AI technology and to develop meaningful regulatory frameworks that set appropriate guardrails for AI without compromising innovation.

12 AI predictions for 2025
This year we’ve seen AI move from pilots into production use cases. In 2025, they’ll expand into fully-scaled, enterprise-wide deployments.
https://www.cio.com/article/3630070/12-ai-predictions-for-2025.html
This year we’ve seen AI move from pilots into production use cases. In 2025, they’ll expand into fully-scaled, enterprise-wide deployments.
1. Small language models and edge computing
Most of the attention this year and last has been on the big language models — specifically on ChatGPT in its various permutations, as well as competitors like Anthropic’s Claude and Meta’s Llama models. But for many business use cases, LLMs are overkill and are too expensive, and too slow, for practical use.
“Looking ahead to 2025, I expect small language models, specifically custom models, to become a more common solution for many businesses,”
2. AI will approach human reasoning ability
In mid-September, OpenAI released a new series of models that thinks through problems much like a person would, it claims. The company says it can achieve PhD-level performance in challenging benchmark tests in physics, chemistry, and biology. For example, the previous best model, GPT-4o, could only solve 13% of the problems on the International Mathematics Olympiad, while the new reasoning model solved 83%.
If AI can reason better, then it will make it possible for AI agents to understand our intent, translate that into a series of steps, and do things on our behalf, says Gartner analyst Arun Chandrasekaran. “Reasoning also helps us use AI as more of a decision support system,”
3. Massive growth in proven use cases
This year, we’ve seen some use cases proven to have ROI, says Monteiro. In 2025, those use cases will see massive adoption, especially if the AI technology is integrated into the software platforms that companies are already using, making it very simple to adopt.
“The fields of customer service, marketing, and customer development are going to see massive adoption,”
4. The evolution of agile development
The agile manifesto was released in 2001 and, since then, the development philosophy has steadily gained over the previous waterfall style of software development.
“For the last 15 years or so, it’s been the de-facto standard for how modern software development works,”
5. Increased regulation
At the end of September, California governor Gavin Newsom signed a law requiring gen AI developers to disclose the data they used to train their systems, which applies to developers who make gen AI systems publicly available to Californians. Developers must comply by the start of 2026.
There are also regulations about the use of deep fakes, facial recognition, and more. The most comprehensive law, the EU’s AI Act, which went into effect last summer, is also something that companies will have to comply with starting in mid-2026, so, again, 2025 is the year when they will need to get ready.
6. AI will become accessible and ubiquitous
With gen AI, people are still at the stage of trying to figure out what gen AI is, how it works, and how to use it.
“There’s going to be a lot less of that,” he says. But gen AI will become ubiquitous and seamlessly woven into workflows, the way the internet is today.
7. Agents will begin replacing services
Software has evolved from big, monolithic systems running on mainframes, to desktop apps, to distributed, service-based architectures, web applications, and mobile apps. Now, it will evolve again, says Malhotra. “Agents are the next phase,” he says. Agents can be more loosely coupled than services, making these architectures more flexible, resilient and smart. And that will bring with it a completely new stack of tools and development processes.
8. The rise of agentic assistants
In addition to agents replacing software components, we’ll also see the rise of agentic assistants, adds Malhotra. Take for example that task of keeping up with regulations.
Today, consultants get continuing education to stay abreast of new laws, or reach out to colleagues who are already experts in them. It takes time for the new knowledge to disseminate and be fully absorbed by employees.
“But an AI agent can be instantly updated to ensure that all our work is compliant with the new laws,” says Malhotra. “This isn’t science fiction.”
9. Multi-agent systems
Sure, AI agents are interesting. But things are going to get really interesting when agents start talking to each other, says Babak Hodjat, CTO of AI at Cognizant. It won’t happen overnight, of course, and companies will need to be careful that these agentic systems don’t go off the rails.
Companies such as Sailes and Salesforce are already developing multi-agent workflows.
10. Multi-modal AI
Humans and the companies we build are multi-modal. We read and write text, we speak and listen, we see and we draw. And we do all these things through time, so we understand that some things come before other things. Today’s AI models are, for the most part, fragmentary. One can create images, another can only handle text, and some recent ones can understand or produce video.
11. Multi-model routing
Not to be confused with multi-modal AI, multi-modal routing is when companies use more than one LLM to power their gen AI applications. Different AI models are better at different things, and some are cheaper than others, or have lower latency. And then there’s the matter of having all your eggs in one basket.
“A number of CIOs I’ve spoken with recently are thinking about the old ERP days of vendor lock,” says Brett Barton, global AI practice leader at Unisys. “And it’s top of mind for many as they look at their application portfolio, specifically as it relates to cloud and AI capabilities.”
Diversifying away from using just a single model for all use cases means a company is less dependent on any one provider and can be more flexible as circumstances change.
12. Mass customization of enterprise software
Today, only the largest companies, with the deepest pockets, get to have custom software developed specifically for them. It’s just not economically feasible to build large systems for small use cases.
“Right now, people are all using the same version of Teams or Slack or what have you,” says Ernst & Young’s Malhotra. “Microsoft can’t make a custom version just for me.” But once AI begins to accelerate the speed of software development while reducing costs, it starts to become much more feasible.

9 IT resolutions for 2025
https://www.cio.com/article/3629833/9-it-resolutions-for-2025.html
1. Innovate
“We’re embracing innovation,”
2. Double down on harnessing the power of AI
Not surprisingly, getting more out of AI is top of mind for many CIOs.
“I am excited about the potential of generative AI, particularly in the security space,”
3. And ensure effective and secure AI rollouts
“AI is everywhere, and while its benefits are extensive, implementing it effectively across a corporation presents challenges. Balancing the rollout with proper training, adoption, and careful measurement of costs and benefits is essential, particularly while securing company assets in tandem,”
4. Focus on responsible AI
The possibilities of AI grow by the day — but so do the risks.
“My resolution is to mature in our execution of responsible AI,”
“AI is the new gold and in order to truly maximize it’s potential, we must first have the proper guardrails in place. Taking a human-first approach to AI will help ensure our state can maintain ethics while taking advantage of the new AI innovations.”
5. Deliver value from generative AI
As organizations move from experimenting and testing generative AI use cases, they’re looking for gen AI to deliver real business value.
“As we go into 2025, we’ll continue to see the evolution of gen AI. But it’s no longer about just standing it up. It’s more about optimizing and maximizing the value we’re getting out of gen AI,”
6. Empower global talent
Although harnessing AI is a top objective for Morgan Stanley’s Wetmur, she says she’s equally committed to harnessing the power of people.
7. Create a wholistic learning culture
Wetmur has another talent-related objective: to create a learning culture — not just in her own department but across all divisions.
8. Deliver better digital experiences
Deltek’s Cilsick has her sights set on improving her company’s digital employee experience, believing that a better DEX will yield benefits in multiple ways.
Cilsick says she first wants to bring in new technologies and automation to “make things as easy as possible,” mirroring the digital experiences most workers have when using consumer technologies.
“It’s really about leveraging tech to make sure [employees] are more efficient and productive,”
“In 2025 my primary focus as CIO will be on transforming operational efficiency, maximizing business productivity, and enhancing employee experiences,”
9. Position the company for long-term success
Lieberman wants to look beyond 2025, saying another resolution for the year is “to develop a longer-term view of our technology roadmap so that we can strategically decide where to invest our resources.”
“My resolutions for 2025 reflect the evolving needs of our organization, the opportunities presented by AI and emerging technologies, and the necessity to balance innovation with operational efficiency,”
Lieberman aims to develop AI capabilities to automate routine tasks.
“Bots will handle common inquiries ranging from sales account summaries to HR benefits, reducing response times and freeing up resources for strategic initiatives,”

Not just hype — here are real-world use cases for AI agents
https://venturebeat.com/ai/not-just-hype-here-are-real-world-use-cases-for-ai-agents/
Just seven or eight months ago, when a customer called in to or emailed Baca Systems with a service question, a human agent handling the query would begin searching for similar cases in the system and analyzing technical documents.
This process would take roughly five to seven minutes; then the agent could offer the “first meaningful response” and finally begin troubleshooting.
But now, with AI agents powered by Salesforce, that time has been shortened to as few as five to 10 seconds.
Now, instead of having to sift through databases for previous customer calls and similar cases, human reps can ask the AI agent to find the relevant information. The AI runs in the background and allows humans to respond right away, Russo noted.
AI can serve as a sales development representative (SDR) to send out general inquires and emails, have a back-and-forth dialogue, then pass the prospect to a member of the sales team, Russo explained.
But once the company implements Salesforce’s Agentforce, a customer needing to modify an order will be able to communicate their needs with AI in natural language, and the AI agent will automatically make adjustments. When more complex issues come up — such as a reconfiguration of an order or an all-out venue change — the AI agent will quickly push the matter up to a human rep.

Open Source in 2025: Strap In, Disruption Straight Ahead
Look for new tensions to arise in the New Year over licensing, the open source AI definition, security and compliance, and how to pay volunteer maintainers.
https://thenewstack.io/open-source-in-2025-strap-in-disruption-straight-ahead/
The trend of widely used open source software moving to more restrictive licensing isn’t new.
In addition to the demands of late-stage capitalism and impatient investors in companies built on open source tools, other outside factors are pressuring the open source world. There’s the promise/threat of generative AI, for instance. Or the shifting geopolitical landscape, which brings new security concerns and governance regulations.
What’s ahead for open source in 2025?
More Consolidation, More Licensing Changes
The Open Source AI Debate: Just Getting Started
Security and Compliance Concerns Will Rise
Paying Maintainers: More Cash, Creativity Needed

Kyberturvallisuuden ja tekoälyn tärkeimmät trendit 2025
https://www.uusiteknologia.fi/2024/11/20/kyberturvallisuuden-ja-tekoalyn-tarkeimmat-trendit-2025/
1. Cyber ​​infrastructure will be centered on a single, unified security platform
2. Big data will give an edge against new entrants
3. AI’s integrated role in 2025 means building trust, governance engagement, and a new kind of leadership
4. Businesses will adopt secure enterprise browsers more widely
5. AI’s energy implications will be more widely recognized in 2025
6. Quantum realities will become clearer in 2025
7. Security and marketing leaders will work more closely together

Presentation: For 2025, ‘AI eats the world’.
https://www.ben-evans.com/presentations

Just like other technologies that have gone before, such as cloud and cybersecurity automation, right now AI lacks maturity.
https://www.securityweek.com/ai-implementing-the-right-technology-for-the-right-use-case/
If 2023 and 2024 were the years of exploration, hype and excitement around AI, 2025 (and 2026) will be the year(s) that organizations start to focus on specific use cases for the most productive implementations of AI and, more importantly, to understand how to implement guardrails and governance so that it is viewed as less of a risk by security teams and more of a benefit to the organization.
Businesses are developing applications that add Large Language Model (LLM) capabilities to provide superior functionality and advanced personalization
Employees are using third party GenAI tools for research and productivity purposes
Developers are leveraging AI-powered code assistants to code faster and meet challenging production deadlines
Companies are building their own LLMs for internal use cases and commercial purposes.
AI is still maturing
However, just like other technologies that have gone before, such as cloud and cybersecurity automation, right now AI lacks maturity. Right now, we very much see AI in this “peak of inflated expectations” phase and predict that it will dip into the “trough of disillusionment”, where organizations realize that it is not the silver bullet they thought it would be. In fact, there are already signs of cynicism as decision-makers are bombarded with marketing messages from vendors and struggle to discern what is a genuine use case and what is not relevant for their organization.
There is also regulation that will come into force, such as the EU AI Act, which is a comprehensive legal framework that sets out rules for the development and use of AI.
AI certainly won’t solve every problem, and it should be used like automation, as part of a collaborative mix of people, process and technology. You simply can’t replace human intuition with AI, and many new AI regulations stipulate that human oversight is maintained.

7 Splunk Predictions for 2025
https://www.splunk.com/en_us/form/future-predictions.html
AI: Projects must prove their worth to anxious boards or risk defunding, and LLMs will go small to reduce operating costs and environmental impact.

OpenAI, Google and Anthropic Are Struggling to Build More Advanced AI
Three of the leading artificial intelligence companies are seeing diminishing returns from their costly efforts to develop newer models.
https://www.bloomberg.com/news/articles/2024-11-13/openai-google-and-anthropic-are-struggling-to-build-more-advanced-ai
Sources: OpenAI, Google, and Anthropic are all seeing diminishing returns from costly efforts to build new AI models; a new Gemini model misses internal targets

It Costs So Much to Run ChatGPT That OpenAI Is Losing Money on $200 ChatGPT Pro Subscriptions
https://futurism.com/the-byte/openai-chatgpt-pro-subscription-losing-money?fbclid=IwY2xjawH8epVleHRuA2FlbQIxMQABHeggEpKe8ZQfjtPRC0f2pOI7A3z9LFtFon8lVG2VAbj178dkxSQbX_2CJQ_aem_N_ll3ETcuQ4OTRrShHqNGg
In a post on X-formerly-Twitter, CEO Sam Altman admitted an “insane” fact: that the company is “currently losing money” on ChatGPT Pro subscriptions, which run $200 per month and give users access to its suite of products including its o1 “reasoning” model.
“People use it much more than we expected,” the cofounder wrote, later adding in response to another user that he “personally chose the price and thought we would make some money.”
Though Altman didn’t explicitly say why OpenAI is losing money on these premium subscriptions, the issue almost certainly comes down to the enormous expense of running AI infrastructure: the massive and increasing amounts of electricity needed to power the facilities that power AI, not to mention the cost of building and maintaining those data centers. Nowadays, a single query on the company’s most advanced models can cost a staggering $1,000.

Tekoäly edellyttää yhä nopeampia verkkoja
https://etn.fi/index.php/opinion/16974-tekoaely-edellyttaeae-yhae-nopeampia-verkkoja
A resilient digital infrastructure is critical to effectively harnessing telecommunications networks for AI innovations and cloud-based services. The increasing demand for data-rich applications related to AI requires a telecommunications network that can handle large amounts of data with low latency, writes Carl Hansson, Partner Solutions Manager at Orange Business.

AI’s Slowdown Is Everyone Else’s Opportunity
Businesses will benefit from some much-needed breathing space to figure out how to deliver that all-important return on investment.
https://www.bloomberg.com/opinion/articles/2024-11-20/ai-slowdown-is-everyone-else-s-opportunity

Näin sirumarkkinoilla käy ensi vuonna
https://etn.fi/index.php/13-news/16984-naein-sirumarkkinoilla-kaey-ensi-vuonna
The growing demand for high-performance computing (HPC) for artificial intelligence and HPC computing continues to be strong, with the market set to grow by more than 15 percent in 2025, IDC estimates in its recent Worldwide Semiconductor Technology Supply Chain Intelligence report.
IDC predicts eight significant trends for the chip market by 2025.
1. AI growth accelerates
2. Asia-Pacific IC Design Heats Up
3. TSMC’s leadership position is strengthening
4. The expansion of advanced processes is accelerating.
5. Mature process market recovers
6. 2nm Technology Breakthrough
7. Restructuring the Packaging and Testing Market
8. Advanced packaging technologies on the rise

2024: The year when MCUs became AI-enabled
https://www-edn-com.translate.goog/2024-the-year-when-mcus-became-ai-enabled/?fbclid=IwZXh0bgNhZW0CMTEAAR1_fEakArfPtgGZfjd-NiPd_MLBiuHyp9qfiszczOENPGPg38wzl9KOLrQ_aem_rLmf2vF2kjDIFGWzRVZWKw&_x_tr_sl=en&_x_tr_tl=fi&_x_tr_hl=fi&_x_tr_pto=wapp
The AI ​​party in the MCU space started in 2024, and in 2025, it is very likely that there will be more advancements in MCUs using lightweight AI models.
Adoption of AI acceleration features is a big step in the development of microcontrollers. The inclusion of AI features in microcontrollers started in 2024, and it is very likely that in 2025, their features and tools will develop further.

Just like other technologies that have gone before, such as cloud and cybersecurity automation, right now AI lacks maturity.
https://www.securityweek.com/ai-implementing-the-right-technology-for-the-right-use-case/
If 2023 and 2024 were the years of exploration, hype and excitement around AI, 2025 (and 2026) will be the year(s) that organizations start to focus on specific use cases for the most productive implementations of AI and, more importantly, to understand how to implement guardrails and governance so that it is viewed as less of a risk by security teams and more of a benefit to the organization.
Businesses are developing applications that add Large Language Model (LLM) capabilities to provide superior functionality and advanced personalization
Employees are using third party GenAI tools for research and productivity purposes
Developers are leveraging AI-powered code assistants to code faster and meet challenging production deadlines
Companies are building their own LLMs for internal use cases and commercial purposes.
AI is still maturing

AI Regulation Gets Serious in 2025 – Is Your Organization Ready?
While the challenges are significant, organizations have an opportunity to build scalable AI governance frameworks that ensure compliance while enabling responsible AI innovation.
https://www.securityweek.com/ai-regulation-gets-serious-in-2025-is-your-organization-ready/
Similar to the GDPR, the EU AI Act will take a phased approach to implementation. The first milestone arrives on February 2, 2025, when organizations operating in the EU must ensure that employees involved in AI use, deployment, or oversight possess adequate AI literacy. Thereafter from August 1 any new AI models based on GPAI standards must be fully compliant with the act. Also similar to GDPR is the threat of huge fines for non-compliance – EUR 35 million or 7 percent of worldwide annual turnover, whichever is higher.
While this requirement may appear manageable on the surface, many organizations are still in the early stages of defining and formalizing their AI usage policies.
Later phases of the EU AI Act, expected in late 2025 and into 2026, will introduce stricter requirements around prohibited and high-risk AI applications. For organizations, this will surface a significant governance challenge: maintaining visibility and control over AI assets.
Tracking the usage of standalone generative AI tools, such as ChatGPT or Claude, is relatively straightforward. However, the challenge intensifies when dealing with SaaS platforms that integrate AI functionalities on the backend. Analysts, including Gartner, refer to this as “embedded AI,” and its proliferation makes maintaining accurate AI asset inventories increasingly complex.
Where frameworks like the EU AI Act grow more complex is their focus on ‘high-risk’ use cases. Compliance will require organizations to move beyond merely identifying AI tools in use; they must also assess how these tools are used, what data is being shared, and what tasks the AI is performing. For instance, an employee using a generative AI tool to summarize sensitive internal documents introduces very different risks than someone using the same tool to draft marketing content.
For security and compliance leaders, the EU AI Act represents just one piece of a broader AI governance puzzle that will dominate 2025.
The next 12-18 months will require sustained focus and collaboration across security, compliance, and technology teams to stay ahead of these developments.

The Global Partnership on Artificial Intelligence (GPAI) is a multi-stakeholder initiative which aims to bridge the gap between theory and practice on AI by supporting cutting-edge research and applied activities on AI-related priorities.
https://gpai.ai/about/#:~:text=The%20Global%20Partnership%20on%20Artificial,activities%20on%20AI%2Drelated%20priorities.

1,988 Comments

  1. Tomi Engdahl says:

    Jay Peters / The Verge:
    OpenAI is adding an image library to ChatGPT to make it easier for users to access their AI-generated images, rolling out to all Free, Plus, and Pro users — The new Library is rolling out now to all Free, Plus, and Pro users. … OpenAI is adding an image library to ChatGPT to make it easier …

    ChatGPT now has a section for your AI-generated images
    The new Library is rolling out now to all Free, Plus, and Pro users.
    https://www.theverge.com/news/649247/chatgpt-image-library

    Reply
  2. Tomi Engdahl says:

    Kyle Wiggers / TechCrunch:
    xAI unveils Grok Studio, a canvas-like tool for its Grok chatbot that can generate and edit documents, run code like Python and JavaScript, and more — Grok, the chatbot from Elon Musk’s AI company, xAI, has gained a canvas-like feature for editing and creating documents and basic apps.

    Grok gains a canvas-like tool for creating docs and apps
    https://techcrunch.com/2025/04/15/grok-gains-a-canvas-like-tool-for-creating-docs-and-apps/

    Grok, the chatbot from Elon Musk’s AI company, xAI, has gained a canvas-like feature for editing and creating documents and basic apps.

    Called Grok Studio, the feature was announced on X late Tuesday. It’s available for both free and paying Grok users on Grok.com

    “Grok can now generate documents, code, reports and browser games,” wrote the official Grok account on X. “Grok Studio will open your content in a separate window, allowing both you and Grok to collaborate on the content together.”

    Grok is the latest chatbot to get a dedicated workspace for tinkering with software and writing projects. OpenAI launched a similar capability, Canvas, for ChatGPT in October. Anthropic was one of the first to the punch, with Artifacts for Claude.

    Reply
  3. Tomi Engdahl says:

    Rest of World:NEW
    Experts say Chinese investors would be wary of backing AI startups building foundation models; PitchBook: in Q1, AI/ML VC funding in China fell 30% YoY to $1.2B

    DeepSeek and chip bans have supercharged AI innovation in China
    Investors believe practical AI applications will fuel the next wave of growth, with startups building on top of DeepSeek’s foundation.
    https://restofworld.org/2025/china-ai-boom-chip-ban-deepseek/

    DeepSeek has rapidly accelerated AI adoption throughout Chinese society.
    Chinese tech giants have released a wave of free, open-source AI models.
    Startups are under pressure to abandon foundational research as investors prioritize practical applications.

    The rise of DeepSeek has reenergized China’s artificial intelligence industry, drawing billions of dollars in state funding and a bigger push toward self-sufficiency. Its success has also sparked a race among startups to build products and services on top of its high-performing open-source technology, even as the country’s tech giants have rushed out competing AI models.

    Widening U.S. chip bans, combined with DeepSeek’s success, have pushed the Chinese AI ecosystem to innovate more quickly, strengthening its position in the global tech landscape, experts and investors told Rest of World. It has also forced startups to pivot toward more practical applications.

    “DeepSeek has proven that Chinese AI labs can produce frontier models even with export control constraints,” Kevin Xu, founder of U.S.-based Interconnected Capital, a hedge fund investing in AI, told Rest of World. “Its success is also pushing more startups to work on building applications and services, and not waste time and resources building models.”

    “DeepSeek has proven that Chinese AI labs can produce frontier models even with export control constraints.”

    China lagged behind the U.S in the number of AI models produced last year, but Chinese models are rapidly closing the performance gap with U.S. models, according to a new report from the Stanford Institute for Human-Centered AI. Yet Chinese companies face several challenges, including expanded U.S. export controls limiting access to advanced chips.

    DeepSeek claimed to have trained its V3 foundation model — a large-scale AI system trained on vast data sets and adaptable to various tasks — using less-advanced Nvidia chips at a cost of just about $6 million, compared to more than $100 million for OpenAI’s GPT-4 model.

    Reply
  4. Tomi Engdahl says:

    Emilia David / VentureBeat:
    Cohere debuts Embed 4, its search engine for enterprise AI assistants, now with a 128K context length, better multimodal abilities for complex documents, more

    Cohere launches Embed 4: New multimodal search model processes 200-page documents
    https://venturebeat.com/ai/cohere-launches-embed-4-new-multimodal-search-model-processes-200-page-documents/

    Enterprise retrieval augmented generation (RAG) remains integral to the current agentic AI craze. Taking advantage of the continued interest in agents, Cohere released the latest version of its embeddings model with longer context windows and more multimodality.

    Cohere’s Embed 4 builds on the multimodal updates of Embed 3 and adds more capabilities around unstructured data. Thanks to a 128,000 token context window, organizations can generate embeddings for documents with around 200 pages.

    “Existing embedding models fail to natively understand complex multimodal business materials,‬‭ leading companies to develop cumbersome data pre-processing pipelines that only slightly‬‭ improve accuracy,” Cohere said in a blog post. “Embed 4 solves this problem, allowing enterprises and their employees to‬‭ efficiently surface insights that are hidden within mountains of unsearchable information.‬”

    Enterprises can deploy Embed 4 on virtual private clouds or on-premise technology stacks for added data security.

    Reply
  5. Tomi Engdahl says:

    Kyle Wiggers / TechCrunch:
    In the first update to its Preparedness Framework since 2023, OpenAI says it “may adjust” its own safeguards if rival frontier labs release “high-risk” AI

    OpenAI may ‘adjust’ its safeguards if rivals release ‘high-risk’ AI
    https://techcrunch.com/2025/04/15/openai-says-it-may-adjust-its-safety-requirements-if-a-rival-lab-releases-high-risk-ai/

    OpenAI has updated its Preparedness Framework — the internal system it uses to assess the safety of AI models and determine necessary safeguards during development and deployment. In the update, OpenAI stated that it may “adjust” its safety requirements if a competing AI lab releases a “high-risk” system without similar protections in place.

    The change reflects the increasing competitive pressures on commercial AI developers to deploy models quickly. OpenAI has been accused of lowering safety standards in favor of faster releases, and of failing to deliver timely reports detailing its safety testing. Last week, 12 former OpenAI employees filed a brief in Elon Musk’s case against OpenAI, arguing the company would be encouraged to cut even more corners on safety should it complete its planned corporate restructuring.

    Perhaps anticipating criticism, OpenAI claims that it wouldn’t make these policy adjustments lightly, and that it would keep its safeguards at “a level more protective.”

    “If another frontier AI developer releases a high-risk system without comparable safeguards, we may adjust our requirements,” wrote OpenAI in a blog post published Tuesday afternoon. “However, we would first rigorously confirm that the risk landscape has actually changed, publicly acknowledge that we are making an adjustment, assess that the adjustment does not meaningfully increase the overall risk of severe harm, and still keep safeguards at a level more protective.”

    Reply
  6. Tomi Engdahl says:

    CNBC:
    Filing: Nvidia says it will record a $5.5B charge after the US said it would require a license to export the H20 chips to China; NVDA drops 6%+ after hours — Nvidia said on Tuesday that it will take a quarterly charge of about $5.5 billion tied to exporting H20 graphics processing units to China and other destinations.

    Nvidia says it will record $5.5 billion charge tied to H20 processors exported to China
    https://www.cnbc.com/2025/04/15/nvidia-says-it-will-record-5point5-billion-quarterly-charge-tied-to-h20-processors-exported-to-china.html

    Nvidia said on Tuesday that it will take a quarterly charge of about $5.5 billion tied to exporting H20 graphics processing units to China and other destinations.
    The stock slid almost 5% in extended trading.
    The U.S. government, during the Biden administration, restricted AI chip exports in 2022 and then updated the rules the following year to prevent the sale of more advanced AI processors.

    Reply
  7. Tomi Engdahl says:

    Gmail-käyttäjien on pian tehtävä valinta
    Gmailin käyttäjien on pian valittava yksityisyyden ja kehittyneempien ominaisuuksien väliltä.
    https://www.iltalehti.fi/digiuutiset/a/92c53069-ddc4-41ed-9f4b-3cd9cf865dc5

    Gmailiin on tulossa uusia ominaisuuksia, joiden käyttöönotto voi vaatia kompromisseja yksityisyyden suhteen.

    Google tarjoaa käyttäjille mahdollisuuden valita, haluavatko he ottaa käyttöön uudet tekoälypohjaiset ominaisuudet vai pitäytyä perinteisemmässä sähköpostikokemuksessa.

    Uudet toiminnot voivat parantaa käyttökokemusta, mutta niiden vaikutuksesta käyttäjän yksityisyyteen on hyvä olla tietoinen.

    Gmailin uusista ominaisuuksista uutisoineen Newsweekin mukaan uudet tekoälytyökalut tarjoavat muun muassa parempia hakutuloksia, älykkäitä vastausehdotuksia sekä sähköpostien yhteenvetoja.

    Työkalujen käyttöönotto päästää Googlen kuitenkin syvemmälle sähköpostilaatikkoosi analysoimaan viestien sisältöä ja metadataa. Googlen mukaan se ei käytä keräämiään tietoja mainosten näyttämiseen suoraan Gmailissa, mutta niillä voi silti olla vaikutusta käyttäjän mainosprofiiliin muilla yhtiön alustoilla.

    Mikäli uusia työkaluja ei halua käyttöönsä, jatkuu Gmailin käyttökokemus perinteisempänä, eikä käyttäjä pääse enää hyödyntämään esimerkiksi tekoälyllä tehostettua hakua.

    Positiivisena puolena on kuitenkin se, että Google kerää käyttäjästä jatkossa vähemmän tietoa parantaen sähköpostien yksityisyyttä.

    Gmail Just Changed—Here’s What To Pick When Google Asks You
    https://www.newsweek.com/gmail-changes-2025-2058086

    If you’re a Gmail user, you may soon see a new prompt in your inbox—and it’s not just another pop-up to click away. Google is now rolling out a privacy-related update that gives users a clear choice: opt into AI-enhanced features that use your personal data, or stick with a more limited version of Gmail. While the decision may feel routine, it marks a significant shift in how the world’s most popular email service handles your data—and your inbox.
    What the New Gmail Prompt Means

    The update is tied to Google’s ongoing rollout of AI-powered search features in Gmail, which aim to make it easier to find relevant emails quickly. The new system ranks search results based on things like how recent an email was, frequency of interaction, and how often you click on certain messages. But to do that, Gmail needs access to more of your behavioral and content data.

    To comply with privacy standards and give users more control, Google is now asking you to make a decision:

    Allow Gmail to use AI-driven tools by enabling “smart features” and data sharing.
    Decline the use of AI tools, which limits certain features but offers stronger data privacy.

    What Happens If You Say ‘Yes’

    If you choose to enable these features, Gmail will continue to offer its latest AI-powered updates, including:

    Smarter search results based on your usage habits.
    Access to “smart compose” and “smart reply” tools.
    Better filtering of spam, phishing, and promotional emails.

    However, agreeing to this also means Google’s AI will analyze more of your inbox activity, including message content and metadata, to personalize your experience. While Google states that this data is not used to serve ads directly in Gmail, it may still feed into your broader ad profile across the company’s platforms.
    What Happens If You Say ‘No’

    Declining the update means you’ll still have access to basic Gmail functionality, but you’ll lose access to features like AI-enhanced search and auto-suggestions. Your inbox may also feel slower or less intuitive when searching for older messages or organizing conversations.

    On the upside, your data footprint with Google remains smaller. Gmail won’t process as much of your message content or behavior to power features, giving you more control over your email privacy.

    Why This Matters Now

    This update lands as concerns about data privacy and AI transparency continue to grow. While Google says it blocks more than 99 percent of spam and phishing attempts, cybersecurity experts warn that AI-generated attacks are getting more sophisticated. At the same time, privacy advocates have raised flags over just how much data is being processed behind the scenes.

    If you’ve ever wondered how much of your inbox activity is really private, now’s the time to take a closer look—and make an informed choice when Google asks.

    Reply
  8. Tomi Engdahl says:

    Claude yrittää Google-käyttäjien Copilotiksi
    https://etn.fi/index.php/13-news/17416-claude-yrittaeae-google-kaeyttaejien-copilotiksi

    OpenAI:n entisten työntekijöiden perustama Anthropic tuo tekoälyavustajansa syvälle Google Workspace -ympäristöön. Käytännössä Claude-tekoäly pyrkii tarjoamaan saman, mitä Microsoftin Copilot tekee 365-käyttäjille.

    Claude saa nyt pääsyn käyttäjän sähköposteihin (Gmail), kalenterimerkintöihin (Google Calendar) sekä dokumentteihin (Google Docs). Samalla Claude saa uuden “Research”-toiminnon, jonka avulla se voi hakea tietoa sekä käyttäjän omista lähteistä että internetistä itsenäisesti.

    Claude voi esimerkiksi koostaa muistiinpanot menneistä kokouksista, poimia toimintakohdat sähköposteista, ja yhdistää dokumenttitietoa eri lähteistä suunnittelun tueksi. Tarkoituksena on säästää käyttäjältä manuaalista työtä ja tarjota nopeasti tilannekuva mistä tahansa projektista tai tehtävästä – suoraan Claude-keskustelussa.

    Claude ei kilpaile Microsoftin Copilotin kanssa samassa käyttöympäristössä, mutta tavoite on käytännössä identtinen: toimia älykkäänä avustajana, joka ymmärtää käyttäjän työtä, tietoja ja aikatauluja kontekstuaalisesti. Erojakin on: siinä missä Copilot on integroitu tiukasti osaksi Microsoftin eri työkaluja, Claude on edelleen chat-tyyppinen agentti omassa sovelluksessaan.

    Toistaiseksi uusi Claude on tarjolla rajoitetusti ja vain maksavien tilaajien beeta-ominaisuutena.

    Reply
  9. Tomi Engdahl says:

    DIY AI Butler Is Simpler And More Useful Than Siri
    https://hackaday.com/2025/04/15/diy-ai-butler-is-simpler-and-more-useful-than-siri/

    [Geoffrey Litt] shows that getting an effective digital assistant that’s tailored to one’s own needs just needs a little DIY, and thanks to the kinds of tools that are available today, it doesn’t even have to be particularly complex. Meet Stevens, the AI assistant who provides the family with useful daily briefs. The back end? Little more than one SQLite table and a few cron jobs.

    Stevens: a hackable AI assistant using a single SQLite table and a handful of cron jobs
    https://www.geoffreylitt.com/2025/04/12/how-i-made-a-useful-ai-assistant-with-one-sqlite-table-and-a-handful-of-cron-jobs

    There’s a lot of hype these days around patterns for building with AI. Agents, memory, RAG, assistants—so many buzzwords! But the reality is, you don’t need fancy techniques or libraries to build useful personal tools with LLMs.

    In this short post, I’ll show you how I built a useful AI assistant for my family using a dead simple architecture: a single SQLite table of memories, and a handful of cron jobs for ingesting memories and sending updates, all hosted on Val.town. The whole thing is so simple that you can easily copy and extend it yourself.

    Meet Stevens

    The assistant is called Stevens, named after the butler in the great Ishiguro novel Remains of the Day. Every morning it sends a brief to me and my wife via Telegram, including our calendar schedules for the day, a preview of the weather forecast, any postal mail or packages we’re expected to receive, and any reminders we’ve asked it to keep track of. All written up nice and formally, just like you’d expect from a proper butler.

    Reply
  10. Tomi Engdahl says:

    US tariffs on chips may slow AI race as firms scramble to localize, costs rise, and global supply chains shift. https://link.ie.social/3zcOXH

    Reply
  11. Tomi Engdahl says:

    Igor Bonifacic / Engadget:
    OpenAI launches o3, its most advanced reasoning model, and o4-mini, a lower cost alternative that still delivers “impressive results”, for ChatGPT paid users — ChatGPT Plus users can begin using both new models starting today. — A mere two days after announcing GPT-4.1, OpenAI is releasing not one but two new models.

    OpenAI’s new o3 and o4-mini models are all about ‘thinking with images’
    ChatGPT Plus users can begin using both new models starting today.
    https://www.engadget.com/ai/openais-new-o3-and-o4-mini-models-are-all-about-thinking-with-images-170043465.html?_fsig=CFotqcb_8sliZqtYP94ArQ–%7EA

    A mere two days after announcing GPT-4.1, OpenAI is releasing not one but two new models. The company today announced the public availability of o3 and o4-mini. Of the former, OpenAI says o3 is its most advanced reasoning model yet, with it showing “strong performance” in coding, math and science tasks. As for o4-mini, OpenAI is billing it as a lower cost alternative that still delivers “impressive results” across those same fields.

    More notably, both models offer novel capabilities not found in OpenAI’s past systems. For first time, the company’s reasoning models can use and combine all of the tools available in ChatGPT, including web browsing and image generation. The company says this capability allows o3 and o4-mini solve challenging, multi-step problems more effectively, and “take real steps toward acting independently.”

    Reply
  12. Tomi Engdahl says:

    Kyle Wiggers / TechCrunch:
    OpenAI debuts Codex CLI, an open-source coding agent tool for terminals that links AI models with local code and computing tasks, like moving files — In a bid to inject AI into more of the programming process, OpenAI is launching Codex CLI, a coding “agent” designed to run locally from terminal software.

    OpenAI debuts Codex CLI, an open source coding tool for terminals
    https://techcrunch.com/2025/04/16/openai-debuts-codex-cli-an-open-source-coding-tool-for-terminals/

    In a bid to inject AI into more of the programming process, OpenAI is launching Codex CLI, a coding “agent” designed to run locally from terminal software.

    Announced on Wednesday alongside OpenAI’s newest AI models, o3 and o4-mini, Codex CLI links OpenAI’s models with local code and computing tasks, OpenAI says. Via Codex CLI, OpenAI’s models can write and edit code on a desktop and take certain actions, like moving files.

    Codex CLI appears to be a small step in the direction of OpenAI’s broader agentic coding vision. Recently, the company’s CFO, Sarah Friar, described what she called the “agentic software engineer,” a set of tools OpenAI intends to build that can take a project description for an app and effectively create it and even quality assurance test it.

    https://github.com/openai/codex

    Reply
  13. Tomi Engdahl says:

    Dan Shipper / Every:
    OpenAI o3 review: it’s “deep research-lite”, giving thorough answers quickly, free from some old ChatGPT limitations, and powerful thanks to ChatGPT integration

    Vibe Check: o3 Is Here—And It’s Great

    The highest praise I can give is that I’m already using it all the time

    https://every.to/chain-of-thought/vibe-check-o3-is-out-and-it-s-great

    OpenAI:
    OpenAI says o3 and o4-mini represent a significant breakthrough in visual perception by reasoning with images in their chain of thought
    https://openai.com/index/thinking-with-images/

    Reply
  14. Tomi Engdahl says:

    Hayden Field / CNBC:
    OpenAI says o3 and o4-mini are its first reasoning models that can use all ChatGPT tools, including web browsing, Python, and image generation

    OpenAI says newest AI model can ‘think with images,’ understanding diagrams and sketches
    https://www.cnbc.com/2025/04/16/openai-releases-most-advanced-ai-model-yet-o3-o4-mini-reasoning-images.html

    OpenAI released its newest AI model that it said can understand uploaded images like whiteboards, sketches and diagrams, even if they’re low quality.
    The company called o3 its most advanced model yet and also released a smaller model called o4-mini.
    OpenAI is racing to stay ahead in generative AI as competitors including Google, Anthropic and Elon Musk’s xAI ramp up development.

    OpenAI has released its latest artificial intelligence model, which it said is capable of “thinking with images,” meaning it can understand and analyze a user’s sketches and diagrams, even if they’re low quality.

    The main new reasoning model from OpenAI is called o3, and the company simultaneously released a smaller model dubbed o4-mini. The rollout follows the September debut of OpenAI’s first reasoning model, o1, which focused on solving complex problems and deliberating over its answers in multiple steps.

    With o3, users can upload whiteboards, sketches and other images and have the AI analyze and discuss them. The models can also rotate, zoom and use other image-editing tools.

    Reply
  15. Tomi Engdahl says:

    Bloomberg:
    Source: OpenAI is in talks to acquire Windsurf, an AI coding tool formerly known as Codeium, for ~$3B; Windsurf was valued at $1.25B in a 2024 funding deal — OpenAI is in talks to acquire Windsurf, an artificial intelligence-assisted coding tool formerly known as Codeium, for about $3 billion …

    OpenAI In Talks to Buy Windsurf for About $3 Billion
    https://www.bloomberg.com/news/articles/2025-04-16/openai-said-to-be-in-talks-to-buy-windsurf-for-about-3-billion?accessToken=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJzb3VyY2UiOiJTdWJzY3JpYmVyR2lmdGVkQXJ0aWNsZSIsImlhdCI6MTc0NDgyNzA2MywiZXhwIjoxNzQ1NDMxODYzLCJhcnRpY2xlSWQiOiJTVVRNVk1UMEcxS1cwMCIsImJjb25uZWN0SWQiOiIyMjNDRDM2NDg0QzY0OTc3QjY5ODE0Rjc1MTYxNDRGNyJ9.Z_OoPEylsGdZvYcyGZQwzeLvlE4OOvP4isdtcS1OLO8&leadSource=uverify%20wall

    OpenAI is in talks to acquire Windsurf, an artificial intelligence-assisted coding tool formerly known as Codeium, for about $3 billion, according to a person familiar with the matter.

    The deal would be OpenAI’s largest acquisition to date, and could help the company take on rising competition in the market for AI-driven coding assistants — systems capable of tasks like writing code based on natural language prompting. The terms of the deal haven’t been finalized and talks could still change or fall apart, said the person, who asked not to be identified because the matter is private.

    Reply
  16. Tomi Engdahl says:

    Samantha Subin / CNBC:
    Tech stocks plunge amid uncertainty over US tariff plans and China export controls: AMD, Nvidia, and ASML dropped 7%, Applied Materials and Lam Research fell 5% — Technology stocks dropped Wednesday, led by a 7% plunge in Nvidia, as the chipmaking sector signaled that President Donald …
    https://www.cnbc.com/2025/04/16/chip-stocks-fall-as-nvidia-amd-warn-of-china-export-control-costs.html

    Reply
  17. Tomi Engdahl says:

    Abner Li / 9to5Google:
    Google is making Gemini Live camera and screen sharing features free for all Android users in the coming weeks after launching them for Advanced users last week — Last week, Google started widely rolling out Gemini Live camera and screen sharing to Advanced subscribers …

    Google makes Gemini Live camera & screen sharing free on Android
    https://9to5google.com/2025/04/16/gemini-live-camera-free/

    Last week, Google started widely rolling out Gemini Live camera and screen sharing to Advanced subscribers, and the Project Astra-powered capabilities will soon be free for all Android users.

    Gemini Live now lets you ask questions about what’s on your screen or camera. Screen sharing can be quickly initiated by launching the Gemini overlay and tapping the new “Share screen with Live” chip. After confirming, you’ll see a count next to the time in your status bar. Google has since rolled out the new phone call-style notification for Live.

    You can scroll an app or website, while talking to Gemini Live. There’s a subtle vibration before Gemini Live starts responding. Pull down your notification shade at any time to “Stop sharing.”

    To launch the camera, open the fullscreen Gemini Live for the new button at the left (screen sharing can also be initiated from here). A viewfinder then appears, with a button in the bottom-right corner letting you switch to the front-facing lens.

    Gemini says: “For better results, capture objects with steady movements.” The display must be active for Gemini Live to capture video.

    Camera and screen sharing joins how you can talk to Gemini Live about an image, PDF, or YouTube video.

    Gemini Live’s Astra-powered visual sharing was first announced at I/O 2024 last May. It started rolling out in March, with wide availability for Gemini Advanced coming last week.

    Gemini app widely rolling out ‘Talk Live about’ images, files, and YouTube
    https://9to5google.com/2025/04/04/gemini-talk-live-about/

    Reply
  18. Tomi Engdahl says:

    Bloomberg:
    Sources: Perplexity is in discussions to integrate its AI assistant on Samsung’s devices and has reached an agreement to preload Perplexity on Motorola’s phones — Perplexity AI Inc., an artificial intelligence startup vying with Google and OpenAI, is ramping up efforts to get prime placement for its virtual assistant on smartphones.

    https://www.bloomberg.com/news/articles/2025-04-16/perplexity-in-talks-to-integrate-assistant-into-samsung-motorola-phones

    Reply
  19. Tomi Engdahl says:

    Larry Dignan / Constellation Research:
    Docusign unveils AI contract agents, which can analyze agreements and surface issues that usually require human intervention, available by the end of 2025 — Docusign launched AI contract agents, which will analyze agreements, highlight risks and surface issues that usually require human intervention.

    https://www.constellationr.com/blog-news/insights/docusign-launches-ai-contract-agents

    Reply
  20. Tomi Engdahl says:

    Reuters:
    Deezer says ~18% of songs uploaded to the service are AI-generated; 20K AI-generated songs are uploaded daily, nearly twice the number reported four months ago

    AI-generated music accounts for 18% of all tracks uploaded to Deezer
    https://www.reuters.com/technology/artificial-intelligence/ai-generated-music-accounts-18-all-tracks-uploaded-deezer-2025-04-16/

    Reply
  21. Tomi Engdahl says:

    Tom Warren / The Verge:
    Microsoft adds a Copilot Studio “computer use” tool that lets AI agents interact with websites and desktop apps and automatically adapts to changes in the apps

    Microsoft lets Copilot Studio use a computer on its own
    Microsoft has its own ‘computer use’ feature for AI agents to automatically use websites and apps.
    https://www.theverge.com/news/649574/microsoft-copilot-studio-computer-use-ai

    Microsoft has enabled a new “computer use” feature for Copilot Studio this week that lets AI agents interact with websites and desktop applications. Much like OpenAI’s Operator or Claude’s identically named “computer use” feature, businesses will be able to use Copilot Studio to build AI agents that treat websites and desktop apps as if they’re tools for automating tasks.

    “Computer use enables agents to interact with websites and desktop apps by clicking buttons, selecting menus, and typing into fields on the screen,” explains Charles Lamanna, corporate vice president of Microsoft’s business & industry Copilot. “This allows agents to handle tasks even when there is no API available to connect to the system directly. If a person can use the app, the agent can too.”

    Copilot Studio will be able to create AI agents that can automate data entry, perform market research, or even process invoices. This new tool will even detect when buttons and screens change in apps or websites and continue working without failing or breaking, according to Microsoft.

    Microsoft added a similar feature, dubbed Actions, to its consumer Copilot earlier this month. Actions can be performed in the background while you work on other tasks. Copilot can now do things like book restaurant reservations, event tickets, and purchase items from online stores.

    Microsoft Copilot can now use the web on your behalf
    Copilot’s Actions feature can book car rentals, concert tickets, and restaurant tables so you don’t have to.
    https://www.theverge.com/news/643276/microsoft-copilot-ai-actions-feature-availability

    Microsoft is making changes to Copilot that will allow the artificial intelligence assistant to complete online tasks for you. Using “simple chat prompts,” these “Actions” can be performed in the background while you work on other tasks, according to Microsoft, and can do things like book restaurant reservations, event tickets, and purchase items to ship to your friends.

    Launch partners include Booking.com, Expedia, Kayak, Tripadvisor, Skyscanner, Viator, Vrbo, and Priceline — a host of travel and vacation-focused services that will allow Copilot to help users plan any upcoming trips — alongside Open Table for meals, and 1-800-Flowers.com for buying floral arrangements. Microsoft says that Copilot’s Actions feature will “work with most websites across the web,” and can sort out “the ride home” for you after events, suggesting it will also support taxi or ride-sharing services.

    Reply
  22. Tomi Engdahl says:

    Dylan Patel / SemiAnalysis:
    A look at Huawei’s AI CloudMatrix 384 rack-scale system, which is less power-efficient than Nvidia’s GB200 NVL72, though this is not a limiting factor in China

    Huawei AI CloudMatrix 384 – China’s Answer to Nvidia GB200 NVL72 China Abundance of Power, 100% Optics, 0% Copper, Power Inefficiency, 2.6x lower FLOP per Watt, 14 Transceivers per Chip, Linear Pluggable Optics
    https://semianalysis.com/2025/04/16/huawei-ai-cloudmatrix-384-chinas-answer-to-nvidia-gb200-nvl72/

    Reply
  23. Tomi Engdahl says:

    Bill Gates says AI can help solve worker shortages in 2 surprising professions
    https://www.businessinsider.com/bill-gates-ai-job-shortages-doctors-teachers-work-free-time-2025-4

    Bill Gates said AI could solve shortages in two key professions: teaching and medicine.
    The billionaire said AI would help plug labor gaps, even in blue-collar roles.
    He also said AI could make early retirement or shorter workweeks possible.

    Reply
  24. Tomi Engdahl says:

    Meet The AI Agent With Multiple Personalities
    A new AI agent from the startup Simular switches between different AI models depending on the task at hand.
    https://www.wired.com/story/simular-ai-agent-multiple-models-personalities/

    Reply
  25. Tomi Engdahl says:

    Top 10 Vibe Coding Tools That Feel Like Magic in 2025
    #
    webdev
    #
    vibecoding
    #
    programming
    AI is no longer just assisting developers—it’s transforming how we build software. Whether you’re a solo indie hacker, a startup engineer, or part of an enterprise team, there’s now an AI tool that aligns with your workflow. The best ones don’t simply automate tasks; they feel intuitive, context-aware, and remarkably efficient.

    https://dev.to/therealmrmumba/top-10-vibe-coding-tools-that-feel-like-magic-in-2025-1md

    Software development has evolved. We’ve moved past the era of simple code autocomplete and template generators. Today, developers demand tools that integrate deeply with their tech stack, adapt to their workflow, and help them ship faster without compromising quality.

    Here are three key trends that shape this new landscape:

    Autonomous Assistance – Developers now rely on AI tools that can complete multi-step tasks with minimal guidance, from writing modules to fixing bugs.
    Context-Aware Coding – Tools that understand your full codebase, API structure, and data flow are no longer a luxury—they’re a necessity.
    Real-Time Collaboration – AI is becoming a collaborative partner in the development process, offering suggestions, generating code, and even reviewing pull requests in real-time.

    The Top 10 Vibe Coding Tools in 2025
    1. Fine by Fine.dev
    2. Cursor
    3. Copilot Workspace
    4. Sweep
    5. 1. Apidog MCP Server
    6. Continue
    7. The Windsurf Editor by Codeium
    Built to Keep you in flow state
    8. Lazy AI
    No-code app creation with AI-powered workflows
    9. Devika
    Your open-source AI software engineer
    10. Cody by Sourcegraph
    Deep code understanding with full repo context

    Final Thoughts
    2025 isn’t about whether you use AI as a developer—it’s about how. The tools listed above aren’t just trendy—they’re reshaping the development experience by embedding intelligence into your daily workflow. Some extend your capabilities. Others remove tedious work. The best ones do both.

    Choose the ones that match your stack, integrate with your workflow, and make you feel like a better developer—not just a faster one.

    Your open-source AI software engineer

    Reply
  26. Tomi Engdahl says:

    OpenAI Releases Codex CLI: An Open-Source Local Coding Agent that Turns Natural Language into Working Code
    https://www.marktechpost.com/2025/04/16/openai-releases-codex-cli-an-open-source-local-coding-agent-that-turns-natural-language-into-working-code/

    Tell Codex CLI what to build, fix, or explain, then watch it bring your ideas to life

    Codex CLI: Bridging Natural Language and Code
    To mitigate these challenges, OpenAI has introduced Codex CLI, an open-source tool designed to operate within terminal environments. Codex CLI enables users to input natural language commands, which are then translated into executable code by OpenAI’s language models. This functionality allows developers to perform tasks such as building features, debugging code, or understanding complex codebases through intuitive, conversational interactions. By integrating natural language processing into the CLI, Codex CLI aims to streamline development workflows and reduce the cognitive load associated with traditional command-line operations.​

    Technical Overview and Benefits
    Codex CLI leverages OpenAI’s advanced language models, including the o3 and o4-mini, to interpret user inputs and execute corresponding actions within the local environment. The tool supports multimodal inputs, allowing users to provide screenshots or sketches alongside textual prompts, enhancing its versatility in handling diverse development tasks. Operating locally ensures that code execution and file manipulations occur within the user’s system, maintaining data privacy and reducing latency. Additionally, Codex CLI offers configurable autonomy levels through the –approval-mode flag, enabling users to control the extent of automated actions, ranging from suggestion-only to full auto-approval modes. This flexibility allows developers to tailor the tool’s behavior to their specific needs and comfort levels.

    While Codex CLI represents a significant step towards AI-assisted development, OpenAI envisions further evolution towards an “agentic software engineer”—a comprehensive system capable of autonomously handling entire development cycles, from conception to deployment. This future direction suggests a continued focus on enhancing the capabilities of AI tools to support more complex and integrated development tasks.​

    Conclusion
    OpenAI’s Codex CLI offers a novel approach to software development by integrating natural language processing into terminal-based workflows. By translating conversational inputs into executable code, it simplifies interactions with complex systems and reduces the barriers associated with traditional CLI usage.

    Reply
  27. Tomi Engdahl says:

    OpenAI is building a social networkIs Sam Altman ready to up his rivalry with Elon Musk and Mark Zuckerberg?
    https://www.theverge.com/openai/648130/openai-social-network-x-competitor

    Reply
  28. Tomi Engdahl says:

    Build Scalable LLM Apps With Kubernetes: A Step-by-Step Guide
    Understanding how to scale AI apps efficiently is the difference between a model stuck in research and one delivering actionable results in production.
    https://thenewstack.io/build-scalable-llm-apps-with-kubernetes-a-step-by-step-guide/

    Reply
  29. Tomi Engdahl says:

    Dozens of tech firms now let their AI agents work together on complex tasks
    Google-led Agent2Agent protocol, which is already being used by PayPal, Salesforce and Workday, lets AI assistants communicate to accomplish tasks
    https://thelogic.co/news/agent2agent-ai-google-cohere-servicenow/

    Reply
  30. Tomi Engdahl says:

    AI agents in SEO: What you need to know
    Behind the scenes of SEO’s next evolution – where agents navigate filters, product feeds, and even send Slack alerts.
    https://searchengineland.com/ai-agents-seo-454279

    Reply
  31. Tomi Engdahl says:

    Traditional RAG Frameworks Fall Short: Megagon Labs Introduces ‘Insight-RAG’, a Novel AI Method Enhancing Retrieval-Augmented Generation through Intermediate Insight Extraction
    https://www.marktechpost.com/2025/04/14/traditional-rag-frameworks-fall-short-megagon-labs-introduces-insight-rag-a-novel-ai-method-enhancing-retrieval-augmented-generation-through-intermediate-insight-extraction/

    Reply
  32. Tomi Engdahl says:

    Sam Altman says OpenAI is no longer “compute-constrained” — after Microsoft lost its exclusive cloud provider status
    https://www.windowscentral.com/software-apps/sam-altman-says-openai-is-no-longer-compute-constrained

    Reply
  33. Tomi Engdahl says:

    Microsoft researchers say they’ve developed a hyper-efficient AI model that can run on CPUs
    https://techcrunch.com/2025/04/16/microsoft-researchers-say-theyve-developed-a-hyper-efficient-ai-model-that-can-run-on-cpus/

    Microsoft researchers claim they’ve developed the largest-scale 1-bit AI model, also known as a “bitnet,” to date. Called BitNet b1.58 2B4T, it’s openly available under an MIT license and can run on CPUs, including Apple’s M2.

    Bitnets are essentially compressed models designed to run on lightweight hardware. In standard models, weights, the values that define the internal structure of a model, are often quantized so the models perform well on a wide range of machines. Quantizing the weights lowers the number of bits — the smallest units a computer can process — needed to represent those weights, enabling models to run on chips with less memory, faster.

    Bitnets quantize weights into just three values: -1, 0, and 1. In theory, that makes them far more memory- and computing-efficient than most models today.

    The Microsoft researchers say that BitNet b1.58 2B4T is the first bitnet with 2 billion parameters, “parameters” being largely synonymous with “weights.” Trained on a dataset of 4 trillion tokens — equivalent to about 33 million books, by one estimate — BitNet b1.58 2B4T outperforms traditional models of similar sizes, the researchers claim.

    BitNet b1.58 2B4T doesn’t sweep the floor with rival 2 billion-parameter models, to be clear, but it seemingly holds its own. According to the researchers’ testing, the model surpasses Meta’s Llama 3.2 1B, Google’s Gemma 3 1B, and Alibaba’s Qwen 2.5 1.5B on benchmarks including GSM8K (a collection of grade-school-level math problems) and PIQA (which tests physical commonsense reasoning

    Reply
  34. Tomi Engdahl says:

    GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet
    14th April 2025

    OpenAI introduced three new models this morning: GPT-4.1, GPT-4.1 mini and GPT-4.1 nano. These are API-only models right now, not available through the ChatGPT interface (though you can try them out in OpenAI’s API playground). All three models can handle 1,047,576 tokens of input and 32,768 tokens of output, and all three have a May 31, 2024 cut-off date (their previous models were mostly September 2023).

    https://simonwillison.net/2025/Apr/14/gpt-4-1/

    Reply
  35. Tomi Engdahl says:

    How AI will change software engineering
    No one can agree how AI will change software engineering. Here’s how to prepare anyway.
    https://leaddev.com/career-development/how-ai-will-change-software-engineering

    Reply
  36. Tomi Engdahl says:

    Google DeepMind Is Hiring a ‘Post-AGI’ Research Scientist
    Emanuel Maiberg
    Emanuel Maiberg
    ·
    Apr 15, 2025 at 11:03 AM
    Google is preparing for a future with AGI, ASI, and machine consciousness.
    https://www.404media.co/google-deepmind-is-hiring-a-post-agi-research-scientist/

    Reply
  37. Tomi Engdahl says:

    “Anybody who claims to have any confidence about where AI is going to be more than a year or two out, the more confident they claim to be, the less you should believe them,” said Charity Majors, cofounder of the observability platform Honeycomb. “There’s so much hype,” she said. “Honestly, if you zoom out a little bit, so much of ‘AI’ you could just replace with ‘automation’.”
    https://leaddev.com/career-development/how-ai-will-change-software-engineering

    Reply
  38. Tomi Engdahl says:

    DeepMind’s New AI Teaches Itself to Play Minecraft From Scratch
    The AI made a “mental map” of the world to collect the game’s most sought-after material.
    https://singularityhub.com/2025/04/11/deepminds-new-ai-teaches-itself-to-play-minecraft-from-scratch/

    Reply

Leave a Comment

Your email address will not be published. Required fields are marked *

*

*