News & Intel
LiveDaily AI-curated intelligence on Palantir Foundry, Ontology, AIP, Apollo, contracts, and community feedback. Updated automatically via GitHub Actions every day at 7 AM UTC.
Better way to log function output and make it redable for LLM
Hi, I have been writing a lot of functions, thanks to AI-FDE, and the MCP server that I can configure with my IDE With the current tools, its become easy to write, commit, wait for CI/Checks and fix the function once the function is working, I can publish it (tag it ) then use is with front end when I have to debug a function that is exposed as a function or a action, I can check the logs in workflow tool, that shows the logs if I add logging to the function But how Im trying to figure out if there is a mcp function where I can prompt the following in my IDE or AI-FDE or continue to say ” check the recent logs for action “create-User” and tell me what input was received and what action really happened ? “ 1 post - 1 participant Read full topic
Exclusive: Andreessen Horowitz backs Deeptune’s $43M Series A to build ‘training gyms’ for AI agents
The company is betting that simulated environments—not more web data—will be key to training the next generation of AI agents.
AIP Document Intelligence for free developer tier
I have had an AIP developer tier (Free) account for a few months. I do not see AIP Document intelligence inside of foundry even though the February 2026 release notes indicate: "AIP Document Intelligence will be generally available on February 4, 2026 and is enabled by default for all AIP enrollments” Is this a bug or when will this be available? 2 posts - 2 participants Read full topic
Jensen Huang just painted the most bold image of AI’s future: 7.5 million agents, 75,000 humans—100 AI workers for every person
The Nvidia CEO said AI agents would work around the clock so human workers don’t have to keep up with them.
Dev Tier: LLM Rate-limiting
I have a dev tier account, and when I use tools like AI-FDE, I feel the rate limits are low enough to prohibit productive development. I appreciate this is a free service, so I want to understand if there’s a way for me to use my own Claude API credentials for AIP? That way I get billed for it directly. Or if there’s a way for me to set limits and just pay per usage, stilling keeping the other platform limitations (user count etc.) such that I won’t actually go to production with it. Is this on the roadmap? 1 post - 1 participant Read full topic
Using Compute Module functions in AIP logic
Hi, I am trying to call two functions from my compute module into AIP Logic. I’m getting an error stating that compute module functions are registering as ‘statefulFunction type which is unsupported in scoped mode’ in the Logic App. I’ve created a function container inside the compute module that calls Flask (a separate scoring container in the same module) over localhost, is there a way to make those functions register as standard functions so Logic can call them? Or should the @function wrappers live in a standard functions repo and call the container over an external HTTP route instead, and if so what is the hostname/port for the container? 1 post - 1 participant Read full topic
AIG, McGill Announce Collaboration to Potentially Transform Subscription Market
McGill and Partners and American International Group said they collaborating to perhaps disrupt the subscription market—using agentic artificial intelligence to deploy capital to clients. AIG said the digital-first platform of the London-based independent bro…
'So much of the work is going to be done by agents': ServiceNow CEO says graduate unemployment could reach 30% because of AI agents
90% of human customer service agent use cases can be replaced by AI, ServiceNow CEO Bill McDermott explains.
NVIDIA Ignites the Next Industrial Revolution in Knowledge Work With Open Agent Development Platform
NVIDIA Agent Toolkit Equips Enterprises to Build and Run AI AgentsNews Summary: NVIDIA Agent Toolkit includes NVIDIA OpenShell open source runtime for building self-evolving agents and claws ...
Daily Hacker News for 2026-03-16
The 10 highest-rated articles on Hacker News on March 16, 2026 which have not appeared on any previous Hacker News Daily are: <ul> <li> LLM Architecture Gallery (comments) </li> <li> Stop Sloppypasta (comments) </li> <li> The 49MB web page (comments) </li> <l…
NVIDIA Debuts Agent Toolkit and NemoClaw at GTC for Faster, Safer AI Agents
During his keynote at GTC 2026, NVIDIA CEO Jensen Huang introduced a wide array of new products, services, AI models, and software to a massive audience at the SAP Center in San Jose, California. The Vera Rubin ecosystem with Groq integration made its debut, …
AI-FDE Core Architecture Library
Palantir explicitly structures AI FDE around two tiers: Modes (the broad task at hand, e.g., data integration or ontology editing) and Skills (granular capabilities usable across multiple modes). This GitHub repository library mirrors this exact two-tier hierarchy — skill.md maps to Palantir’s Skills layer, agents.md maps to full Mode-level agent configurations. Your prompt library will sit as the execution layer beneath both. https://github.com/s-andthat/palantir-ai-fde-library Hopefully someone will find this useful. 1 post - 1 participant Read full topic
AI agents could easily send college grad unemployment over 30%, ServiceNow CEO says
Technology companies, including Block and Atlassian, have already cut jobs this year due to AI adoption.
Using LLM Models in Functions
Hi All, I was trying to do some summarisation in Functions, and I am wondering why I can bring in LLM models such as GPT-5 in Typescript v1 (Via importing models), however I can’t find a way to do it in Typescript V2, and Python. I am wondering if it’s possible to do this? Also, I understand there are ways around this, such as creating an AIP Logic, which would make more sense. I am just interested in why it isn’t possible? Kind Regards Sam 2 posts - 2 participants Read full topic
RemoteException: INTERNAL (Default:Internal) when writing embeddings to Ontology object via AIP Logic Action
Problem I’m building an AIP Logic pipeline that processes a PDF and creates a Knowledge Article object. The workflow is: Extract text from a PDF media reference Use an LLM to generate structured fields (Title, Short Description, Content, Category, Tags, Key Points, Content Type, Entities) Generate text embeddings using text-embedding-ada-002 Convert the embedding output to Array Call a Create Knowledge Articles ontology action to write the data to an object All upstream blocks execute successfully during preview (PDF extraction, LLM output, embedding generation). However, the final Ontology Action fails. Error RemoteException: INTERNAL (Default:Internal) Trace ID: 11896d79c7214279 Error ID: b4d93c5-95df-47fb-a560-f765603daa8b The failure occurs specifically in the Create Knowledge Articles action block. In the debugger, the action shows: Arguments: {} Even though all fields appear correctly mapped in
New NVIDIA Nemotron 3 Super Delivers 5x Higher Throughput for Agentic AI
Launched today, NVIDIA Nemotron 3 Super is a 120‑billion‑parameter open model with 12 billion active parameters designed to run complex agentic AI systems at scale. Available now, the model combines advanced reasoning capabilities to efficiently complete tas…
Nvidia’s Nemotron Super 3 model for agentic systems launches with five-times higher throughput
With so much talk about its upcoming Vera Rubin graphics processing units, it’s easy to forget that Nvidia Corp. doesn’t just supply the hardware for artificial intelligence. It also develops its own series of AI models, and today it announced the availabilit…
Is AI our agent, or are our governments becoming agents for AI?
Yet more billions are being spent on agentic AI, despite warnings of its potentially extreme fallibility. Just who are governments serving when they spout the messaging of Big Tech companies?