Skip to content

Latest_Editorials

VIEW_ALL →

Latest_Videos

VIEW_ALL →

Latest_Articles

VIEW_ALL →
#NewsArticle

Google is making it easier to import another AI’s memory into Gemini

Google is rolling out two new Gemini tools on Thursday which could attract users who are thinking about switching their preferred AI chatbot but don't want to "retrain" a new AI. Gemini's new "Import Memory" and "Import Chat History" features help users quickly copy over everything their current AI already knows about them with a couple of pre-written prompts. To use the "Import Memory" tool, users copy and paste a suggested prompt from Gemini into their previous AI, then paste the output from the previous AI into Gemini, which should get Gemini caught up on their preferences. The "Import Chat History" feature has users request an export o … Read the full story at The Verge.

src: The Verge AIOPEN_SOURCE
#NewsArticle

Apple will reportedly allow other AI chatbots to plug into Siri

Siri currently has an integration with OpenAI’s ChatGPT. | Photo: Allison Johnson / The Verge Apple's iOS 27 update will allow users to choose the AI chatbot they want to link with Siri. That's according to a report from Bloomberg's Mark Gurman, who says third-party chatbots downloaded from the App Store, like Google's Gemini or Anthropic's Claude, will be able to fetch replies for Siri - similar to how the voice assistant already works with OpenAI's ChatGPT. The new system, reportedly called "Extensions," will give users the ability to enable or disable the chatbots they want to connect with Siri on the iPhone, iPad, and Mac, Bloomberg reports. The upcoming integrations will also work with the standalone app Apple reportedly plans … Read the full story at The Verge.

src: The Verge AIOPEN_SOURCE
#DiscussionReddit

Why prompt engineer never became a real job?

around late 2024 I saw on various different websites, first about how agents will take over, which has happened in a way, though I will say different from how it was advertised, and second how prompt engineer will become the new staple tech job. given how claude is the most personality synergising chatbot and trained on undisclosed methods that anthropic keeps as their business secrets, and there's still a massive gap how people expect their prototype to look like vs what it actually gets stitched togther by llm. there are tons of other jobs that sound way less believable like thumbnail designer and social media intern, and everytime something gets accomplished ever there are some dudes pointing out how the poster could've done it more efficiently. just why isn't there the word wizard that figures out how to change prompt into what each person needs, saving them bunch of time and tokens, and no the clowns tooting their horn about 10x productivity prompts don't count, as they're surface

src: r/ClaudeAIOPEN_SOURCE
#DiscussionReddit

By What Year will AGI Arrive - Poll

It's 2026 so here is the obligatory AGI poll. By what year do you predict AGI? I'll use the definition for AGI that I used in previous polls. The definition of AGI for this poll: an AI capable of learning to accomplish any intellectual task that humans or animals can perform. Alternatively, any autonomous system that surpasses human capabilities in the majority of economically valuable tasks. My last poll was December 2024. Amazingly, more than a fifth of respondents though we'd have AGI by the above definition by 2025. Obviously, that did not happen, but we're fast approaching some dates popularised by the likes of Ray Kurzweil. [View Poll](https://www.reddit.com/poll/1s4kfhl)

src: r/singularityOPEN_SOURCE
#DiscussionReddit

Apple stopped selling 512gb URAM mac studios, now the max amount is 256GB!

THe memory supply crisis is hitting apple too. IT is probably too expensive and/or not enough supply for them to sell 512gb ram m3 ultras. U can look at [https://www.apple.com/shop/buy-mac/mac-studio](https://www.apple.com/shop/buy-mac/mac-studio) to see it is no longer available.. MAybe that is why the m5 max only has a max of 128gb, i think they couldve added 256gb to it... Yeah they probably wont make the m5 ultra with 1tb of ram; at best 512 gb of ram, maybe even only 256 gb of ram...

src: r/LocalLLaMAOPEN_SOURCE
#DiscussionReddit

Which system for 2x RTX 6000 blackwell max-q

I am trying to decide which system to run these cards in. 1) Supermicro X10Dri-T, 2x E5-2699v4, 1TB ddr4 ecc ram (16x 64GB lrdimm 2400mhz), PCI-E 3.0 slots 2) Supermicro X13SAE-F, i9-13900k, 128GB ddr5 ecc ram (4x 32GB udimm 4800mhz), PCI-E 5.0 slots For ssds I have 2x Micron 9300 Pro 15.36TB. I haven't had much luck with offloading to the cpu/ram on the 1TB ddr4. Probably can tweak it up a little. For the large models running just on cpu I get 1.8 tok/s (still impressive they even run at all). So question is: Is there any point in trying to offload to ram? or just go for the higher pci 5 speed?

src: r/LocalLLaMAOPEN_SOURCE
#DiscussionReddit

Am I expecting too much?

Hi there, I work in the IT department of a financial industry and dabbled with creating our local ai. I got the following requirements: \-Local AI / should be able to work as an assistant (so give a daily overview etc) / be able to read our data from clients without exposing it to the outside As far as I understand, I can run LlaMA on a Mac Studio inside our local network without any problems and will be able to connect via MCP to Powerbi, Excel and Outlook. I wanted to expose it to Open Web UI, give it a static URl and then let it run (would also work when somebody connects via VPN to the server) . I was also asked to be able to create an audit log of the requests (so which user, what prompts, documents, etc). Claude gave me this: nginx reverse proxy , which I definetly have to read into. Am I just babbled by the AI Hype or is this reasonable to run this? (Initially with 5-10 users and then upscale the equipment maybe? for 50)

src: r/LocalLLaMAOPEN_SOURCE
#DiscussionReddit

The "Preamble" Problem: How do you actually force an LLM to output RAW text only?

​ I am struggling with a persistent issue across Llama.cpp-qwen3.5—where they won't stop adding introductory and concluding "fluff." Even when I explicitly command the model to provide the result and nothing else, I still get hit with "Here is your summary..." or "Note: The following changes were made..." This is becoming a major headache for automation. I’m currently working on two specific use cases where this extra text breaks everything: \* . Despite telling the model: "Do not provide any output outside of the sentence format" and "Do not give me opening lines like 'Here is your phrass...'", it still prepends "Here's my attempt at creating a sentence ..." This ruins the script's ability to parse the file directly. \* Text Readability Reformatting: I'm using qwen3.5 generare sentence for tts. I’ve tried a 10-point instruction list, where point #10 is literally: "Answer back the revised text without additional comments." It is completely ignored. What's weirder i

src: r/LocalLLaMAOPEN_SOURCE
#DiscussionReddit

Unsloth says MLX fine-tuning is coming early next month: this could be huge for local AI

Yesterday, the Unsloth dev actually responded to my question over in r/unsloth and confirmed that MLX fine-tuning support is expected sometime early next month in unsloth studio. If they actually nail this and ship it properly, it’s going to be a pretty huge moment for anyone doing local AI work on MacBooks and Mac Studios. Up until now, those of us on Apple Silicon have mostly been stuck doing inference and complicated mlx training demos. Proper training and fine-tuning has always felt like the missing layer on these machines, which is a shame considering how much raw unified memory and efficiency they pack. If this lands well, it feels like it could unlock a true end-to-end local workflow. Obviously, this isn't going to suddenly replace serious NVIDIA setups for large-scale training. The interesting shift is just how much more we'll realistically be able to do locally. Less dependency on cloud compute, and a lot more freedom to just build and experiment. Personally, I’m running 2

src: r/LocalLLaMAOPEN_SOURCE
#TutorialsArticle

Gap-Up and Fade Is Not Random

A Quantitative AI Framework for Identifying Stocks Prone to Overnight Strength and Intraday Reversal Continue reading on Towards AI »

src: Towards AIOPEN_SOURCE
#DiscussionArticle

Show HN: Fio: 3D World editor/game engine – inspired by Radiant and Hammer

A liminal brush-based CSG editor and game engine with unified (forward) renderer inspired by Radiant and Worldcraft/Hammer Compact and lightweight (target: Snapdragon 8CX, OpenGL 3.3) Real-time lighting with stencil shadows without the need for pre-baked compilation Comments URL: https://news.ycombinator.com/item?id=47535626 Points: 5 # Comments: 0

src: Hacker News Front PageOPEN_SOURCE
#DiscussionReddit

Just got Pro... why do the limits seem worse?

Title. Simple questions using Sonnet (non-extended) use minimum 10% of my session limit. I feel like it was nowhere near this bad when I was just on a free account. I guess this would be a good time to ask, is there something I'm missing or a quick guide on how to optimize usage? I thought simply using Sonnet for simple tasks would be the way but 10 questions max on Sonnet every 4 hours is a complete rip off

src: r/ClaudeAIOPEN_SOURCE
Connecting to live updates