“Long context models with massive custom prompts (~2M) may soon replace fine-tuning for new knowledge! Let’s explore why: (1/10)” / X – https://twitter.com/amanrsanger/status/1772742457937060288

Redpajama-Data-v2 is Incredible : r/LocalLLaMA – https://www.reddit.com/r/LocalLLaMA/comments/17om8xf/redpajamadatav2_is_incredible/

“One of the most imaginative LLM papers I’ve read in a while: use evolution to merge models from HuggingFace to unlock new capabilities, such as Japanese understanding. It’s a form of sophisticated model surgery that requires much smaller compute than traditional LLM training. By…  https://twitter.com/DrJimFan/status/1771927650883522899

“There are a growing number of voices expressing disillusionment with fine-tuning. I’m curious about the sentiment more generally. (I am withholding sharing my opinion rn). Tweets below are from @mlpowered @abacaj @emollick  https://twitter.com/HamelHusain/status/1772426234032541962

We Raised $6.7M to Replace GPT-4 with Your Own Fine-Tuned Models – OpenPipe – https://openpipe.ai/blog/announcing-6-7m-seed-raise

Building and testing C extensions for SQLite with ChatGPT Code Interpreter – https://simonwillison.net/2024/Mar/23/building-c-extensions-for-sqlite-with-chatgpt-code-interpreter/ 

Confluent | Apache Kafka Reinvented for the Cloud – https://www.confluent.io/ 

“Building AI models is faster and cheaper than you probably think. In this ep of Lightcone, we highlight 25 startups that developed or fine-tuned foundation models on a budget — achieving breakthroughs in music, protein design, weather forecasting, and robotics in just 3 months.  https://twitter.com/ycombinator/status/1773402596595896635

“🚀 We’re setting new standards in #GenAI with our latest innovation: the Vectara Factual Consistency Score! 🎯 A calibrated score translates to factual accuracy, empowering devs & ensuring trust. Dive in:  https://twitter.com/vectara/status/1773393615844192611 

Untangling concerns about consolidation in AI – https://generatingconversation.substack.com/p/untangling-concerns-about-consolidation 

[2403.14291v1] Open-Vocabulary Attention Maps with Token Optimization for Semantic Segmentation in Diffusion Models – https://arxiv.org/abs/2403.14291v1 

Towards 1-bit Machine Learning Models – https://mobiusml.github.io/1bit_blog/ 

My binary vector search is better than your FP32 vectors – https://blog.pgvecto.rs/my-binary-vector-search-is-better-than-your-fp32-vectors 

PAID Project Page – https://qy-h00.github.io/attention-interpolation-diffusion/

“📦Introducing AI Gallery: Package, share and run AI with SkyPilot! AI gallery is a community-driven collection of ready-to-run recipes for AI frameworks, models & apps. Launch any recipe with a single command in your own clouds or k8s. Check out recipes for vLLM, TGI…  https://twitter.com/skypilot_org/status/1772660457779958223 

Introducing Superpipe – Superpipe – https://superpipe.ai/blog/2024/03/26/introducing-superpipe/ 

“These things are amazing especially the Claude models Lumentis can now do Mermaid diagrams and Latex for less than 10 cents with Haiku – go check it out  https://twitter.com/hrishioa/status/1772651749326946455 

“RAFT – Retrieval Augmented Fine Tuning 🔥 RAFT offers a method to fine-tune pre-trained LLMs for specific domain RAG settings. Conventional RAG is like an open-book exam, retrieving documents from an index to provide context for answering queries. This makes it more effective…  https://twitter.com/llama_index/status/1772662480210198809 

The Need for Speed: Pruning Transformers with One Recipe – https://www.samirkhaki.com/optin-transformer-pruning/ 

[2403.17694v1] AniPortrait: Audio-Driven Synthesis of Photorealistic Portrait Animation – https://arxiv.org/abs/2403.17694v1 

[2403.17881v1] Deepfake Generation and Detection: A Benchmark and Survey – https://arxiv.org/abs/2403.17881v1 

“The Unreasonable Ineffectiveness of the Deeper Layers Finds minimal degradation of performance of LLMs on various QA benchmarks until after a large fraction (up to half) of the layers are removed  https://twitter.com/arankomatsuzaki/status/1772803686965694684

“InternLM2 Technical Report – Presents an open-source LLM (1.8 ~ 20B params) trained over 2T tokens – Equipped with GQA and trained on up to 32k contexts – Intermediate checkpoints and detailed description of training framework and dataset available  https://twitter.com/arankomatsuzaki/status/1772816281785217087

“🏓 Stream intermediate steps from LangServe @LangChainAI 🦜🔗 JS/TS lets you use chains hosted on LangServe just like local ones. This now includes streamEvents for showing internal steps as they occur! Think source docs for RAG. Thank you rahilvora!  https://twitter.com/Hacubu/status/1772651174384341314

What Computers Cannot Do: The Consequences of Turing-Completeness | Yzena, LLC – https://yzena.com/2024/03/what-computers-cannot-do-the-consequences-of-turing-completeness/

Heads up! You’ve scrolled to the end of this category. There may have been just one or two links (above), so go back up and double check to be sure you didn’t quickly scroll down past it.

Be Sure To Read This Week’s Main Post:

This week’s executive overview and top links are here:

AI News #25: Week Ending 03/22/2024 with Executive Summary and Top 55 Links

The post you just read is an deep dive extension of my weekly newsletter, This Week In AI, an executive summary of the top things to know in AI. Each week, I create an accessible overview for laypeople to feel confident they are conversant with the week’s AI developments. I include a curated list of must-click links of the week, to offer everyone a hands-on opportunity to explore the most intriguing updates in artificial intelligence across various categories, including robotics, imagery, video, AR/VR, science, ethics, and more. Beyond the overview, I post these topic-based deeper dives (below). If you haven’t read this week’s overview, I recommend starting there.

Credits/Sources

Most of these weekly links come from just a few prolific oversharing sources. Please follow them, as they work hard to find the news each week and they make it a lot easier for me to compile.

For previous issues, please visit the archives!

Thanks for reading!

Leave a Reply

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading