Gemma is expanding.... Google announced CodeGemma, a version of Gemma tuned for code generation. And bonus... Gemma is now bumped to v1.1, addressing lots of feedback we got.
Googleblog
Google for Developers Blog - News about Web, Mobile, AI and Cloud
Meta confirmed its GPT-4 competitor, Llama 3, is coming within the month.
At an event in London, Meta confirmed that it plans an initial release of Llama 3, its GPT-4 competitor, within the next month.
The company did not disclose the size of the parameters used in Llama 3, but it's expected to have about 140 billion parameters.
At an event in London, Meta confirmed that it plans an initial release of Llama 3, its GPT-4 competitor, within the next month.
The company did not disclose the size of the parameters used in Llama 3, but it's expected to have about 140 billion parameters.
TechCrunch
Meta confirms that its Llama 3 open source LLM is coming in the next month
Meta's Llama families, built as open-source products, represent a different philosophical approach to how AI should develop as a wider technology.
❤4
DROID: A Large-Scale In-The-Wild Robot Manipulation Dataset
Distributed RObot Interaction Dataset: A diverse robot manipulation dataset with 76k demonstrations, collected across 564 scenes and 84 tasks over the course of a year.
Paper.
Distributed RObot Interaction Dataset: A diverse robot manipulation dataset with 76k demonstrations, collected across 564 scenes and 84 tasks over the course of a year.
Paper.
arXiv.org
DROID: A Large-Scale In-The-Wild Robot Manipulation Dataset
The creation of large, diverse, high-quality robot manipulation datasets is an important stepping stone on the path toward more capable and robust robotic manipulation policies. However, creating...
This media is not supported in your browser
VIEW IN TELEGRAM
GE HealthCare’s Vscan Air SL with Caption AI software provides real-time guidance that shows healthcare professionals how to maneuver the probe to capture diagnostic-quality standard cardiac images.
With the help of on-device AI, there's now a way for handheld ultrasound users to confidently acquire cardiac views for rapid assessments at the point of care.
With the help of on-device AI, there's now a way for handheld ultrasound users to confidently acquire cardiac views for rapid assessments at the point of care.
Meta announced 2nd-gen inference chip MTIAv2
- 708TF/s Int8 / 353TF/s BF16
- 256MB SRAM, 128GB memory
- 90W TDP. 24 chips per node, 3 nodes per rack.
- standard PyTorch stack (Dynamo, Inductor, Triton) for flexibility
Fabbed on TSMC's 5nm process, its fully programmable via the standard PyTorch stack, driven via Triton for software kernels.
This chip is an inference power-house and the software work is entirely driven by the PyTorch team, making usability a first; and its been great to see it in action on various Meta workloads.
- 708TF/s Int8 / 353TF/s BF16
- 256MB SRAM, 128GB memory
- 90W TDP. 24 chips per node, 3 nodes per rack.
- standard PyTorch stack (Dynamo, Inductor, Triton) for flexibility
Fabbed on TSMC's 5nm process, its fully programmable via the standard PyTorch stack, driven via Triton for software kernels.
This chip is an inference power-house and the software work is entirely driven by the PyTorch team, making usability a first; and its been great to see it in action on various Meta workloads.
Meta
Our next generation Meta Training and Inference Accelerator
We are sharing details of our next generation chip in our Meta Training and Inference Accelerator (MTIA) family. MTIA is a long-term bet to provide the most efficient architecture for Meta’s unique workloads.
New paper from Berkeley on Autonomous Evaluation and Refinement of Digital Agents
VLM/LLM-based evaluators can significantly improve the performance of agents for web browsing and device control, advancing sotas by 29% to 75%.
VLM/LLM-based evaluators can significantly improve the performance of agents for web browsing and device control, advancing sotas by 29% to 75%.
arXiv.org
Autonomous Evaluation and Refinement of Digital Agents
We show that domain-general automatic evaluators can significantly improve the performance of agents for web navigation and device control. We experiment with multiple evaluation models that trade...
The music industry just had its 'ChatGPT for music' moment with Udio.
A new AI-powered music creation app called Udio from former Google DeepMind researchers just launched.
It allows users to generate full audio tracks in under 40 seconds with simple prompts and secured early funding from a16z, will i am, Common, and more.
A new AI-powered music creation app called Udio from former Google DeepMind researchers just launched.
It allows users to generate full audio tracks in under 40 seconds with simple prompts and secured early funding from a16z, will i am, Common, and more.
Udio
Udio | AI Music Generator - Official Website
Discover, create, and share music with the world. Use the latest technology to create AI music in seconds.
A newly revealed patent from Microsoft Bing detailed ‘Visual Search’.
The patent describes a reverse image search with personal results tailored to user preferences and interests.
The patent describes a reverse image search with personal results tailored to user preferences and interests.
⚡5
Integrated data visualization and analysis for workhorse biological assays
Make publication-quality figures in under a minute. No more fiddling with Excel, Prism, ggplot, and PowerPoint.
Make publication-quality figures in under a minute. No more fiddling with Excel, Prism, ggplot, and PowerPoint.
⚡4
Google presents Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
1B model that was fine-tuned on up to 5K sequence length passkey instances solves the 1M length problem.
The future of attention just dropped, and it looks a lot like a state space model (finite size, continual updates)
Little doubt now that a mixture of architectures will support the long-term, gradually conditioned memory needed for highly capable agents
1B model that was fine-tuned on up to 5K sequence length passkey instances solves the 1M length problem.
The future of attention just dropped, and it looks a lot like a state space model (finite size, continual updates)
Little doubt now that a mixture of architectures will support the long-term, gradually conditioned memory needed for highly capable agents
arXiv.org
Leave No Context Behind: Efficient Infinite Context Transformers...
This work introduces an efficient method to scale Transformer-based Large Language Models (LLMs) to infinitely long inputs with bounded memory and computation. A key component in our proposed...
⚡4
MR_for_autism_1712848311.pdf
2.6 MB
Eye tracking as a window onto conscious and non-conscious processing in the brain.
The goals are two-fold and relatively simple:
1. propose and test a method for familiarizing individuals with severe autism spectrum disorder (ASD) with the HoloLens 2 headset and the use of MR technology through a tutorial.
2. obtain quantitative learning indicators in MR, such as execution speed and eye tracking, by comparing individuals with ASD to neurotypical individuals.
Over 80% of individuals with ASD successfully familiarized themselves with MR after several sessions.
In addition, the visual activity of individuals with ASD did not differ from that of neurotypical individuals when they successfully familiarized themselves.
This opens a lot of doors for potential learning opportunities in this population.
The goals are two-fold and relatively simple:
1. propose and test a method for familiarizing individuals with severe autism spectrum disorder (ASD) with the HoloLens 2 headset and the use of MR technology through a tutorial.
2. obtain quantitative learning indicators in MR, such as execution speed and eye tracking, by comparing individuals with ASD to neurotypical individuals.
Over 80% of individuals with ASD successfully familiarized themselves with MR after several sessions.
In addition, the visual activity of individuals with ASD did not differ from that of neurotypical individuals when they successfully familiarized themselves.
This opens a lot of doors for potential learning opportunities in this population.
⚡4
Apple nears production of first M4 chips with AI upgrades and plans to bring the new chips to all of its Macs, including new MacBook Pros and Airs, Mac Pro, Mac Studio, Mac mini and iMac across the end of this year and 2025.
Bloomberg.com
Apple Plans to Overhaul Entire Mac Line With AI-Focused M4 Chips
Apple Inc., aiming to boost sluggish computer sales, is preparing to overhaul its entire Mac line with a new family of in-house processors designed to highlight artificial intelligence.
⚡4
Sanctuary AI announced a partnership with European automaker Magna.
The collab features Sanctuary AI’s development of general-purpose AI robots for deployment in Magna’s manufacturing operations.
The collab features Sanctuary AI’s development of general-purpose AI robots for deployment in Magna’s manufacturing operations.
Bloomberg.com
Robotics Startup Sanctuary Signs Deal for Factory Tests, Funds
Humanoid robot-making startup Sanctuary AI has struck a deal with a major auto-parts manufacturer for deployment in its factories and additional equity.
👍4⚡3
Forbes released their AI 50 list.
Here’s a categorized list of the companies, including their current funding amounts.
Here’s a categorized list of the companies, including their current funding amounts.
⚡6
Germany published the report "Generative AI Models - Opportunities and Risks for Industry and Authorities."
Quotes & comments:
1. “LLMs are trained based on huge text corpora. The origin of these texts and their quality are generally not fully verified due to the large amount of data. Therefore, personal or copyrighted data, as well as texts with questionable, false, or discriminatory content (e.g., disinformation, propaganda, or hate messages), may be included in the training set. When generating outputs, these contents may appear in these outputs either verbatim or slightly altered (Weidinger, et al., 2022). Imbalances in the training data can also lead to biases in the model" (page 9)
2. “If individual data points are disproportionately present in the training data, there is a risk that the model cannot adequately learn the desired data distribution and, depending on the extent, tends to produce repetitive, one-sided, or incoherent outputs (known as model collapse). It is expected that this problem will increasingly occur in the future, as LLM-generated data becomes more available on the internet and is used to train new LLMs (Shumailov, et al., 2023). This could lead to self-reinforcing effects, which is particularly critical in cases where texts with abuse potential have been generated, or when a bias in text data becomes entrenched. This happens, for example, as more and more relevant texts are produced and used again for training new models, which in turn generate a multitude of texts (Bender, et al., 2021)." (page 10)
3. “The high linguistic quality of the model outputs, combined with user-friendly access via APIs and the enormous flexibility of responses from currently popular LLMs, makes it easier for criminals to misuse the models for a targeted generation of misinformation (De Angelis, et al., 2023), propaganda texts, hate messages, product reviews, or posts for social media."
According to the report, special attention should be given to the following aspects:
➵ Raising awareness of users;
➵ Testing;
➵ Handling sensitive data;
➵ Establishing transparency;
➵ Auditing of inputs and outputs;
➵ Paying attention to (indirect) prompt injections;
➵ Selection and management of training data;
➵ Developing practical expertise.
Quotes & comments:
1. “LLMs are trained based on huge text corpora. The origin of these texts and their quality are generally not fully verified due to the large amount of data. Therefore, personal or copyrighted data, as well as texts with questionable, false, or discriminatory content (e.g., disinformation, propaganda, or hate messages), may be included in the training set. When generating outputs, these contents may appear in these outputs either verbatim or slightly altered (Weidinger, et al., 2022). Imbalances in the training data can also lead to biases in the model" (page 9)
2. “If individual data points are disproportionately present in the training data, there is a risk that the model cannot adequately learn the desired data distribution and, depending on the extent, tends to produce repetitive, one-sided, or incoherent outputs (known as model collapse). It is expected that this problem will increasingly occur in the future, as LLM-generated data becomes more available on the internet and is used to train new LLMs (Shumailov, et al., 2023). This could lead to self-reinforcing effects, which is particularly critical in cases where texts with abuse potential have been generated, or when a bias in text data becomes entrenched. This happens, for example, as more and more relevant texts are produced and used again for training new models, which in turn generate a multitude of texts (Bender, et al., 2021)." (page 10)
3. “The high linguistic quality of the model outputs, combined with user-friendly access via APIs and the enormous flexibility of responses from currently popular LLMs, makes it easier for criminals to misuse the models for a targeted generation of misinformation (De Angelis, et al., 2023), propaganda texts, hate messages, product reviews, or posts for social media."
According to the report, special attention should be given to the following aspects:
➵ Raising awareness of users;
➵ Testing;
➵ Handling sensitive data;
➵ Establishing transparency;
➵ Auditing of inputs and outputs;
➵ Paying attention to (indirect) prompt injections;
➵ Selection and management of training data;
➵ Developing practical expertise.
⚡4
Google DeepMind published new work on using reinforcement learning to train robots to be more agile (like soccer players).
They tested a new method of adding disruptive forces and randomness, which led to impressive results compared to the baseline.
They tested a new method of adding disruptive forces and randomness, which led to impressive results compared to the baseline.
Science Robotics
Learning agile soccer skills for a bipedal robot with deep reinforcement learning
OP3 humanoid robots learned to play agile soccer using deep reinforcement learning.
⚡4
Researchers at Tsinghua University in China have developed a revolutionary new AI chip that uses light instead of electricity to process data.
Dubbed “Taichi,” the chip is reportedly over 1,000 times more energy-efficient than Nvidia’s high performance H100 GPU chip.
Dubbed “Taichi,” the chip is reportedly over 1,000 times more energy-efficient than Nvidia’s high performance H100 GPU chip.
Interesting Engineering
Light-based chip: China's Taichi could power artificial general intelligence
Researchers have developed a highly energy-efficient photonic AI chip called Taichi, which could accelerate the development of advanced computing solutions.
⚡5🔥2
The AI Index 2024 annual report by Stanford University is finally here.
This is the resource that will impact our conversations on the topic in the following months, highlighting the latest technical developments, its impact on global organizations, society and research, and the existential need for clear regulation and governance.
This is the resource that will impact our conversations on the topic in the following months, highlighting the latest technical developments, its impact on global organizations, society and research, and the existential need for clear regulation and governance.
⚡4
Executive exodus and a 10% cut to workforce at Tesla
Drew Baglino resigned from Tesla. Baglino had been at Tesla since 2006 and was SVP of Powertrain/Energy Engineering. He was one of just four named executive officers and integral to the work the company was doing on everything from EVs to energy storage and next gen 4680 cells. Drew posted on social media a few hours after our story to confirm his departure.
Tesla's Public Policy chief Rohan Patel also has left the company, we reported.
When you take in to account it follows key names in Tesla's semicondcutor team leaving earlier this year and Zach Kirkhorn leaving in August, its a lot of intellectual capital and experience out.
Actually things are going well - particularly in the energy division - and a reason for Baglino leaving is he felt the place was in good hands.
Last night Musk told staff around the world Tesla would cut 10% of more of its global workforce. The reasons are clear: pursuit of cost cuts and productivity in a tough environment. But Tesla's grown quickly. And Musk cited the number of duplicate roles in his reason for the RIF. 10% of Tesla's staff is around 14,000 people. So its sizable, the biggest RIF ever.
Drew Baglino resigned from Tesla. Baglino had been at Tesla since 2006 and was SVP of Powertrain/Energy Engineering. He was one of just four named executive officers and integral to the work the company was doing on everything from EVs to energy storage and next gen 4680 cells. Drew posted on social media a few hours after our story to confirm his departure.
Tesla's Public Policy chief Rohan Patel also has left the company, we reported.
When you take in to account it follows key names in Tesla's semicondcutor team leaving earlier this year and Zach Kirkhorn leaving in August, its a lot of intellectual capital and experience out.
Actually things are going well - particularly in the energy division - and a reason for Baglino leaving is he felt the place was in good hands.
Last night Musk told staff around the world Tesla would cut 10% of more of its global workforce. The reasons are clear: pursuit of cost cuts and productivity in a tough environment. But Tesla's grown quickly. And Musk cited the number of duplicate roles in his reason for the RIF. 10% of Tesla's staff is around 14,000 people. So its sizable, the biggest RIF ever.
Bloomberg.com
Tesla Executive Baglino Leaves as Musk Loses Another Top Deputy
Two of Tesla Inc.’s top executives have left in the midst of the carmaker’s largest-ever round of job cuts, as slowing electric-vehicle demand leads the company to reduce its global headcount by more than 10%.
⚡4
Paradigm announced an open source Reth Alphanet.
Reth AlphaNet is an OP Stack-compatible testnet rollup that maximizes Reth performance and enables experimentation of bleeding edge Ethereum Research.
Reth AlphaNet is a testnet rollup built on OP Stack & OP Reth.
Reth AlphaNet is aimed at experimentation of Ethereum research at the bleeding edge, and comes with 3 EIPs not available anywhere else yet. EIP-3074, EIP-71212, EIP-2537.
These EIPs are built with best-practices in mind, are optimized, and tested.
Reth AlphaNet is an OP Stack-compatible testnet rollup that maximizes Reth performance and enables experimentation of bleeding edge Ethereum Research.
Reth AlphaNet is a testnet rollup built on OP Stack & OP Reth.
Reth AlphaNet is aimed at experimentation of Ethereum research at the bleeding edge, and comes with 3 EIPs not available anywhere else yet. EIP-3074, EIP-71212, EIP-2537.
These EIPs are built with best-practices in mind, are optimized, and tested.
Paradigm
Reth AlphaNet - Paradigm
Paradigm is a research-driven crypto investment firm that funds companies and protocols from their earliest stages.
This is Microsoft Asia team overtaking original GPT-4 with their evol tuning
Today they are announcing WizardLM-2, next generation state-of-the-art LLM.
New family includes three cutting-edge models: WizardLM-2 8x22B, 70B, and 7B - demonstrates highly competitive performance compared to leading proprietary LLMs.
Model weights.
Today they are announcing WizardLM-2, next generation state-of-the-art LLM.
New family includes three cutting-edge models: WizardLM-2 8x22B, 70B, and 7B - demonstrates highly competitive performance compared to leading proprietary LLMs.
Model weights.
wizardlm.github.io
TWITTER BANNER TITLE META TAG
TWITTER BANNER DESCRIPTION META TAG
⚡4