deep learning - TechTalks https://bdtechtalks.com Technology solving problems... and creating new ones Mon, 21 Apr 2025 12:50:16 +0000 en-US hourly 1 https://i0.wp.com/bdtechtalks.com/wp-content/uploads/2018/02/cropped-TechTalks-logo.jpg?fit=32%2C32&ssl=1 deep learning - TechTalks https://bdtechtalks.com 32 32 99082954 Are we at the cusp of a new era for artificial intelligence? https://bdtechtalks.com/2025/04/21/are-we-at-the-cusp-of-a-new-era-for-artificial-intelligence/?utm_source=rss&utm_medium=rss&utm_campaign=are-we-at-the-cusp-of-a-new-era-for-artificial-intelligence Mon, 21 Apr 2025 12:50:14 +0000 https://bdtechtalks.com/?p=24412 The "Era of Experience" envisions AI's evolution beyond human data, emphasizing self-learning from real-world interactions. But challenges loom for this vision.

The post Are we at the cusp of a new era for artificial intelligence? first appeared on TechTalks.

]]>
24412
Everything you need to know about Grok-3 https://bdtechtalks.com/2025/02/20/what-is-grok-3/?utm_source=rss&utm_medium=rss&utm_campaign=what-is-grok-3 https://bdtechtalks.com/2025/02/20/what-is-grok-3/#respond Thu, 20 Feb 2025 14:54:29 +0000 https://bdtechtalks.com/?p=23892 Grok-3 storms the AI scene, boasting superior capabilities and competitive benchmarks. Here's everything to know about this new LLM and LRM from xAI.

The post Everything you need to know about Grok-3 first appeared on TechTalks.

]]>
https://bdtechtalks.com/2025/02/20/what-is-grok-3/feed/ 0 23892
New training paradigm prevents machine learning models from learning spurious correlations https://bdtechtalks.com/2025/01/20/memorization-aware-training-machine-learning/?utm_source=rss&utm_medium=rss&utm_campaign=memorization-aware-training-machine-learning https://bdtechtalks.com/2025/01/20/memorization-aware-training-machine-learning/#respond Mon, 20 Jan 2025 14:26:35 +0000 https://bdtechtalks.com/?p=23563 Meta researchers show how memorization-aware training can help machine learning models avoid developing dangerous biases.

The post New training paradigm prevents machine learning models from learning spurious correlations first appeared on TechTalks.

]]>
https://bdtechtalks.com/2025/01/20/memorization-aware-training-machine-learning/feed/ 0 23563
GEAR turbo-charges LLMs with advanced graph-based RAG capabilities https://bdtechtalks.com/2025/01/13/gear-graph-based-llm-rag/?utm_source=rss&utm_medium=rss&utm_campaign=gear-graph-based-llm-rag https://bdtechtalks.com/2025/01/13/gear-graph-based-llm-rag/#respond Mon, 13 Jan 2025 20:44:56 +0000 https://bdtechtalks.com/?p=23513 GEAR enhances RAG by automatically extracting triples and using beam search to create and iterate over graph representations from retrieved documents.

The post GEAR turbo-charges LLMs with advanced graph-based RAG capabilities first appeared on TechTalks.

]]>
https://bdtechtalks.com/2025/01/13/gear-graph-based-llm-rag/feed/ 0 23513
How treating LLMs as “actors” can produce better results https://bdtechtalks.com/2024/11/25/llm-method-actors/?utm_source=rss&utm_medium=rss&utm_campaign=llm-method-actors https://bdtechtalks.com/2024/11/25/llm-method-actors/#respond Mon, 25 Nov 2024 13:56:08 +0000 https://bdtechtalks.com/?p=22989 Think of LLMs as actors, prompts as scripts, and LLM outputs as performances.

The post How treating LLMs as “actors” can produce better results first appeared on TechTalks.

]]>
https://bdtechtalks.com/2024/11/25/llm-method-actors/feed/ 0 22989
Mistral expands its reach in the SLM space with Ministral models https://bdtechtalks.com/2024/10/16/mistral-slm-ministral/?utm_source=rss&utm_medium=rss&utm_campaign=mistral-slm-ministral https://bdtechtalks.com/2024/10/16/mistral-slm-ministral/#respond Wed, 16 Oct 2024 20:10:07 +0000 https://bdtechtalks.com/?p=22583 The new Ministral models outperforms other small language models, including Gemma 2, Phi 3.5, and Llama 3.2.

The post Mistral expands its reach in the SLM space with Ministral models first appeared on TechTalks.

]]>
https://bdtechtalks.com/2024/10/16/mistral-slm-ministral/feed/ 0 22583
Meta SAM 2 is the most impressive object segmentation model https://bdtechtalks.com/2024/08/05/meta-sam-2-object-segmentation-model/?utm_source=rss&utm_medium=rss&utm_campaign=meta-sam-2-object-segmentation-model https://bdtechtalks.com/2024/08/05/meta-sam-2-object-segmentation-model/#respond Mon, 05 Aug 2024 07:05:33 +0000 https://bdtechtalks.com/?p=22021 Meta's new object segmentation model, SAM 2, provides near-real-time inference on a wide variety of objects and environments.

The post Meta SAM 2 is the most impressive object segmentation model first appeared on TechTalks.

]]>
https://bdtechtalks.com/2024/08/05/meta-sam-2-object-segmentation-model/feed/ 0 22021
Energy-Based World Models bring human-like cognition to AI https://bdtechtalks.com/2024/06/24/energy-based-world-models/?utm_source=rss&utm_medium=rss&utm_campaign=energy-based-world-models https://bdtechtalks.com/2024/06/24/energy-based-world-models/#respond Mon, 24 Jun 2024 12:12:15 +0000 https://bdtechtalks.com/?p=21738 Energy-based world models (EBWM) enable AI systems to reflect on their predictions and achieve human-like cognitive abilities missing in autoregressive models.

The post Energy-Based World Models bring human-like cognition to AI first appeared on TechTalks.

]]>
https://bdtechtalks.com/2024/06/24/energy-based-world-models/feed/ 0 21738
Self-assembling neural networks can open new directions for AI research https://bdtechtalks.com/2023/11/13/self-assembling-neural-networks-ndp/?utm_source=rss&utm_medium=rss&utm_campaign=self-assembling-neural-networks-ndp https://bdtechtalks.com/2023/11/13/self-assembling-neural-networks-ndp/#respond Mon, 13 Nov 2023 14:00:00 +0000 https://bdtechtalks.com/?p=17671 A new software architecture uses neural development programs (NDP) to self-assemble deep learning models from basic units, like their biological counterparts.

The post Self-assembling neural networks can open new directions for AI research first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/11/13/self-assembling-neural-networks-ndp/feed/ 0 17671
A simple guide to gradient descent in machine learning https://bdtechtalks.com/2023/07/31/what-is-gradient-descent/?utm_source=rss&utm_medium=rss&utm_campaign=what-is-gradient-descent https://bdtechtalks.com/2023/07/31/what-is-gradient-descent/#respond Mon, 31 Jul 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16965 Gradient descent is the main technique for training machine learning and deep learning models. Read all about it.

The post A simple guide to gradient descent in machine learning first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/07/31/what-is-gradient-descent/feed/ 0 16965
The complete guide to LLM fine-tuning https://bdtechtalks.com/2023/07/10/llm-fine-tuning/?utm_source=rss&utm_medium=rss&utm_campaign=llm-fine-tuning https://bdtechtalks.com/2023/07/10/llm-fine-tuning/#comments Mon, 10 Jul 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16807 Everything to know about LLM fine-tuning, supervised fine-tuning, reinforcement learning from human feedback (RLHF), and parameter-efficient fine-tuning (PEFT)

The post The complete guide to LLM fine-tuning first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/07/10/llm-fine-tuning/feed/ 1 16807
How to teach AI to imitate human thought and action https://bdtechtalks.com/2023/07/03/ai-thought-cloning/?utm_source=rss&utm_medium=rss&utm_campaign=ai-thought-cloning https://bdtechtalks.com/2023/07/03/ai-thought-cloning/#respond Mon, 03 Jul 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16768 A new technique called "Thought Cloning" trains AI systems on both behavior and reasoning data, making them robust and interpretable.

The post How to teach AI to imitate human thought and action first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/07/03/ai-thought-cloning/feed/ 0 16768
What is low-rank adaptation (LoRA)? https://bdtechtalks.com/2023/05/22/what-is-lora/?utm_source=rss&utm_medium=rss&utm_campaign=what-is-lora https://bdtechtalks.com/2023/05/22/what-is-lora/#respond Mon, 22 May 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16415 Low-rank adaptation (LoRA) is a technique that cuts the costs of fine-tuning large language models (LLM) to a fraction of its actual figure.

The post What is low-rank adaptation (LoRA)? first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/05/22/what-is-lora/feed/ 0 16415
Are the emergent abilities of LLMs like GPT-4 a mirage? https://bdtechtalks.com/2023/05/17/llm-emergent-abilities-mirage/?utm_source=rss&utm_medium=rss&utm_campaign=llm-emergent-abilities-mirage https://bdtechtalks.com/2023/05/17/llm-emergent-abilities-mirage/#respond Wed, 17 May 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16381 A new study byStanford University suggests that the emergent abilities of large language models (LLM) are caused by a poor choice of evaluation metrics.

The post Are the emergent abilities of LLMs like GPT-4 a mirage? first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/05/17/llm-emergent-abilities-mirage/feed/ 0 16381
How to customize LLMs like ChatGPT with your own data and documents https://bdtechtalks.com/2023/05/01/customize-chatgpt-llm-embeddings/?utm_source=rss&utm_medium=rss&utm_campaign=customize-chatgpt-llm-embeddings https://bdtechtalks.com/2023/05/01/customize-chatgpt-llm-embeddings/#comments Mon, 01 May 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16265 ChatGPT and other LLMs are limited to their training data. Here's how you can customize them with embeddings and your own documents.

The post How to customize LLMs like ChatGPT with your own data and documents first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/05/01/customize-chatgpt-llm-embeddings/feed/ 11 16265
What we learned from the deep learning revolution https://bdtechtalks.com/2023/04/10/deep-learning-revolution-terry-sejnowski/?utm_source=rss&utm_medium=rss&utm_campaign=deep-learning-revolution-terry-sejnowski https://bdtechtalks.com/2023/04/10/deep-learning-revolution-terry-sejnowski/#respond Mon, 10 Apr 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16102 Neuroscientist Terry Sejnowski discusses the early struggles of deep learning, its explosion into the mainstream, and the lessons learned from decades of research and development.

The post What we learned from the deep learning revolution first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/04/10/deep-learning-revolution-terry-sejnowski/feed/ 0 16102
What to know about augmented language models https://bdtechtalks.com/2023/04/03/augmented-language-models/?utm_source=rss&utm_medium=rss&utm_campaign=augmented-language-models https://bdtechtalks.com/2023/04/03/augmented-language-models/#respond Mon, 03 Apr 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16059 Large language models suffer from fundamental problems, such as failing at math and reasoning. Augmented language models address some of these problems.

The post What to know about augmented language models first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/04/03/augmented-language-models/feed/ 0 16059
Microsoft and OpenAI get ahead in the LLM competition https://bdtechtalks.com/2023/03/28/microsoft-openai-llm-competition/?utm_source=rss&utm_medium=rss&utm_campaign=microsoft-openai-llm-competition https://bdtechtalks.com/2023/03/28/microsoft-openai-llm-competition/#respond Tue, 28 Mar 2023 13:00:00 +0000 https://bdtechtalks.com/?p=16025 Microsoft and OpenAI released a bunch of good new LLM products. Google has released Bard, but it is still lagging behind.

The post Microsoft and OpenAI get ahead in the LLM competition first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/03/28/microsoft-openai-llm-competition/feed/ 0 16025
What you need to know about multimodal language models https://bdtechtalks.com/2023/03/13/multimodal-large-language-models/?utm_source=rss&utm_medium=rss&utm_campaign=multimodal-large-language-models https://bdtechtalks.com/2023/03/13/multimodal-large-language-models/#comments Mon, 13 Mar 2023 14:00:00 +0000 https://bdtechtalks.com/?p=15936 Multimodal language models bring together text, images, and other datatypes to solve some of the problems current artificial intelligence systems suffer from.

The post What you need to know about multimodal language models first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/03/13/multimodal-large-language-models/feed/ 1 15936
To understand language models, we must separate “language” from “thought” https://bdtechtalks.com/2023/02/20/llm-dissociating-language-and-thought/?utm_source=rss&utm_medium=rss&utm_campaign=llm-dissociating-language-and-thought https://bdtechtalks.com/2023/02/20/llm-dissociating-language-and-thought/#respond Mon, 20 Feb 2023 14:00:00 +0000 https://bdtechtalks.com/?p=15828 To understand the power and limits of large language models (LLM), we must separate “formal” from “functional” linguistic competence.

The post To understand language models, we must separate “language” from “thought” first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/02/20/llm-dissociating-language-and-thought/feed/ 0 15828
AI21 Labs’ mission to make large language models get their facts right https://bdtechtalks.com/2023/01/30/ai21labs-llms-ralm/?utm_source=rss&utm_medium=rss&utm_campaign=ai21labs-llms-ralm https://bdtechtalks.com/2023/01/30/ai21labs-llms-ralm/#respond Mon, 30 Jan 2023 14:00:00 +0000 https://bdtechtalks.com/?p=15719 AI21 Labs chief scientist Yoav Levine explains how retrieval augmented language modeling can solve one of the biggest problems of LLMs.

The post AI21 Labs’ mission to make large language models get their facts right first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/01/30/ai21labs-llms-ralm/feed/ 0 15719
The definitive guide to adversarial machine learning https://bdtechtalks.com/2023/01/23/adversarial-machine-learning-book/?utm_source=rss&utm_medium=rss&utm_campaign=adversarial-machine-learning-book https://bdtechtalks.com/2023/01/23/adversarial-machine-learning-book/#respond Mon, 23 Jan 2023 14:00:00 +0000 https://bdtechtalks.com/?p=15670 "Adversarial Robustness for Machine Learning" provides a comprehensive overview of adversarial ML.

The post The definitive guide to adversarial machine learning first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/01/23/adversarial-machine-learning-book/feed/ 0 15670
Why ChatGPT is not a threat to Google Search https://bdtechtalks.com/2023/01/02/chatgpt-google-search/?utm_source=rss&utm_medium=rss&utm_campaign=chatgpt-google-search https://bdtechtalks.com/2023/01/02/chatgpt-google-search/#comments Mon, 02 Jan 2023 14:00:00 +0000 https://bdtechtalks.com/?p=15511 ChatGPT is a remarkable LLM, with potential applications for online search. But it might be a bit of a stretch to say that it will dethrone Google.

The post Why ChatGPT is not a threat to Google Search first appeared on TechTalks.

]]>
https://bdtechtalks.com/2023/01/02/chatgpt-google-search/feed/ 9 15511
What is the “forward-forward” algorithm, Geoffrey Hinton’s new AI technique? https://bdtechtalks.com/2022/12/19/forward-forward-algorithm-geoffrey-hinton/?utm_source=rss&utm_medium=rss&utm_campaign=forward-forward-algorithm-geoffrey-hinton https://bdtechtalks.com/2022/12/19/forward-forward-algorithm-geoffrey-hinton/#comments Mon, 19 Dec 2022 14:00:00 +0000 https://bdtechtalks.com/?p=15397 In a new NeurIPS paper, Geoffrey Hinton introduced the “forward-forward algorithm,” a new learning algorithm for artificial neural networks inspired by the brain.

The post What is the “forward-forward” algorithm, Geoffrey Hinton’s new AI technique? first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/12/19/forward-forward-algorithm-geoffrey-hinton/feed/ 7 15397
What to (not) expect from OpenAI’s ChatGPT https://bdtechtalks.com/2022/12/05/openai-chatgpt/?utm_source=rss&utm_medium=rss&utm_campaign=openai-chatgpt https://bdtechtalks.com/2022/12/05/openai-chatgpt/#comments Mon, 05 Dec 2022 14:00:00 +0000 https://bdtechtalks.com/?p=15277 OpenAI's ChatGPT, with all its successes and failures, is a reflection of the short but rich history of large language models (LLM).

The post What to (not) expect from OpenAI’s ChatGPT first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/12/05/openai-chatgpt/feed/ 2 15277
The power of wide transformers models https://bdtechtalks.com/2022/10/31/wide-transformers-models/?utm_source=rss&utm_medium=rss&utm_campaign=wide-transformers-models https://bdtechtalks.com/2022/10/31/wide-transformers-models/#respond Mon, 31 Oct 2022 14:00:00 +0000 https://bdtechtalks.com/?p=15053 Switching transformer models from deep to wide architecture results in significant improvements in speed, memory, and interpretability.

The post The power of wide transformers models first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/10/31/wide-transformers-models/feed/ 0 15053
DeepMind AlphaTensor: The delicate balance between human and artificial intelligence https://bdtechtalks.com/2022/10/10/deepmind-alphatensor/?utm_source=rss&utm_medium=rss&utm_campaign=deepmind-alphatensor https://bdtechtalks.com/2022/10/10/deepmind-alphatensor/#comments Mon, 10 Oct 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14894 DeepMind AlphaTensor shows how the right combination of human and artificial intelligence can find solutions to complicated problems.

The post DeepMind AlphaTensor: The delicate balance between human and artificial intelligence first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/10/10/deepmind-alphatensor/feed/ 2 14894
Self-attention can be big for TinyML applications https://bdtechtalks.com/2022/09/26/self-attention-tinyml/?utm_source=rss&utm_medium=rss&utm_campaign=self-attention-tinyml https://bdtechtalks.com/2022/09/26/self-attention-tinyml/#comments Mon, 26 Sep 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14769 Researchers and the University of Waterloo and DarwinAI present a new deep learning architecture that brings self-attention to TinyML applications.

The post Self-attention can be big for TinyML applications first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/09/26/self-attention-tinyml/feed/ 1 14769
Can GPT-3 be honest when it speaks nonsense? https://bdtechtalks.com/2022/09/05/llm-uncertainty-verbalized-probability/?utm_source=rss&utm_medium=rss&utm_campaign=llm-uncertainty-verbalized-probability https://bdtechtalks.com/2022/09/05/llm-uncertainty-verbalized-probability/#respond Mon, 05 Sep 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14605 A study by researchers at OpenAI and the University of Oxford shows large language models can be calibrated to express their uncertainty in their answers.

The post Can GPT-3 be honest when it speaks nonsense? first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/09/05/llm-uncertainty-verbalized-probability/feed/ 0 14605
AI scientists are studying the “emergent” abilities of large language models https://bdtechtalks.com/2022/08/22/llm-emergent-abilities/?utm_source=rss&utm_medium=rss&utm_campaign=llm-emergent-abilities https://bdtechtalks.com/2022/08/22/llm-emergent-abilities/#respond Mon, 22 Aug 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14487 A new study by researchers at Google, Stanford, DeepMind, and the University of North Carolina explores the emergent abilities of large language models as they become larger.

The post AI scientists are studying the “emergent” abilities of large language models first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/08/22/llm-emergent-abilities/feed/ 0 14487
Reinforcement learning models are prone to membership inference attacks https://bdtechtalks.com/2022/08/15/reinforcement-learning-membership-inference-attacks/?utm_source=rss&utm_medium=rss&utm_campaign=reinforcement-learning-membership-inference-attacks https://bdtechtalks.com/2022/08/15/reinforcement-learning-membership-inference-attacks/#respond Mon, 15 Aug 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14436 A new study by researchers at McGill University, Mila, and the University of Waterloo highlights the privacy threats of deep reinforcement learning algorithms.

The post Reinforcement learning models are prone to membership inference attacks first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/08/15/reinforcement-learning-membership-inference-attacks/feed/ 0 14436
How to solve AI’s “common sense” problem https://bdtechtalks.com/2022/08/08/machines-like-us-review/?utm_source=rss&utm_medium=rss&utm_campaign=machines-like-us-review https://bdtechtalks.com/2022/08/08/machines-like-us-review/#comments Mon, 08 Aug 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14366 "Machines Like Us" argues that the artificial intelligence community needs to revisit symbolic reasoning to solve AI's "common sense" problem.

The post How to solve AI’s “common sense” problem first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/08/08/machines-like-us-review/feed/ 4 14366
Democratizing the hardware side of large language models https://bdtechtalks.com/2022/08/01/cerebras-large-language-models/?utm_source=rss&utm_medium=rss&utm_campaign=cerebras-large-language-models https://bdtechtalks.com/2022/08/01/cerebras-large-language-models/#comments Mon, 01 Aug 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14306 Cerebras CEO Andrew Feldman discusses the hardware challenges of LLMs and his vision to reduce the costs and complexity of training and running large neural networks.

The post Democratizing the hardware side of large language models first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/08/01/cerebras-large-language-models/feed/ 1 14306
Large language models can’t plan, even if they write fancy essays https://bdtechtalks.com/2022/07/25/large-language-models-cant-plan/?utm_source=rss&utm_medium=rss&utm_campaign=large-language-models-cant-plan https://bdtechtalks.com/2022/07/25/large-language-models-cant-plan/#respond Mon, 25 Jul 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14237 A study by researchers at Arizona State University shows large language models perform very poorly at tasks that require methodical planning.

The post Large language models can’t plan, even if they write fancy essays first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/07/25/large-language-models-cant-plan/feed/ 0 14237
BLOOM can set a new culture for AI research—but challenges remain https://bdtechtalks.com/2022/07/18/bloom-large-language-model-ai-research/?utm_source=rss&utm_medium=rss&utm_campaign=bloom-large-language-model-ai-research https://bdtechtalks.com/2022/07/18/bloom-large-language-model-ai-research/#respond Mon, 18 Jul 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14184 The open-source example that BLOOM has set can be very beneficial to the future of research in LLMs and AI. But some of the challenges that are inherent to large language models remain to be solved.

The post BLOOM can set a new culture for AI research—but challenges remain first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/07/18/bloom-large-language-model-ai-research/feed/ 0 14184
Large language models might reason—if you know how to speak to them https://bdtechtalks.com/2022/07/11/large-language-models-zero-shot-reasoning/?utm_source=rss&utm_medium=rss&utm_campaign=large-language-models-zero-shot-reasoning https://bdtechtalks.com/2022/07/11/large-language-models-zero-shot-reasoning/#respond Mon, 11 Jul 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14127 A new study by the University of Tokyo shows that with the right prompt, large language models can perform zero-shot reasoning.

The post Large language models might reason—if you know how to speak to them first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/07/11/large-language-models-zero-shot-reasoning/feed/ 0 14127
Large language models have a reasoning problem https://bdtechtalks.com/2022/06/27/large-language-models-logical-reasoning/?utm_source=rss&utm_medium=rss&utm_campaign=large-language-models-logical-reasoning https://bdtechtalks.com/2022/06/27/large-language-models-logical-reasoning/#comments Mon, 27 Jun 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14050 According to a research paper by scientists at UCLA, transformers, the deep learning architectures used in LLMs, don’t learn to emulate reasoning functions.

The post Large language models have a reasoning problem first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/06/27/large-language-models-logical-reasoning/feed/ 1 14050
“Sentience” is the wrong discussion to have on AI right now https://bdtechtalks.com/2022/06/20/lamda-large-language-models-sentient-ai/?utm_source=rss&utm_medium=rss&utm_campaign=lamda-large-language-models-sentient-ai https://bdtechtalks.com/2022/06/20/lamda-large-language-models-sentient-ai/#comments Mon, 20 Jun 2022 13:00:00 +0000 https://bdtechtalks.com/?p=14013 Instead of "sentience" and "consciousness," we should discuss human compatibility and trust issues with current AI systems.

The post “Sentience” is the wrong discussion to have on AI right now first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/06/20/lamda-large-language-models-sentient-ai/feed/ 1 14013
This deep learning technique solves one of the tough challenges of robotics https://bdtechtalks.com/2022/05/09/diffskill-robotics-deformable-object-manipulation/?utm_source=rss&utm_medium=rss&utm_campaign=diffskill-robotics-deformable-object-manipulation https://bdtechtalks.com/2022/05/09/diffskill-robotics-deformable-object-manipulation/#respond Mon, 09 May 2022 13:00:00 +0000 https://bdtechtalks.com/?p=13684 DiffSkill is a deep learning technique that makes robots much more stable at handling deformable objects.

The post This deep learning technique solves one of the tough challenges of robotics first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/05/09/diffskill-robotics-deformable-object-manipulation/feed/ 0 13684
Machine learning: What is the transformer architecture? https://bdtechtalks.com/2022/05/02/what-is-the-transformer/?utm_source=rss&utm_medium=rss&utm_campaign=what-is-the-transformer https://bdtechtalks.com/2022/05/02/what-is-the-transformer/#respond Mon, 02 May 2022 13:03:52 +0000 https://bdtechtalks.com/?p=13637 The transformer model has become one of the main highlights of advances in deep learning and deep neural networks.

The post Machine learning: What is the transformer architecture? first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/05/02/what-is-the-transformer/feed/ 0 13637
FOMO is a TinyML neural network for real-time object detection https://bdtechtalks.com/2022/04/18/fomo-tinyml-object-detection/?utm_source=rss&utm_medium=rss&utm_campaign=fomo-tinyml-object-detection https://bdtechtalks.com/2022/04/18/fomo-tinyml-object-detection/#respond Mon, 18 Apr 2022 13:00:00 +0000 https://bdtechtalks.com/?p=13486 FOMO is a deep learning object detection model that weighs less than 200 kilobytes.

The post FOMO is a TinyML neural network for real-time object detection first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/04/18/fomo-tinyml-object-detection/feed/ 0 13486
DALL-E 2, the future of AI research, and OpenAI’s business model https://bdtechtalks.com/2022/04/11/openai-dall-e-2/?utm_source=rss&utm_medium=rss&utm_campaign=openai-dall-e-2 https://bdtechtalks.com/2022/04/11/openai-dall-e-2/#comments Mon, 11 Apr 2022 13:00:00 +0000 https://bdtechtalks.com/?p=13438 OpenAI's DALL-E 2 shows how far the AI research community has come toward harnessing the power of deep learning and addressing some of its limits.

The post DALL-E 2, the future of AI research, and OpenAI’s business model first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/04/11/openai-dall-e-2/feed/ 1 13438
Meta’s Yann LeCun on his vision for human-level AI https://bdtechtalks.com/2022/03/07/yann-lecun-ai-self-supervised-learning/?utm_source=rss&utm_medium=rss&utm_campaign=yann-lecun-ai-self-supervised-learning https://bdtechtalks.com/2022/03/07/yann-lecun-ai-self-supervised-learning/#comments Mon, 07 Mar 2022 13:51:32 +0000 https://bdtechtalks.com/?p=13103 Yann LeCun, deep learning pioneer and Chief AI Scientist at Meta, explains a vision for getting closer to human-level AI.

The post Meta’s Yann LeCun on his vision for human-level AI first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/03/07/yann-lecun-ai-self-supervised-learning/feed/ 1 13103
What is neural architecture search? https://bdtechtalks.com/2022/02/28/what-is-neural-architecture-search/?utm_source=rss&utm_medium=rss&utm_campaign=what-is-neural-architecture-search https://bdtechtalks.com/2022/02/28/what-is-neural-architecture-search/#respond Mon, 28 Feb 2022 14:00:00 +0000 https://bdtechtalks.com/?p=13059 Neural architecture search NAS is a series of machine learning techniques that can help discover optimal neural networks for a given problem.

The post What is neural architecture search? first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/02/28/what-is-neural-architecture-search/feed/ 0 13059
What DeepMind’s AlphaCode is and isn’t https://bdtechtalks.com/2022/02/07/deepmind-alphacode-competitive-programming/?utm_source=rss&utm_medium=rss&utm_campaign=deepmind-alphacode-competitive-programming https://bdtechtalks.com/2022/02/07/deepmind-alphacode-competitive-programming/#respond Mon, 07 Feb 2022 14:00:00 +0000 https://bdtechtalks.com/?p=12880 This article is part of our reviews of AI research papers, a series of posts that explore the latest findings in artificial intelligence. DeepMind is the latest AI research lab to introduce a deep learning model that can generate software source code with remarkable results. Called AlphaCode, the model is based on Transformers, the same architecture […]

The post What DeepMind’s AlphaCode is and isn’t first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/02/07/deepmind-alphacode-competitive-programming/feed/ 0 12880
TinyML is bringing neural networks to microcontrollers https://bdtechtalks.com/2022/01/17/mcunetv2-tinyml-deep-learning-microcontrollers/?utm_source=rss&utm_medium=rss&utm_campaign=mcunetv2-tinyml-deep-learning-microcontrollers https://bdtechtalks.com/2022/01/17/mcunetv2-tinyml-deep-learning-microcontrollers/#comments Mon, 17 Jan 2022 14:00:00 +0000 https://bdtechtalks.com/?p=12687 IBM and MIT have created a new deep learning technique that can run CNNs on low-power, low-memory microcontrollers.

The post TinyML is bringing neural networks to microcontrollers first appeared on TechTalks.

]]>
https://bdtechtalks.com/2022/01/17/mcunetv2-tinyml-deep-learning-microcontrollers/feed/ 1 12687
How can we tell if artificial intelligence understands our language? https://bdtechtalks.com/2021/12/20/artificial-intelligence-large-language-understanding/?utm_source=rss&utm_medium=rss&utm_campaign=artificial-intelligence-large-language-understanding https://bdtechtalks.com/2021/12/20/artificial-intelligence-large-language-understanding/#respond Mon, 20 Dec 2021 14:00:00 +0000 https://bdtechtalks.com/?p=12464 AI scientist Blaise Aguera y Arcas argues that large language models have a great deal to teach us about “the nature of language, understanding, intelligence, sociality, and personhood.”

The post How can we tell if artificial intelligence understands our language? first appeared on TechTalks.

]]>
https://bdtechtalks.com/2021/12/20/artificial-intelligence-large-language-understanding/feed/ 0 12464
DeepMind’s AI can untangle knots. But does it guide human intuition? https://bdtechtalks.com/2021/12/13/deepminds-machine-learning-mathematics/?utm_source=rss&utm_medium=rss&utm_campaign=deepminds-machine-learning-mathematics https://bdtechtalks.com/2021/12/13/deepminds-machine-learning-mathematics/#respond Mon, 13 Dec 2021 14:00:00 +0000 https://bdtechtalks.com/?p=12402 Deep learning can help discover mathematical relations that evade human scientists, a recent paper by researchers at DeepMind shows.

The post DeepMind’s AI can untangle knots. But does it guide human intuition? first appeared on TechTalks.

]]>
https://bdtechtalks.com/2021/12/13/deepminds-machine-learning-mathematics/feed/ 0 12402
Neural networks can hide malware, researchers find https://bdtechtalks.com/2021/12/09/evilmodel-neural-networks-malware/?utm_source=rss&utm_medium=rss&utm_campaign=evilmodel-neural-networks-malware https://bdtechtalks.com/2021/12/09/evilmodel-neural-networks-malware/#respond Thu, 09 Dec 2021 14:00:00 +0000 https://bdtechtalks.com/?p=12370 researchers at the University of California, San Diego, and the University of Illinois have discovered a technique to embed malware in deep neural networks.

The post Neural networks can hide malware, researchers find first appeared on TechTalks.

]]>
https://bdtechtalks.com/2021/12/09/evilmodel-neural-networks-malware/feed/ 0 12370
Google Research: Self-supervised learning is a game-changer for medical imaging https://bdtechtalks.com/2021/11/08/google-research-self-supervised-learning-medical-imaging/?utm_source=rss&utm_medium=rss&utm_campaign=google-research-self-supervised-learning-medical-imaging https://bdtechtalks.com/2021/11/08/google-research-self-supervised-learning-medical-imaging/#respond Mon, 08 Nov 2021 14:00:00 +0000 https://bdtechtalks.com/?p=12048 Self-supervised learning reduces the need for annotated data in medical imaging applications, new research from Google AI shows.

The post Google Research: Self-supervised learning is a game-changer for medical imaging first appeared on TechTalks.

]]>
https://bdtechtalks.com/2021/11/08/google-research-self-supervised-learning-medical-imaging/feed/ 0 12048